diff --git a/ci/error_logs/HECULES_PR_LOGS/C48_S2SW_logs_2021032312_gfs_atmos_prod_f024-f026.log b/ci/error_logs/HECULES_PR_LOGS/C48_S2SW_logs_2021032312_gfs_atmos_prod_f024-f026.log deleted file mode 100644 index 6c5cccae..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/C48_S2SW_logs_2021032312_gfs_atmos_prod_f024-f026.log +++ /dev/null @@ -1,33529 +0,0 @@ -+ . /work2/noaa/global/mterry/global-workflow_forked/ush/load_fv3gfs_modules.sh -++ [[ hxB == *x* ]] -++ set_x=YES -++ [[ NO == \N\O ]] -++ echo 'Loading modules quietly...' -Loading modules quietly... -++ set +x -Running "module reset". Resetting modules to system default. The following $MODULEPATH directories have been removed: /work2/noaa/global/mterry/global-workflow_forked/modulefiles - -Currently Loaded Modules: - 1) contrib/0.1 43) ip/4.3.0 - 2) intel-oneapi-compilers/2023.1.0 44) grib-util/1.3.0 - 3) stack-intel/2021.9.0 45) g2tmpl/1.13.0 - 4) intel-oneapi-mpi/2021.9.0 46) gsi-ncdiag/1.1.2 - 5) stack-intel-oneapi-mpi/2021.9.0 47) crtm-fix/2.4.0.1_emc - 6) intel-oneapi-mkl/2023.1.0 48) git-lfs/3.1.2 - 7) zlib/1.2.13 49) crtm/2.4.0.1 - 8) pigz/2.7 50) openblas/0.3.24 - 9) zstd/1.5.2 51) py-setuptools/63.4.3 - 10) tar/1.34 52) py-numpy/1.23.4 - 11) gettext/0.21.1 53) bufr/11.7.0 - 12) libxcrypt/4.4.35 54) gmake/4.2.1 - 13) sqlite/3.43.2 55) wgrib2/2.0.8 - 14) util-linux-uuid/2.38.1 56) py-f90nml/1.4.3 - 15) python/3.11.6 57) py-cftime/1.0.3.4 - 16) libjpeg/2.1.0 58) py-netcdf4/1.5.8 - 17) jasper/2.0.32 59) libyaml/0.2.5 - 18) libpng/1.6.37 60) py-pyyaml/6.0 - 19) openjpeg/2.3.1 61) py-markupsafe/2.1.3 - 20) eccodes/2.32.0 62) py-jinja2/3.1.2 - 21) fftw/3.3.10 63) py-bottleneck/1.3.7 - 22) nghttp2/1.57.0 64) py-numexpr/2.8.4 - 23) curl/8.4.0 65) py-et-xmlfile/1.0.1 - 24) proj/9.2.1 66) py-openpyxl/3.1.2 - 25) udunits/2.2.28 67) py-pytz/2023.3 - 26) cdo/2.2.0 68) py-pyxlsb/1.0.10 - 27) hdf5/1.14.0 69) py-xlrd/2.0.1 - 28) snappy/1.1.10 70) py-xlsxwriter/3.1.7 - 29) c-blosc/1.21.5 71) py-xlwt/1.3.0 - 30) netcdf-c/4.9.2 72) py-pandas/1.5.3 - 31) netcdf-fortran/4.6.1 73) py-six/1.16.0 - 32) parallel-netcdf/1.12.2 74) py-python-dateutil/2.8.2 - 33) parallelio/2.5.10 75) g2c/1.8.0 - 34) esmf/8.6.0 76) netcdf-cxx4/4.3.1 - 35) antlr/2.7.7 77) met/9.1.3 - 36) gsl/2.7.1 78) metplus/3.1.1 - 37) nco/5.0.6 79) py-packaging/23.1 - 38) bacio/2.4.1 80) py-xarray/2023.7.0 - 39) w3emc/2.10.0 81) prepobs/1.1.0 - 40) prod_util/2.1.1 82) fit2obs/1.1.7.1 - 41) g2/3.4.5 83) globus-cli/3.35.2 - 42) sp/2.5.0 84) module_base.hercules - - - -++ [[ -d /work2/noaa/global/mterry/global-workflow_forked/sorc/wxflow/src ]] -++ PYTHONPATH=/apps/other/globus-cli-3.35.2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:/work2/noaa/global/mterry/global-workflow_forked/sorc/wxflow/src:/work2/noaa/global/mterry/global-workflow_forked/sorc/wxflow/src -++ PYTHONPATH=/apps/other/globus-cli-3.35.2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:/work2/noaa/global/mterry/global-workflow_forked/sorc/wxflow/src:/work2/noaa/global/mterry/global-workflow_forked/sorc/wxflow/src:/work2/noaa/global/mterry/global-workflow_forked/ush/python -++ export PYTHONPATH -++ ulimit -S -s 16384 -++ unset ulimit_s -+ status=0 -+ (( status != 0 )) -+ export job=atmos_products -+ job=atmos_products -+ IFS=', ' -+ read -r -a fhr_list -+ export FORECAST_HOUR jobid -+ for FORECAST_HOUR in "${fhr_list[@]}" -++ printf %03d 24 -+ fhr3=024 -+ jobid=atmos_products_f024.533601 -+ /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_ATMOS_PRODUCTS -Begin /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_ATMOS_PRODUCTS at Tue Jul 29 02:24:21 UTC 2025 -++ jjob_header.sh[46]OPTIND=1 -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[50]env_job=atmos_products -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[49]read -ra configs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[61]shift 4 -++ jjob_header.sh[63][[ -z atmos_products ]] -++ jjob_header.sh[71]export DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601 -++ jjob_header.sh[71]DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601 -++ jjob_header.sh[72][[ YES == \Y\E\S ]] -++ jjob_header.sh[73]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601 -++ jjob_header.sh[75]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601 -++ jjob_header.sh[76]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601 -++ jjob_header.sh[85]export pid=534058 -++ jjob_header.sh[85]pid=534058 -++ jjob_header.sh[86]export pgmout=OUTPUT.534058 -++ jjob_header.sh[86]pgmout=OUTPUT.534058 -++ jjob_header.sh[87]export pgmerr=errfile -++ jjob_header.sh[87]pgmerr=errfile -++ jjob_header.sh[90]export pgm= -++ jjob_header.sh[90]pgm= -++ jjob_header.sh[96]export cycle=t12z -++ jjob_header.sh[96]cycle=t12z -++ jjob_header.sh[97]setpdy.sh -+ setpdy.sh[20]'[' /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601 == /home/mterry ']' -+ setpdy.sh[25][[ ! t12z =~ t??z ]] -+ setpdy.sh[30]case $# in -+ setpdy.sh[31]dates_before_PDY=7 -+ setpdy.sh[32]dates_after_PDY=7 -+ setpdy.sh[50]COMDATEROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+ setpdy.sh[53]'[' -z 20210323 ']' -+ setpdy.sh[57]sed 's/[0-9]\{8\}/20210323/' /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z -sed: can't read /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z: No such file or directory -++ jjob_header.sh[97]true -++ jjob_header.sh[98]source ./PDY -/work2/noaa/global/mterry/global-workflow_forked/ush/jjob_header.sh: line 98: ./PDY: No such file or directory -++ jjob_header.sh[98]true -++ jjob_header.sh[104]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[104]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.base -+++ config.base[6]echo 'BEGIN: config.base' -BEGIN: config.base -+++ config.base[9]export machine=HERCULES -+++ config.base[9]machine=HERCULES -+++ config.base[12]export RUN_ENVIR=emc -+++ config.base[12]RUN_ENVIR=emc -+++ config.base[15]export ACCOUNT=fv3-cpu -+++ config.base[15]ACCOUNT=fv3-cpu -+++ config.base[16]export QUEUE=batch -+++ config.base[16]QUEUE=batch -+++ config.base[17]export QUEUE_SERVICE=batch -+++ config.base[17]QUEUE_SERVICE=batch -+++ config.base[18]export QUEUE_DTN=batch -+++ config.base[18]QUEUE_DTN=batch -+++ config.base[19]export PARTITION_BATCH=hercules -+++ config.base[19]PARTITION_BATCH=hercules -+++ config.base[20]export PARTITION_SERVICE=service -+++ config.base[20]PARTITION_SERVICE=service -+++ config.base[21]export PARTITION_DTN= -+++ config.base[21]PARTITION_DTN= -+++ config.base[22]export RESERVATION= -+++ config.base[22]RESERVATION= -+++ config.base[23]export CLUSTERS= -+++ config.base[23]CLUSTERS= -+++ config.base[24]export CLUSTERS_SERVICE= -+++ config.base[24]CLUSTERS_SERVICE= -+++ config.base[25]export CLUSTERS_DTN= -+++ config.base[25]CLUSTERS_DTN= -+++ config.base[28]export HPSS_PROJECT=emc-global -+++ config.base[28]HPSS_PROJECT=emc-global -+++ config.base[31]export HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[31]HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[32]export EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[32]EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[33]export FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[33]FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[34]export PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[34]PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[35]export SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[35]SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[36]export USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[36]USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[38]export FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[38]FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[39]export FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[39]FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[40]export FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[40]FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[41]export FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[41]FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[42]export FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[42]FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[43]export FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[43]FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[44]export FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[44]FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[45]export FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[45]FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[50]export PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[50]PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[51]export COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[51]COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[52]export COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[52]COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[53]export DMPDIR=/work/noaa/rstprod/dump -+++ config.base[53]DMPDIR=/work/noaa/rstprod/dump -+++ config.base[57]export COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[57]COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[58]export COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[58]COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[59]export COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[59]COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[62]export HOMEDIR=/work2/noaa/global/mterry -+++ config.base[62]HOMEDIR=/work2/noaa/global/mterry -+++ config.base[63]export STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[63]STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]export PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[65]export NOSCRUB=/work2/noaa/global/mterry -+++ config.base[65]NOSCRUB=/work2/noaa/global/mterry -+++ config.base[68]export BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[68]BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[71]export BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[71]BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[74]export DO_PREP_SFC=NO -+++ config.base[74]DO_PREP_SFC=NO -+++ config.base[77]export DO_GOES=NO -+++ config.base[77]DO_GOES=NO -+++ config.base[78]export DO_BUFRSND=NO -+++ config.base[78]DO_BUFRSND=NO -+++ config.base[79]export DO_GEMPAK=NO -+++ config.base[79]DO_GEMPAK=NO -+++ config.base[80]export DO_AWIPS=NO -+++ config.base[80]DO_AWIPS=NO -+++ config.base[81]export DO_NPOESS=NO -+++ config.base[81]DO_NPOESS=NO -+++ config.base[82]export DO_TRACKER=YES -+++ config.base[82]DO_TRACKER=YES -+++ config.base[83]export DO_GENESIS=YES -+++ config.base[83]DO_GENESIS=YES -+++ config.base[84]export DO_GENESIS_FSU=NO -+++ config.base[84]DO_GENESIS_FSU=NO -+++ config.base[85]export DO_VERFOZN=YES -+++ config.base[85]DO_VERFOZN=YES -+++ config.base[86]export DO_VERFRAD=YES -+++ config.base[86]DO_VERFRAD=YES -+++ config.base[87]export DO_VMINMON=YES -+++ config.base[87]DO_VMINMON=YES -+++ config.base[88]export DO_ANLSTAT=NO -+++ config.base[88]DO_ANLSTAT=NO -+++ config.base[91]export MODE=forecast-only -+++ config.base[91]MODE=forecast-only -+++ config.base[92]export DO_TEST_MODE=YES -+++ config.base[92]DO_TEST_MODE=YES -+++ config.base[101]export FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[101]FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[102]export HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[102]HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[103]export HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v1.2.0 -+++ config.base[103]HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v1.2.0 -+++ config.base[106]export NMV=/bin/mv -+++ config.base[106]NMV=/bin/mv -+++ config.base[107]export 'NLN=/bin/ln -sf' -+++ config.base[107]NLN='/bin/ln -sf' -+++ config.base[108]export VERBOSE=YES -+++ config.base[108]VERBOSE=YES -+++ config.base[109]export KEEPDATA=NO -+++ config.base[109]KEEPDATA=NO -+++ config.base[110]export DEBUG_POSTSCRIPT=NO -+++ config.base[110]DEBUG_POSTSCRIPT=NO -+++ config.base[111]export CHGRP_RSTPROD=YES -+++ config.base[111]CHGRP_RSTPROD=YES -+++ config.base[112]export 'CHGRP_CMD=chgrp rstprod' -+++ config.base[112]CHGRP_CMD='chgrp rstprod' -+++ config.base[113]export NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[113]NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[114]export NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[114]NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[117]export BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[117]BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[120]export SDATE=2021032312 -+++ config.base[120]SDATE=2021032312 -+++ config.base[121]export EDATE=2021032312 -+++ config.base[121]EDATE=2021032312 -+++ config.base[122]export EXP_WARM_START=.false. -+++ config.base[122]EXP_WARM_START=.false. -+++ config.base[123]export assim_freq=6 -+++ config.base[123]assim_freq=6 -+++ config.base[124]export PSLOT=C48_S2SW -+++ config.base[124]PSLOT=C48_S2SW -+++ config.base[125]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[125]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[126]export ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[126]ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[127]export DUMP_SUFFIX= -+++ config.base[127]DUMP_SUFFIX= -+++ config.base[128][[ 2021032312 -ge 2019092100 ]] -+++ config.base[128][[ 2021032312 -le 2019110700 ]] -+++ config.base[131]export ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[131]ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[132]export ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[132]ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[133]export FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[133]FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[136]export envir=prod -+++ config.base[136]envir=prod -+++ config.base[137]export NET=gfs -+++ config.base[137]NET=gfs -+++ config.base[138]export RUN=gfs -+++ config.base[138]RUN=gfs -+++ config.base[141]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.com -++++ config.com[4]echo 'BEGIN: config.com' -BEGIN: config.com -++++ config.com[38][[ emc == \n\c\o ]] -++++ config.com[43]COM_OBSPROC_TMPL='${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos' -++++ config.com[44]COM_RTOFS_TMPL='${DMPDIR}' -++++ config.com[45]COM_TCVITAL_TMPL='${DMPDIR}/${RUN}.${YMD}/${HH}/atmos' -++++ config.com[47]declare -rx 'COM_OBS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/obs' -++++ config.com[48]declare -rx COM_OBSPROC_TMPL COM_RTOFS_TMPL -++++ config.com[50]COM_BASE='${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}' -++++ config.com[52]declare -rx 'COM_TOP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}' -++++ config.com[54]declare -rx 'COM_CONF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf' -++++ config.com[55]declare -rx 'COM_OBS_JEDI=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi' -++++ config.com[57]declare -rx 'COM_ATMOS_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input' -++++ config.com[58]declare -rx 'COM_ATMOS_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart' -++++ config.com[59]declare -rx 'COM_ATMOS_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos' -++++ config.com[60]declare -rx 'COM_SNOW_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow' -++++ config.com[61]declare -rx 'COM_SNOW_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon' -++++ config.com[62]declare -rx 'COM_ATMOS_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history' -++++ config.com[63]declare -rx 'COM_ATMOS_MASTER_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master' -++++ config.com[64]declare -rx 'COM_ATMOS_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2' -++++ config.com[65]declare -rx 'COM_ATMOS_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}' -++++ config.com[66]declare -rx 'COM_ATMOS_BUFR_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr' -++++ config.com[67]declare -rx 'COM_ATMOS_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}' -++++ config.com[68]declare -rx 'COM_ATMOS_GENESIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital' -++++ config.com[69]declare -rx 'COM_ATMOS_TRACK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks' -++++ config.com[70]declare -rx 'COM_ATMOS_GOES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim' -++++ config.com[71]declare -rx 'COM_ATMOS_IMAGERY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery' -++++ config.com[72]declare -rx 'COM_ATMOS_OZNMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon' -++++ config.com[73]declare -rx 'COM_ATMOS_RADMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon' -++++ config.com[74]declare -rx 'COM_ATMOS_MINMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon' -++++ config.com[75]declare -rx 'COM_ATMOS_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon' -++++ config.com[76]declare -rx 'COM_ATMOS_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo' -++++ config.com[78]declare -rx 'COM_WAVE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart' -++++ config.com[79]declare -rx 'COM_WAVE_PREP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep' -++++ config.com[80]declare -rx 'COM_WAVE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history' -++++ config.com[81]declare -rx 'COM_WAVE_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded' -++++ config.com[82]declare -rx 'COM_WAVE_GRID_RES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}' -++++ config.com[83]declare -rx 'COM_WAVE_STATION_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station' -++++ config.com[84]declare -rx 'COM_WAVE_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak' -++++ config.com[85]declare -rx 'COM_WAVE_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo' -++++ config.com[87]declare -rx 'COM_OCEAN_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history' -++++ config.com[88]declare -rx 'COM_OCEAN_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart' -++++ config.com[89]declare -rx 'COM_OCEAN_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input' -++++ config.com[90]declare -rx 'COM_OCEAN_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean' -++++ config.com[91]declare -rx 'COM_OCEAN_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon' -++++ config.com[92]declare -rx 'COM_OCEAN_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf' -++++ config.com[93]declare -rx 'COM_OCEAN_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean' -++++ config.com[94]declare -rx 'COM_OCEAN_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf' -++++ config.com[95]declare -rx 'COM_OCEAN_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2' -++++ config.com[96]declare -rx 'COM_OCEAN_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}' -++++ config.com[98]declare -rx 'COM_ICE_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice' -++++ config.com[99]declare -rx 'COM_ICE_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf' -++++ config.com[100]declare -rx 'COM_ICE_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon' -++++ config.com[101]declare -rx 'COM_ICE_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice' -++++ config.com[102]declare -rx 'COM_ICE_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input' -++++ config.com[103]declare -rx 'COM_ICE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history' -++++ config.com[104]declare -rx 'COM_ICE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart' -++++ config.com[105]declare -rx 'COM_ICE_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf' -++++ config.com[106]declare -rx 'COM_ICE_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2' -++++ config.com[107]declare -rx 'COM_ICE_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}' -++++ config.com[109]declare -rx 'COM_CHEM_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history' -++++ config.com[110]declare -rx 'COM_CHEM_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem' -++++ config.com[111]declare -rx 'COM_CHEM_BMAT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix' -++++ config.com[112]declare -rx 'COM_CHEM_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon' -++++ config.com[114]declare -rx 'COM_MED_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart' -+++ config.base[143]export LOGSCRIPT= -+++ config.base[143]LOGSCRIPT= -+++ config.base[145]export 'REDOUT=1>' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.atmos_products -+++ config.atmos_products[6]echo 'BEGIN: config.atmos_products' -BEGIN: config.atmos_products -+++ config.atmos_products[9]. /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources atmos_products -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=atmos_products -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[1019]walltime=00:15:00 -++++ config.resources[1020]ntasks=24 -++++ config.resources[1021]threads_per_task=1 -++++ config.resources[1022]tasks_per_node=24 -++++ config.resources[1023]export is_exclusive=True -++++ config.resources[1023]is_exclusive=True -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n '' ]] -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n '' ]] -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.atmos_products[12]export MAX_TASKS=25 -+++ config.atmos_products[12]MAX_TASKS=25 -+++ config.atmos_products[15]export INTERP_ATMOS_MASTERSH=/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh -+++ config.atmos_products[15]INTERP_ATMOS_MASTERSH=/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh -+++ config.atmos_products[16]export INTERP_ATMOS_SFLUXSH=/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_sflux.sh -+++ config.atmos_products[16]INTERP_ATMOS_SFLUXSH=/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_sflux.sh -+++ config.atmos_products[18][[ gfs == \g\d\a\s ]] -+++ config.atmos_products[24][[ gfs == \g\f\s ]] -+++ config.atmos_products[25]export downset=2 -+++ config.atmos_products[25]downset=2 -+++ config.atmos_products[26]export FHOUT_PGBS=3 -+++ config.atmos_products[26]FHOUT_PGBS=3 -+++ config.atmos_products[27]export FLXGF=YES -+++ config.atmos_products[27]FLXGF=YES -+++ config.atmos_products[28]export WGNE=YES -+++ config.atmos_products[28]WGNE=YES -+++ config.atmos_products[29]export FHMAX_WGNE=180 -+++ config.atmos_products[29]FHMAX_WGNE=180 -+++ config.atmos_products[32]export APCP_MSG=597 -+++ config.atmos_products[32]APCP_MSG=597 -+++ config.atmos_products[35]export paramlista=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.fFFF.paramlist.a.txt -+++ config.atmos_products[35]paramlista=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.fFFF.paramlist.a.txt -+++ config.atmos_products[36]export paramlista_anl=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.anl.paramlist.a.txt -+++ config.atmos_products[36]paramlista_anl=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.anl.paramlist.a.txt -+++ config.atmos_products[37]export paramlista_f000=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.f000.paramlist.a.txt -+++ config.atmos_products[37]paramlista_f000=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.f000.paramlist.a.txt -+++ config.atmos_products[38]export paramlistb=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.fFFF.paramlist.b.txt -+++ config.atmos_products[38]paramlistb=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.fFFF.paramlist.b.txt -+++ config.atmos_products[40]echo 'END: config.atmos_products' -END: config.atmos_products -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env atmos_products -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=atmos_products -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 520192000 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 24 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 24 ]] -+++ HERCULES.env[34]max_threads_per_task=3 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 3 ]] -+++ HERCULES.env[40][[ 1 -gt 3 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 24' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[281]export USE_CFP=YES -+++ HERCULES.env[281]USE_CFP=YES -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_ATMOS_PRODUCTS[11]YMD=20210323 -+ JGLOBAL_ATMOS_PRODUCTS[11]HH=12 -+ JGLOBAL_ATMOS_PRODUCTS[11]declare_from_tmpl -rx COMIN_ATMOS_ANALYSIS:COM_ATMOS_ANALYSIS_TMPL COMIN_ATMOS_HISTORY:COM_ATMOS_HISTORY_TMPL COMIN_ATMOS_MASTER:COM_ATMOS_MASTER_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_ATMOS_ANALYSIS=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//analysis/atmos -declare_from_tmpl :: COMIN_ATMOS_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/history -declare_from_tmpl :: COMIN_ATMOS_MASTER=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master -+ JGLOBAL_ATMOS_PRODUCTS[16]for grid in '0p25' '0p50' '1p00' -+ JGLOBAL_ATMOS_PRODUCTS[17]prod_dir=COMOUT_ATMOS_GRIB_0p25 -+ JGLOBAL_ATMOS_PRODUCTS[18]GRID=0p25 -+ JGLOBAL_ATMOS_PRODUCTS[18]YMD=20210323 -+ JGLOBAL_ATMOS_PRODUCTS[18]HH=12 -+ JGLOBAL_ATMOS_PRODUCTS[18]declare_from_tmpl -rx COMOUT_ATMOS_GRIB_0p25:COM_ATMOS_GRIB_GRID_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_ATMOS_GRIB_0p25=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25 -+ JGLOBAL_ATMOS_PRODUCTS[19][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25 ]] -+ JGLOBAL_ATMOS_PRODUCTS[16]for grid in '0p25' '0p50' '1p00' -+ JGLOBAL_ATMOS_PRODUCTS[17]prod_dir=COMOUT_ATMOS_GRIB_0p50 -+ JGLOBAL_ATMOS_PRODUCTS[18]GRID=0p50 -+ JGLOBAL_ATMOS_PRODUCTS[18]YMD=20210323 -+ JGLOBAL_ATMOS_PRODUCTS[18]HH=12 -+ JGLOBAL_ATMOS_PRODUCTS[18]declare_from_tmpl -rx COMOUT_ATMOS_GRIB_0p50:COM_ATMOS_GRIB_GRID_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_ATMOS_GRIB_0p50=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50 -+ JGLOBAL_ATMOS_PRODUCTS[19][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50 ]] -+ JGLOBAL_ATMOS_PRODUCTS[16]for grid in '0p25' '0p50' '1p00' -+ JGLOBAL_ATMOS_PRODUCTS[17]prod_dir=COMOUT_ATMOS_GRIB_1p00 -+ JGLOBAL_ATMOS_PRODUCTS[18]GRID=1p00 -+ JGLOBAL_ATMOS_PRODUCTS[18]YMD=20210323 -+ JGLOBAL_ATMOS_PRODUCTS[18]HH=12 -+ JGLOBAL_ATMOS_PRODUCTS[18]declare_from_tmpl -rx COMOUT_ATMOS_GRIB_1p00:COM_ATMOS_GRIB_GRID_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_ATMOS_GRIB_1p00=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00 -+ JGLOBAL_ATMOS_PRODUCTS[19][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00 ]] -+ JGLOBAL_ATMOS_PRODUCTS[23]export PREFIX=gfs.t12z. -+ JGLOBAL_ATMOS_PRODUCTS[23]PREFIX=gfs.t12z. -+ JGLOBAL_ATMOS_PRODUCTS[27]/work2/noaa/global/mterry/global-workflow_forked/scripts/exglobal_atmos_products.sh -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ exglobal_atmos_products.sh[4]INTERP_ATMOS_MASTERSH=/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh -+ exglobal_atmos_products.sh[5]INTERP_ATMOS_SFLUXSH=/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_sflux.sh -+ exglobal_atmos_products.sh[8]downset=2 -+ exglobal_atmos_products.sh[9]ntasks_atmos_products=8 -+ exglobal_atmos_products.sh[12]WGNE=YES -+ exglobal_atmos_products.sh[13]FHMAX_WGNE=180 -+ exglobal_atmos_products.sh[15]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601 -+ exglobal_atmos_products.sh[19][[ 24 -le 0 ]] -++ exglobal_atmos_products.sh[30]printf f%03d 24 -+ exglobal_atmos_products.sh[30]fhr3=f024 -+ exglobal_atmos_products.sh[31](( FORECAST_HOUR%FHOUT_PGBS == 0 )) -+ exglobal_atmos_products.sh[32]PGBS=YES -+ exglobal_atmos_products.sh[40]MASTER_FILE=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.master.grb2f024 -+ exglobal_atmos_products.sh[45]wgrib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.master.grb2f024 -+ exglobal_atmos_products.sh[45]wgrib2 -i -grib tmpfile_f024 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.master.grb2f024 -+ exglobal_atmos_products.sh[45]grep -F -f /work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.fFFF.paramlist.a.txt -1:0:d=2021032312:PRMSL:mean sea level:24 hour fcst: -2:25731:d=2021032312:CLMR:1 hybrid level:24 hour fcst: -3:29524:d=2021032312:ICMR:1 hybrid level:24 hour fcst: -4:37100:d=2021032312:RWMR:1 hybrid level:24 hour fcst: -5:48509:d=2021032312:SNMR:1 hybrid level:24 hour fcst: -6:59635:d=2021032312:GRLE:1 hybrid level:24 hour fcst: -7:62405:d=2021032312:REFD:1 hybrid level:24 hour fcst: -8:88221:d=2021032312:REFD:2 hybrid level:24 hour fcst: -9:114004:d=2021032312:REFC:entire atmosphere (considered as a single layer):24 hour fcst: -10:141439:d=2021032312:VIS:surface:24 hour fcst: -11:173907:d=2021032312:UGRD:planetary boundary layer:24 hour fcst: -12:187864:d=2021032312:VGRD:planetary boundary layer:24 hour fcst: -13:202409:d=2021032312:VRATE:planetary boundary layer:24 hour fcst: -14:217517:d=2021032312:GUST:surface:24 hour fcst: -15:232221:d=2021032312:HGT:0.01 mb:24 hour fcst: -16:257938:d=2021032312:TMP:0.01 mb:24 hour fcst: -17:268909:d=2021032312:RH:0.01 mb:24 hour fcst: -18:275412:d=2021032312:SPFH:0.01 mb:24 hour fcst: -19:298176:d=2021032312:VVEL:0.01 mb:24 hour fcst: -20:325074:d=2021032312:DZDT:0.01 mb:24 hour fcst: -21:352568:d=2021032312:UGRD:0.01 mb:24 hour fcst: -22:363684:d=2021032312:VGRD:0.01 mb:24 hour fcst: -23:374056:d=2021032312:ABSV:0.01 mb:24 hour fcst: -24:389577:d=2021032312:O3MR:0.01 mb:24 hour fcst: -25:403656:d=2021032312:HGT:0.02 mb:24 hour fcst: -26:429080:d=2021032312:TMP:0.02 mb:24 hour fcst: -27:443922:d=2021032312:RH:0.02 mb:24 hour fcst: -28:451652:d=2021032312:SPFH:0.02 mb:24 hour fcst: -29:475844:d=2021032312:VVEL:0.02 mb:24 hour fcst: -30:499814:d=2021032312:DZDT:0.02 mb:24 hour fcst: -31:527345:d=2021032312:UGRD:0.02 mb:24 hour fcst: -32:539841:d=2021032312:VGRD:0.02 mb:24 hour fcst: -33:551546:d=2021032312:ABSV:0.02 mb:24 hour fcst: -34:568507:d=2021032312:O3MR:0.02 mb:24 hour fcst: -35:584083:d=2021032312:HGT:0.04 mb:24 hour fcst: -36:608394:d=2021032312:TMP:0.04 mb:24 hour fcst: -37:625667:d=2021032312:RH:0.04 mb:24 hour fcst: -38:631654:d=2021032312:SPFH:0.04 mb:24 hour fcst: -39:657290:d=2021032312:VVEL:0.04 mb:24 hour fcst: -40:684752:d=2021032312:DZDT:0.04 mb:24 hour fcst: -41:709597:d=2021032312:UGRD:0.04 mb:24 hour fcst: -42:729530:d=2021032312:VGRD:0.04 mb:24 hour fcst: -43:749243:d=2021032312:ABSV:0.04 mb:24 hour fcst: -44:767626:d=2021032312:O3MR:0.04 mb:24 hour fcst: -45:791532:d=2021032312:HGT:0.07 mb:24 hour fcst: -46:815629:d=2021032312:TMP:0.07 mb:24 hour fcst: -47:832541:d=2021032312:RH:0.07 mb:24 hour fcst: -48:840370:d=2021032312:SPFH:0.07 mb:24 hour fcst: -49:864128:d=2021032312:VVEL:0.07 mb:24 hour fcst: -50:893051:d=2021032312:DZDT:0.07 mb:24 hour fcst: -51:918481:d=2021032312:UGRD:0.07 mb:24 hour fcst: -52:938837:d=2021032312:VGRD:0.07 mb:24 hour fcst: -53:959146:d=2021032312:ABSV:0.07 mb:24 hour fcst: -54:978056:d=2021032312:O3MR:0.07 mb:24 hour fcst: -55:997869:d=2021032312:HGT:0.1 mb:24 hour fcst: -56:1019441:d=2021032312:TMP:0.1 mb:24 hour fcst: -57:1036109:d=2021032312:RH:0.1 mb:24 hour fcst: -58:1044096:d=2021032312:SPFH:0.1 mb:24 hour fcst: -59:1065442:d=2021032312:VVEL:0.1 mb:24 hour fcst: -60:1095289:d=2021032312:DZDT:0.1 mb:24 hour fcst: -61:1121161:d=2021032312:UGRD:0.1 mb:24 hour fcst: -62:1141154:d=2021032312:VGRD:0.1 mb:24 hour fcst: -63:1161201:d=2021032312:ABSV:0.1 mb:24 hour fcst: -64:1179757:d=2021032312:O3MR:0.1 mb:24 hour fcst: -65:1203381:d=2021032312:HGT:0.2 mb:24 hour fcst: -66:1224532:d=2021032312:TMP:0.2 mb:24 hour fcst: -67:1240846:d=2021032312:RH:0.2 mb:24 hour fcst: -68:1251778:d=2021032312:SPFH:0.2 mb:24 hour fcst: -69:1271448:d=2021032312:VVEL:0.2 mb:24 hour fcst: -70:1300400:d=2021032312:DZDT:0.2 mb:24 hour fcst: -71:1327937:d=2021032312:UGRD:0.2 mb:24 hour fcst: -72:1347357:d=2021032312:VGRD:0.2 mb:24 hour fcst: -73:1366271:d=2021032312:ABSV:0.2 mb:24 hour fcst: -74:1383935:d=2021032312:O3MR:0.2 mb:24 hour fcst: -75:1409717:d=2021032312:HGT:0.4 mb:24 hour fcst: -76:1430479:d=2021032312:TMP:0.4 mb:24 hour fcst: -77:1446494:d=2021032312:RH:0.4 mb:24 hour fcst: -78:1455603:d=2021032312:SPFH:0.4 mb:24 hour fcst: -79:1475567:d=2021032312:VVEL:0.4 mb:24 hour fcst: -80:1500487:d=2021032312:DZDT:0.4 mb:24 hour fcst: -81:1531015:d=2021032312:UGRD:0.4 mb:24 hour fcst: -82:1549911:d=2021032312:VGRD:0.4 mb:24 hour fcst: -83:1568182:d=2021032312:ABSV:0.4 mb:24 hour fcst: -84:1584980:d=2021032312:O3MR:0.4 mb:24 hour fcst: -85:1609600:d=2021032312:HGT:0.7 mb:24 hour fcst: -86:1627732:d=2021032312:TMP:0.7 mb:24 hour fcst: -87:1643578:d=2021032312:RH:0.7 mb:24 hour fcst: -88:1652546:d=2021032312:SPFH:0.7 mb:24 hour fcst: -89:1671792:d=2021032312:VVEL:0.7 mb:24 hour fcst: -90:1697702:d=2021032312:DZDT:0.7 mb:24 hour fcst: -91:1726842:d=2021032312:UGRD:0.7 mb:24 hour fcst: -92:1745287:d=2021032312:VGRD:0.7 mb:24 hour fcst: -93:1763146:d=2021032312:ABSV:0.7 mb:24 hour fcst: -94:1779740:d=2021032312:O3MR:0.7 mb:24 hour fcst: -95:1805811:d=2021032312:HGT:1 mb:24 hour fcst: -96:1824058:d=2021032312:TMP:1 mb:24 hour fcst: -97:1839923:d=2021032312:RH:1 mb:24 hour fcst: -98:1849749:d=2021032312:SPFH:1 mb:24 hour fcst: -99:1868149:d=2021032312:VVEL:1 mb:24 hour fcst: -100:1895003:d=2021032312:DZDT:1 mb:24 hour fcst: -101:1921096:d=2021032312:UGRD:1 mb:24 hour fcst: -102:1939697:d=2021032312:VGRD:1 mb:24 hour fcst: -103:1957398:d=2021032312:ABSV:1 mb:24 hour fcst: -104:1973752:d=2021032312:O3MR:1 mb:24 hour fcst: -105:1994820:d=2021032312:HGT:2 mb:24 hour fcst: -106:2015132:d=2021032312:TMP:2 mb:24 hour fcst: -107:2031736:d=2021032312:RH:2 mb:24 hour fcst: -108:2040317:d=2021032312:SPFH:2 mb:24 hour fcst: -109:2059312:d=2021032312:VVEL:2 mb:24 hour fcst: -110:2087999:d=2021032312:DZDT:2 mb:24 hour fcst: -111:2114485:d=2021032312:UGRD:2 mb:24 hour fcst: -112:2126070:d=2021032312:VGRD:2 mb:24 hour fcst: -113:2144416:d=2021032312:ABSV:2 mb:24 hour fcst: -114:2161499:d=2021032312:O3MR:2 mb:24 hour fcst: -115:2184317:d=2021032312:HGT:3 mb:24 hour fcst: -116:2202153:d=2021032312:TMP:3 mb:24 hour fcst: -117:2218626:d=2021032312:RH:3 mb:24 hour fcst: -118:2224656:d=2021032312:SPFH:3 mb:24 hour fcst: -119:2245914:d=2021032312:VVEL:3 mb:24 hour fcst: -120:2275970:d=2021032312:DZDT:3 mb:24 hour fcst: -121:2302386:d=2021032312:UGRD:3 mb:24 hour fcst: -122:2313769:d=2021032312:VGRD:3 mb:24 hour fcst: -123:2331906:d=2021032312:ABSV:3 mb:24 hour fcst: -124:2348920:d=2021032312:O3MR:3 mb:24 hour fcst: -125:2372293:d=2021032312:HGT:5 mb:24 hour fcst: -126:2389968:d=2021032312:TMP:5 mb:24 hour fcst: -127:2406597:d=2021032312:RH:5 mb:24 hour fcst: -128:2415639:d=2021032312:SPFH:5 mb:24 hour fcst: -129:2436387:d=2021032312:VVEL:5 mb:24 hour fcst: -130:2465571:d=2021032312:DZDT:5 mb:24 hour fcst: -131:2491636:d=2021032312:UGRD:5 mb:24 hour fcst: -132:2503228:d=2021032312:VGRD:5 mb:24 hour fcst: -133:2521798:d=2021032312:ABSV:5 mb:24 hour fcst: -134:2539419:d=2021032312:O3MR:5 mb:24 hour fcst: -135:2562079:d=2021032312:HGT:7 mb:24 hour fcst: -136:2579838:d=2021032312:TMP:7 mb:24 hour fcst: -137:2596027:d=2021032312:RH:7 mb:24 hour fcst: -138:2602146:d=2021032312:SPFH:7 mb:24 hour fcst: -139:2621569:d=2021032312:VVEL:7 mb:24 hour fcst: -140:2646243:d=2021032312:DZDT:7 mb:24 hour fcst: -141:2671752:d=2021032312:UGRD:7 mb:24 hour fcst: -142:2683166:d=2021032312:VGRD:7 mb:24 hour fcst: -143:2701858:d=2021032312:ABSV:7 mb:24 hour fcst: -144:2719667:d=2021032312:O3MR:7 mb:24 hour fcst: -145:2743441:d=2021032312:HGT:10 mb:24 hour fcst: -146:2763289:d=2021032312:TMP:10 mb:24 hour fcst: -147:2779445:d=2021032312:RH:10 mb:24 hour fcst: -148:2787087:d=2021032312:SPFH:10 mb:24 hour fcst: -149:2805333:d=2021032312:VVEL:10 mb:24 hour fcst: -150:2830855:d=2021032312:DZDT:10 mb:24 hour fcst: -151:2856265:d=2021032312:UGRD:10 mb:24 hour fcst: -152:2867606:d=2021032312:VGRD:10 mb:24 hour fcst: -153:2886269:d=2021032312:ABSV:10 mb:24 hour fcst: -154:2904103:d=2021032312:O3MR:10 mb:24 hour fcst: -155:2922881:d=2021032312:HGT:15 mb:24 hour fcst: -156:2942392:d=2021032312:TMP:15 mb:24 hour fcst: -157:2958577:d=2021032312:RH:15 mb:24 hour fcst: -158:2963889:d=2021032312:SPFH:15 mb:24 hour fcst: -159:2979362:d=2021032312:VVEL:15 mb:24 hour fcst: -160:3005936:d=2021032312:DZDT:15 mb:24 hour fcst: -161:3036315:d=2021032312:UGRD:15 mb:24 hour fcst: -162:3047581:d=2021032312:VGRD:15 mb:24 hour fcst: -163:3065874:d=2021032312:ABSV:15 mb:24 hour fcst: -164:3083430:d=2021032312:O3MR:15 mb:24 hour fcst: -165:3102481:d=2021032312:HGT:20 mb:24 hour fcst: -166:3121899:d=2021032312:TMP:20 mb:24 hour fcst: -167:3137954:d=2021032312:RH:20 mb:24 hour fcst: -168:3144158:d=2021032312:SPFH:20 mb:24 hour fcst: -169:3158568:d=2021032312:VVEL:20 mb:24 hour fcst: -170:3185926:d=2021032312:DZDT:20 mb:24 hour fcst: -171:3216210:d=2021032312:UGRD:20 mb:24 hour fcst: -172:3234848:d=2021032312:VGRD:20 mb:24 hour fcst: -173:3253386:d=2021032312:ABSV:20 mb:24 hour fcst: -174:3270995:d=2021032312:O3MR:20 mb:24 hour fcst: -175:3295921:d=2021032312:HGT:30 mb:24 hour fcst: -176:3315236:d=2021032312:TMP:30 mb:24 hour fcst: -177:3331242:d=2021032312:RH:30 mb:24 hour fcst: -178:3338836:d=2021032312:SPFH:30 mb:24 hour fcst: -179:3354139:d=2021032312:VVEL:30 mb:24 hour fcst: -180:3382646:d=2021032312:DZDT:30 mb:24 hour fcst: -181:3412625:d=2021032312:UGRD:30 mb:24 hour fcst: -182:3431328:d=2021032312:VGRD:30 mb:24 hour fcst: -183:3449908:d=2021032312:ABSV:30 mb:24 hour fcst: -184:3467608:d=2021032312:O3MR:30 mb:24 hour fcst: -185:3494438:d=2021032312:HGT:40 mb:24 hour fcst: -186:3513288:d=2021032312:TMP:40 mb:24 hour fcst: -187:3529623:d=2021032312:RH:40 mb:24 hour fcst: -188:3539149:d=2021032312:SPFH:40 mb:24 hour fcst: -189:3556233:d=2021032312:VVEL:40 mb:24 hour fcst: -190:3585634:d=2021032312:DZDT:40 mb:24 hour fcst: -191:3615576:d=2021032312:UGRD:40 mb:24 hour fcst: -192:3634336:d=2021032312:VGRD:40 mb:24 hour fcst: -193:3653066:d=2021032312:ABSV:40 mb:24 hour fcst: -194:3670911:d=2021032312:O3MR:40 mb:24 hour fcst: -195:3697855:d=2021032312:HGT:50 mb:24 hour fcst: -196:3716760:d=2021032312:TMP:50 mb:24 hour fcst: -197:3732896:d=2021032312:RH:50 mb:24 hour fcst: -198:3743536:d=2021032312:TCDC:50 mb:24 hour fcst: -199:3743715:d=2021032312:SPFH:50 mb:24 hour fcst: -200:3762260:d=2021032312:VVEL:50 mb:24 hour fcst: -201:3792355:d=2021032312:DZDT:50 mb:24 hour fcst: -202:3821970:d=2021032312:UGRD:50 mb:24 hour fcst: -203:3840451:d=2021032312:VGRD:50 mb:24 hour fcst: -204:3858977:d=2021032312:ABSV:50 mb:24 hour fcst: -205:3876672:d=2021032312:CLMR:50 mb:24 hour fcst: -206:3876851:d=2021032312:ICMR:50 mb:24 hour fcst: -207:3877030:d=2021032312:RWMR:50 mb:24 hour fcst: -208:3877209:d=2021032312:SNMR:50 mb:24 hour fcst: -209:3877388:d=2021032312:GRLE:50 mb:24 hour fcst: -210:3877567:d=2021032312:O3MR:50 mb:24 hour fcst: -211:3903927:d=2021032312:HGT:70 mb:24 hour fcst: -212:3923007:d=2021032312:TMP:70 mb:24 hour fcst: -213:3939512:d=2021032312:RH:70 mb:24 hour fcst: -215:3947922:d=2021032312:SPFH:70 mb:24 hour fcst: -216:3968899:d=2021032312:VVEL:70 mb:24 hour fcst: -217:4000320:d=2021032312:DZDT:70 mb:24 hour fcst: -218:4032423:d=2021032312:UGRD:70 mb:24 hour fcst: -219:4051116:d=2021032312:VGRD:70 mb:24 hour fcst: -220:4069767:d=2021032312:ABSV:70 mb:24 hour fcst: -226:4088471:d=2021032312:O3MR:70 mb:24 hour fcst: -227:4114275:d=2021032312:HGT:100 mb:24 hour fcst: -228:4133924:d=2021032312:TMP:100 mb:24 hour fcst: -229:4150842:d=2021032312:RH:100 mb:24 hour fcst: -230:4160991:d=2021032312:TCDC:100 mb:24 hour fcst: -231:4161170:d=2021032312:SPFH:100 mb:24 hour fcst: -232:4187004:d=2021032312:VVEL:100 mb:24 hour fcst: -233:4212061:d=2021032312:DZDT:100 mb:24 hour fcst: -234:4236673:d=2021032312:UGRD:100 mb:24 hour fcst: -235:4256039:d=2021032312:VGRD:100 mb:24 hour fcst: -236:4275471:d=2021032312:ABSV:100 mb:24 hour fcst: -237:4293660:d=2021032312:CLMR:100 mb:24 hour fcst: -238:4293839:d=2021032312:ICMR:100 mb:24 hour fcst: -239:4294645:d=2021032312:RWMR:100 mb:24 hour fcst: -240:4294824:d=2021032312:SNMR:100 mb:24 hour fcst: -241:4296791:d=2021032312:GRLE:100 mb:24 hour fcst: -242:4296970:d=2021032312:O3MR:100 mb:24 hour fcst: -259:4537904:d=2021032312:HGT:150 mb:24 hour fcst: -260:4558304:d=2021032312:TMP:150 mb:24 hour fcst: -261:4575273:d=2021032312:RH:150 mb:24 hour fcst: -262:4586732:d=2021032312:TCDC:150 mb:24 hour fcst: -263:4590763:d=2021032312:SPFH:150 mb:24 hour fcst: -264:4615140:d=2021032312:VVEL:150 mb:24 hour fcst: -265:4642452:d=2021032312:DZDT:150 mb:24 hour fcst: -266:4667945:d=2021032312:UGRD:150 mb:24 hour fcst: -267:4687918:d=2021032312:VGRD:150 mb:24 hour fcst: -268:4707944:d=2021032312:ABSV:150 mb:24 hour fcst: -269:4726822:d=2021032312:CLMR:150 mb:24 hour fcst: -270:4727001:d=2021032312:ICMR:150 mb:24 hour fcst: -271:4732273:d=2021032312:RWMR:150 mb:24 hour fcst: -272:4732452:d=2021032312:SNMR:150 mb:24 hour fcst: -273:4735971:d=2021032312:GRLE:150 mb:24 hour fcst: -274:4736523:d=2021032312:O3MR:150 mb:24 hour fcst: -291:4981186:d=2021032312:HGT:200 mb:24 hour fcst: -292:5002224:d=2021032312:TMP:200 mb:24 hour fcst: -293:5020329:d=2021032312:RH:200 mb:24 hour fcst: -294:5035415:d=2021032312:TCDC:200 mb:24 hour fcst: -295:5041907:d=2021032312:SPFH:200 mb:24 hour fcst: -296:5067676:d=2021032312:VVEL:200 mb:24 hour fcst: -297:5094411:d=2021032312:DZDT:200 mb:24 hour fcst: -298:5120760:d=2021032312:UGRD:200 mb:24 hour fcst: -299:5134163:d=2021032312:VGRD:200 mb:24 hour fcst: -300:5155440:d=2021032312:ABSV:200 mb:24 hour fcst: -301:5175388:d=2021032312:CLMR:200 mb:24 hour fcst: -302:5175567:d=2021032312:ICMR:200 mb:24 hour fcst: -303:5183978:d=2021032312:RWMR:200 mb:24 hour fcst: -304:5184157:d=2021032312:SNMR:200 mb:24 hour fcst: -305:5189465:d=2021032312:GRLE:200 mb:24 hour fcst: -306:5189884:d=2021032312:O3MR:200 mb:24 hour fcst: -323:5446502:d=2021032312:HGT:250 mb:24 hour fcst: -324:5468427:d=2021032312:TMP:250 mb:24 hour fcst: -325:5487313:d=2021032312:RH:250 mb:24 hour fcst: -326:5505104:d=2021032312:TCDC:250 mb:24 hour fcst: -327:5515642:d=2021032312:SPFH:250 mb:24 hour fcst: -328:5543945:d=2021032312:VVEL:250 mb:24 hour fcst: -329:5572330:d=2021032312:DZDT:250 mb:24 hour fcst: -330:5599477:d=2021032312:UGRD:250 mb:24 hour fcst: -331:5613777:d=2021032312:VGRD:250 mb:24 hour fcst: -332:5628594:d=2021032312:ABSV:250 mb:24 hour fcst: -333:5650043:d=2021032312:CLMR:250 mb:24 hour fcst: -334:5650222:d=2021032312:ICMR:250 mb:24 hour fcst: -335:5664485:d=2021032312:RWMR:250 mb:24 hour fcst: -336:5664664:d=2021032312:SNMR:250 mb:24 hour fcst: -337:5673469:d=2021032312:GRLE:250 mb:24 hour fcst: -338:5673813:d=2021032312:O3MR:250 mb:24 hour fcst: -355:5950611:d=2021032312:HGT:300 mb:24 hour fcst: -356:5972921:d=2021032312:TMP:300 mb:24 hour fcst: -357:5991489:d=2021032312:RH:300 mb:24 hour fcst: -358:6010488:d=2021032312:TCDC:300 mb:24 hour fcst: -359:6023220:d=2021032312:SPFH:300 mb:24 hour fcst: -360:6049924:d=2021032312:VVEL:300 mb:24 hour fcst: -361:6073839:d=2021032312:DZDT:300 mb:24 hour fcst: -362:6101987:d=2021032312:UGRD:300 mb:24 hour fcst: -363:6116492:d=2021032312:VGRD:300 mb:24 hour fcst: -364:6131814:d=2021032312:ABSV:300 mb:24 hour fcst: -365:6154123:d=2021032312:CLMR:300 mb:24 hour fcst: -366:6154302:d=2021032312:ICMR:300 mb:24 hour fcst: -367:6170121:d=2021032312:RWMR:300 mb:24 hour fcst: -368:6170300:d=2021032312:SNMR:300 mb:24 hour fcst: -369:6181701:d=2021032312:GRLE:300 mb:24 hour fcst: -370:6182112:d=2021032312:O3MR:300 mb:24 hour fcst: -387:6469045:d=2021032312:HGT:350 mb:24 hour fcst: -388:6491182:d=2021032312:TMP:350 mb:24 hour fcst: -389:6509396:d=2021032312:RH:350 mb:24 hour fcst: -390:6528509:d=2021032312:TCDC:350 mb:24 hour fcst: -391:6541601:d=2021032312:SPFH:350 mb:24 hour fcst: -392:6569784:d=2021032312:VVEL:350 mb:24 hour fcst: -393:6594311:d=2021032312:DZDT:350 mb:24 hour fcst: -394:6622631:d=2021032312:UGRD:350 mb:24 hour fcst: -395:6637221:d=2021032312:VGRD:350 mb:24 hour fcst: -396:6652471:d=2021032312:ABSV:350 mb:24 hour fcst: -397:6674740:d=2021032312:CLMR:350 mb:24 hour fcst: -398:6674919:d=2021032312:ICMR:350 mb:24 hour fcst: -399:6691733:d=2021032312:RWMR:350 mb:24 hour fcst: -400:6691912:d=2021032312:SNMR:350 mb:24 hour fcst: -401:6705191:d=2021032312:GRLE:350 mb:24 hour fcst: -402:6705698:d=2021032312:O3MR:350 mb:24 hour fcst: -419:6999373:d=2021032312:HGT:400 mb:24 hour fcst: -420:7021126:d=2021032312:TMP:400 mb:24 hour fcst: -421:7039310:d=2021032312:RH:400 mb:24 hour fcst: -422:7058511:d=2021032312:TCDC:400 mb:24 hour fcst: -423:7071628:d=2021032312:SPFH:400 mb:24 hour fcst: -424:7099496:d=2021032312:VVEL:400 mb:24 hour fcst: -425:7124452:d=2021032312:DZDT:400 mb:24 hour fcst: -426:7152845:d=2021032312:UGRD:400 mb:24 hour fcst: -427:7166920:d=2021032312:VGRD:400 mb:24 hour fcst: -428:7189630:d=2021032312:ABSV:400 mb:24 hour fcst: -429:7211592:d=2021032312:CLMR:400 mb:24 hour fcst: -430:7211966:d=2021032312:ICMR:400 mb:24 hour fcst: -431:7228951:d=2021032312:RWMR:400 mb:24 hour fcst: -432:7229239:d=2021032312:SNMR:400 mb:24 hour fcst: -433:7243162:d=2021032312:GRLE:400 mb:24 hour fcst: -434:7243974:d=2021032312:O3MR:400 mb:24 hour fcst: -451:7548487:d=2021032312:HGT:450 mb:24 hour fcst: -452:7572449:d=2021032312:TMP:450 mb:24 hour fcst: -453:7590825:d=2021032312:RH:450 mb:24 hour fcst: -454:7610235:d=2021032312:TCDC:450 mb:24 hour fcst: -455:7622962:d=2021032312:SPFH:450 mb:24 hour fcst: -456:7652439:d=2021032312:VVEL:450 mb:24 hour fcst: -457:7677597:d=2021032312:DZDT:450 mb:24 hour fcst: -458:7705746:d=2021032312:UGRD:450 mb:24 hour fcst: -459:7727430:d=2021032312:VGRD:450 mb:24 hour fcst: -460:7749758:d=2021032312:ABSV:450 mb:24 hour fcst: -461:7771200:d=2021032312:CLMR:450 mb:24 hour fcst: -462:7772139:d=2021032312:ICMR:450 mb:24 hour fcst: -463:7788700:d=2021032312:RWMR:450 mb:24 hour fcst: -464:7789219:d=2021032312:SNMR:450 mb:24 hour fcst: -465:7804097:d=2021032312:GRLE:450 mb:24 hour fcst: -466:7805366:d=2021032312:O3MR:450 mb:24 hour fcst: -483:8106885:d=2021032312:HGT:500 mb:24 hour fcst: -484:8130612:d=2021032312:TMP:500 mb:24 hour fcst: -485:8149153:d=2021032312:RH:500 mb:24 hour fcst: -486:8168728:d=2021032312:TCDC:500 mb:24 hour fcst: -487:8180930:d=2021032312:SPFH:500 mb:24 hour fcst: -488:8209277:d=2021032312:VVEL:500 mb:24 hour fcst: -489:8234350:d=2021032312:DZDT:500 mb:24 hour fcst: -490:8262073:d=2021032312:UGRD:500 mb:24 hour fcst: -491:8283510:d=2021032312:VGRD:500 mb:24 hour fcst: -492:8305550:d=2021032312:ABSV:500 mb:24 hour fcst: -493:8326719:d=2021032312:CLMR:500 mb:24 hour fcst: -494:8328363:d=2021032312:ICMR:500 mb:24 hour fcst: -495:8345751:d=2021032312:RWMR:500 mb:24 hour fcst: -496:8346692:d=2021032312:SNMR:500 mb:24 hour fcst: -497:8360966:d=2021032312:GRLE:500 mb:24 hour fcst: -498:8362874:d=2021032312:O3MR:500 mb:24 hour fcst: -515:8665310:d=2021032312:HGT:550 mb:24 hour fcst: -516:8688877:d=2021032312:TMP:550 mb:24 hour fcst: -517:8707550:d=2021032312:RH:550 mb:24 hour fcst: -518:8727398:d=2021032312:TCDC:550 mb:24 hour fcst: -519:8739577:d=2021032312:SPFH:550 mb:24 hour fcst: -520:8769168:d=2021032312:VVEL:550 mb:24 hour fcst: -521:8794415:d=2021032312:DZDT:550 mb:24 hour fcst: -522:8821952:d=2021032312:UGRD:550 mb:24 hour fcst: -523:8843125:d=2021032312:VGRD:550 mb:24 hour fcst: -524:8864888:d=2021032312:ABSV:550 mb:24 hour fcst: -525:8885842:d=2021032312:CLMR:550 mb:24 hour fcst: -526:8889037:d=2021032312:ICMR:550 mb:24 hour fcst: -527:8905609:d=2021032312:RWMR:550 mb:24 hour fcst: -528:8907128:d=2021032312:SNMR:550 mb:24 hour fcst: -529:8921523:d=2021032312:GRLE:550 mb:24 hour fcst: -530:8923364:d=2021032312:O3MR:550 mb:24 hour fcst: -547:9228612:d=2021032312:HGT:600 mb:24 hour fcst: -548:9252837:d=2021032312:TMP:600 mb:24 hour fcst: -549:9271858:d=2021032312:RH:600 mb:24 hour fcst: -550:9291935:d=2021032312:TCDC:600 mb:24 hour fcst: -551:9304051:d=2021032312:SPFH:600 mb:24 hour fcst: -552:9334496:d=2021032312:VVEL:600 mb:24 hour fcst: -553:9359909:d=2021032312:DZDT:600 mb:24 hour fcst: -554:9387468:d=2021032312:UGRD:600 mb:24 hour fcst: -555:9408513:d=2021032312:VGRD:600 mb:24 hour fcst: -556:9430135:d=2021032312:ABSV:600 mb:24 hour fcst: -557:9450895:d=2021032312:CLMR:600 mb:24 hour fcst: -558:9455670:d=2021032312:ICMR:600 mb:24 hour fcst: -559:9469603:d=2021032312:RWMR:600 mb:24 hour fcst: -560:9473144:d=2021032312:SNMR:600 mb:24 hour fcst: -561:9488965:d=2021032312:GRLE:600 mb:24 hour fcst: -562:9491363:d=2021032312:O3MR:600 mb:24 hour fcst: -579:9807719:d=2021032312:HGT:650 mb:24 hour fcst: -580:9831895:d=2021032312:TMP:650 mb:24 hour fcst: -581:9851233:d=2021032312:RH:650 mb:24 hour fcst: -582:9871483:d=2021032312:TCDC:650 mb:24 hour fcst: -583:9883200:d=2021032312:SPFH:650 mb:24 hour fcst: -584:9911875:d=2021032312:VVEL:650 mb:24 hour fcst: -585:9937427:d=2021032312:DZDT:650 mb:24 hour fcst: -586:9964823:d=2021032312:UGRD:650 mb:24 hour fcst: -587:9985700:d=2021032312:VGRD:650 mb:24 hour fcst: -588:10007395:d=2021032312:ABSV:650 mb:24 hour fcst: -589:10028299:d=2021032312:CLMR:650 mb:24 hour fcst: -590:10033521:d=2021032312:ICMR:650 mb:24 hour fcst: -591:10045957:d=2021032312:RWMR:650 mb:24 hour fcst: -592:10050527:d=2021032312:SNMR:650 mb:24 hour fcst: -593:10065594:d=2021032312:GRLE:650 mb:24 hour fcst: -594:10067872:d=2021032312:O3MR:650 mb:24 hour fcst: -611:10385881:d=2021032312:HGT:700 mb:24 hour fcst: -612:10410219:d=2021032312:TMP:700 mb:24 hour fcst: -613:10429826:d=2021032312:RH:700 mb:24 hour fcst: -614:10450117:d=2021032312:TCDC:700 mb:24 hour fcst: -615:10461649:d=2021032312:SPFH:700 mb:24 hour fcst: -616:10490261:d=2021032312:VVEL:700 mb:24 hour fcst: -617:10515779:d=2021032312:DZDT:700 mb:24 hour fcst: -618:10542944:d=2021032312:UGRD:700 mb:24 hour fcst: -619:10563779:d=2021032312:VGRD:700 mb:24 hour fcst: -620:10585396:d=2021032312:ABSV:700 mb:24 hour fcst: -621:10606431:d=2021032312:CLMR:700 mb:24 hour fcst: -622:10613380:d=2021032312:ICMR:700 mb:24 hour fcst: -623:10624522:d=2021032312:RWMR:700 mb:24 hour fcst: -624:10630041:d=2021032312:SNMR:700 mb:24 hour fcst: -625:10644269:d=2021032312:GRLE:700 mb:24 hour fcst: -626:10646518:d=2021032312:O3MR:700 mb:24 hour fcst: -643:10963655:d=2021032312:HGT:750 mb:24 hour fcst: -644:10988060:d=2021032312:TMP:750 mb:24 hour fcst: -645:11007952:d=2021032312:RH:750 mb:24 hour fcst: -646:11028404:d=2021032312:TCDC:750 mb:24 hour fcst: -647:11040737:d=2021032312:SPFH:750 mb:24 hour fcst: -648:11069865:d=2021032312:VVEL:750 mb:24 hour fcst: -649:11095568:d=2021032312:DZDT:750 mb:24 hour fcst: -650:11122739:d=2021032312:UGRD:750 mb:24 hour fcst: -651:11143583:d=2021032312:VGRD:750 mb:24 hour fcst: -652:11165267:d=2021032312:ABSV:750 mb:24 hour fcst: -653:11186378:d=2021032312:CLMR:750 mb:24 hour fcst: -654:11195754:d=2021032312:ICMR:750 mb:24 hour fcst: -655:11207446:d=2021032312:RWMR:750 mb:24 hour fcst: -656:11213376:d=2021032312:SNMR:750 mb:24 hour fcst: -657:11226892:d=2021032312:GRLE:750 mb:24 hour fcst: -658:11228826:d=2021032312:O3MR:750 mb:24 hour fcst: -675:11551203:d=2021032312:HGT:800 mb:24 hour fcst: -676:11575723:d=2021032312:TMP:800 mb:24 hour fcst: -677:11596125:d=2021032312:RH:800 mb:24 hour fcst: -678:11616602:d=2021032312:TCDC:800 mb:24 hour fcst: -679:11629867:d=2021032312:SPFH:800 mb:24 hour fcst: -680:11659291:d=2021032312:VVEL:800 mb:24 hour fcst: -681:11684991:d=2021032312:DZDT:800 mb:24 hour fcst: -682:11712035:d=2021032312:UGRD:800 mb:24 hour fcst: -683:11732967:d=2021032312:VGRD:800 mb:24 hour fcst: -684:11754719:d=2021032312:ABSV:800 mb:24 hour fcst: -685:11775936:d=2021032312:CLMR:800 mb:24 hour fcst: -686:11787285:d=2021032312:ICMR:800 mb:24 hour fcst: -687:11798238:d=2021032312:RWMR:800 mb:24 hour fcst: -688:11805950:d=2021032312:SNMR:800 mb:24 hour fcst: -689:11818869:d=2021032312:GRLE:800 mb:24 hour fcst: -690:11821181:d=2021032312:O3MR:800 mb:24 hour fcst: -707:12149635:d=2021032312:HGT:850 mb:24 hour fcst: -708:12174429:d=2021032312:TMP:850 mb:24 hour fcst: -709:12195117:d=2021032312:RH:850 mb:24 hour fcst: -710:12215407:d=2021032312:TCDC:850 mb:24 hour fcst: -711:12230125:d=2021032312:SPFH:850 mb:24 hour fcst: -712:12259799:d=2021032312:VVEL:850 mb:24 hour fcst: -713:12285079:d=2021032312:DZDT:850 mb:24 hour fcst: -714:12311639:d=2021032312:UGRD:850 mb:24 hour fcst: -715:12332794:d=2021032312:VGRD:850 mb:24 hour fcst: -716:12354612:d=2021032312:ABSV:850 mb:24 hour fcst: -717:12375829:d=2021032312:CLMR:850 mb:24 hour fcst: -718:12389617:d=2021032312:ICMR:850 mb:24 hour fcst: -719:12400240:d=2021032312:RWMR:850 mb:24 hour fcst: -720:12411036:d=2021032312:SNMR:850 mb:24 hour fcst: -721:12422734:d=2021032312:GRLE:850 mb:24 hour fcst: -722:12425193:d=2021032312:O3MR:850 mb:24 hour fcst: -739:12750973:d=2021032312:HGT:900 mb:24 hour fcst: -740:12776009:d=2021032312:TMP:900 mb:24 hour fcst: -741:12796816:d=2021032312:RH:900 mb:24 hour fcst: -742:12816725:d=2021032312:TCDC:900 mb:24 hour fcst: -743:12829704:d=2021032312:SPFH:900 mb:24 hour fcst: -744:12859120:d=2021032312:VVEL:900 mb:24 hour fcst: -745:12883955:d=2021032312:DZDT:900 mb:24 hour fcst: -746:12910017:d=2021032312:UGRD:900 mb:24 hour fcst: -747:12931191:d=2021032312:VGRD:900 mb:24 hour fcst: -748:12953095:d=2021032312:ABSV:900 mb:24 hour fcst: -749:12974291:d=2021032312:CLMR:900 mb:24 hour fcst: -750:12984661:d=2021032312:ICMR:900 mb:24 hour fcst: -751:12993136:d=2021032312:RWMR:900 mb:24 hour fcst: -752:13005480:d=2021032312:SNMR:900 mb:24 hour fcst: -753:13015761:d=2021032312:GRLE:900 mb:24 hour fcst: -754:13018820:d=2021032312:O3MR:900 mb:24 hour fcst: -755:13044922:d=2021032312:HGT:925 mb:24 hour fcst: -756:13070093:d=2021032312:TMP:925 mb:24 hour fcst: -757:13091240:d=2021032312:RH:925 mb:24 hour fcst: -758:13110967:d=2021032312:TCDC:925 mb:24 hour fcst: -759:13122797:d=2021032312:SPFH:925 mb:24 hour fcst: -760:13152536:d=2021032312:VVEL:925 mb:24 hour fcst: -761:13176973:d=2021032312:DZDT:925 mb:24 hour fcst: -762:13202569:d=2021032312:UGRD:925 mb:24 hour fcst: -763:13223998:d=2021032312:VGRD:925 mb:24 hour fcst: -764:13245902:d=2021032312:ABSV:925 mb:24 hour fcst: -765:13267194:d=2021032312:CLMR:925 mb:24 hour fcst: -766:13275897:d=2021032312:ICMR:925 mb:24 hour fcst: -767:13283525:d=2021032312:RWMR:925 mb:24 hour fcst: -768:13296031:d=2021032312:SNMR:925 mb:24 hour fcst: -769:13305511:d=2021032312:GRLE:925 mb:24 hour fcst: -770:13308695:d=2021032312:O3MR:925 mb:24 hour fcst: -771:13334726:d=2021032312:HGT:950 mb:24 hour fcst: -772:13360079:d=2021032312:TMP:950 mb:24 hour fcst: -773:13381198:d=2021032312:RH:950 mb:24 hour fcst: -774:13400726:d=2021032312:TCDC:950 mb:24 hour fcst: -775:13410891:d=2021032312:SPFH:950 mb:24 hour fcst: -776:13438093:d=2021032312:VVEL:950 mb:24 hour fcst: -777:13461916:d=2021032312:DZDT:950 mb:24 hour fcst: -778:13486595:d=2021032312:UGRD:950 mb:24 hour fcst: -779:13508033:d=2021032312:VGRD:950 mb:24 hour fcst: -780:13529933:d=2021032312:ABSV:950 mb:24 hour fcst: -781:13551160:d=2021032312:CLMR:950 mb:24 hour fcst: -782:13557585:d=2021032312:ICMR:950 mb:24 hour fcst: -783:13564760:d=2021032312:RWMR:950 mb:24 hour fcst: -784:13576076:d=2021032312:SNMR:950 mb:24 hour fcst: -785:13584828:d=2021032312:GRLE:950 mb:24 hour fcst: -786:13587713:d=2021032312:O3MR:950 mb:24 hour fcst: -787:13613707:d=2021032312:HINDEX:surface:24 hour fcst: -788:13618800:d=2021032312:HGT:975 mb:24 hour fcst: -789:13644312:d=2021032312:TMP:975 mb:24 hour fcst: -790:13665217:d=2021032312:RH:975 mb:24 hour fcst: -791:13684336:d=2021032312:TCDC:975 mb:24 hour fcst: -792:13692613:d=2021032312:SPFH:975 mb:24 hour fcst: -793:13719679:d=2021032312:VVEL:975 mb:24 hour fcst: -794:13742425:d=2021032312:DZDT:975 mb:24 hour fcst: -795:13765985:d=2021032312:UGRD:975 mb:24 hour fcst: -796:13787379:d=2021032312:VGRD:975 mb:24 hour fcst: -797:13809211:d=2021032312:ABSV:975 mb:24 hour fcst: -798:13830205:d=2021032312:CLMR:975 mb:24 hour fcst: -799:13834907:d=2021032312:ICMR:975 mb:24 hour fcst: -800:13840155:d=2021032312:RWMR:975 mb:24 hour fcst: -801:13851966:d=2021032312:SNMR:975 mb:24 hour fcst: -802:13859638:d=2021032312:GRLE:975 mb:24 hour fcst: -803:13862333:d=2021032312:O3MR:975 mb:24 hour fcst: -804:13888223:d=2021032312:TMP:1000 mb:24 hour fcst: -805:13908998:d=2021032312:RH:1000 mb:24 hour fcst: -806:13927884:d=2021032312:TCDC:1000 mb:24 hour fcst: -807:13932463:d=2021032312:SPFH:1000 mb:24 hour fcst: -808:13959553:d=2021032312:VVEL:1000 mb:24 hour fcst: -809:13980852:d=2021032312:DZDT:1000 mb:24 hour fcst: -810:14002200:d=2021032312:UGRD:1000 mb:24 hour fcst: -811:14022989:d=2021032312:VGRD:1000 mb:24 hour fcst: -812:14044380:d=2021032312:ABSV:1000 mb:24 hour fcst: -813:14064912:d=2021032312:CLMR:1000 mb:24 hour fcst: -814:14066827:d=2021032312:ICMR:1000 mb:24 hour fcst: -815:14069347:d=2021032312:RWMR:1000 mb:24 hour fcst: -816:14079056:d=2021032312:SNMR:1000 mb:24 hour fcst: -817:14082375:d=2021032312:GRLE:1000 mb:24 hour fcst: -818:14083410:d=2021032312:O3MR:1000 mb:24 hour fcst: -819:14109342:d=2021032312:MSLET:mean sea level:24 hour fcst: -820:14135950:d=2021032312:HGT:1000 mb:24 hour fcst: -821:14161667:d=2021032312:SDEN:surface:24 hour fcst: -822:14172627:d=2021032312:REFD:4000 m above ground:24 hour fcst: -823:14185655:d=2021032312:REFD:1000 m above ground:24 hour fcst: -824:14209858:d=2021032312:PRES:surface:24 hour fcst: -825:14235507:d=2021032312:HGT:surface:24 hour fcst: -826:14251577:d=2021032312:TMP:surface:24 hour fcst: -827:14268207:d=2021032312:TSOIL:0-0.1 m below ground:24 hour fcst: -828:14280138:d=2021032312:SOILW:0-0.1 m below ground:24 hour fcst: -829:14289049:d=2021032312:SOILL:0-0.1 m below ground:24 hour fcst: -830:14297846:d=2021032312:TSOIL:0.1-0.4 m below ground:24 hour fcst: -831:14313306:d=2021032312:SOILW:0.1-0.4 m below ground:24 hour fcst: -832:14322323:d=2021032312:SOILL:0.1-0.4 m below ground:24 hour fcst: -833:14331217:d=2021032312:TSOIL:0.4-1 m below ground:24 hour fcst: -834:14345656:d=2021032312:SOILW:0.4-1 m below ground:24 hour fcst: -835:14354789:d=2021032312:SOILL:0.4-1 m below ground:24 hour fcst: -836:14363693:d=2021032312:TSOIL:1-2 m below ground:24 hour fcst: -837:14377955:d=2021032312:SOILW:1-2 m below ground:24 hour fcst: -838:14387015:d=2021032312:SOILL:1-2 m below ground:24 hour fcst: -839:14396091:d=2021032312:CNWAT:surface:24 hour fcst: -840:14401676:d=2021032312:WEASD:surface:24 hour fcst: -841:14414473:d=2021032312:SNOD:surface:24 hour fcst: -842:14428819:d=2021032312:PEVPR:surface:24 hour fcst: -843:14444469:d=2021032312:ICETK:surface:24 hour fcst: -845:14455904:d=2021032312:TMP:2 m above ground:24 hour fcst: -846:14476664:d=2021032312:SPFH:2 m above ground:24 hour fcst: -847:14505531:d=2021032312:DPT:2 m above ground:24 hour fcst: -848:14526770:d=2021032312:RH:2 m above ground:24 hour fcst: -849:14545307:d=2021032312:APTMP:2 m above ground:24 hour fcst: -850:14559685:d=2021032312:TMAX:2 m above ground:18-24 hour max fcst: -851:14580381:d=2021032312:TMIN:2 m above ground:18-24 hour min fcst: -852:14601038:d=2021032312:UGRD:10 m above ground:24 hour fcst: -853:14621553:d=2021032312:VGRD:10 m above ground:24 hour fcst: -854:14642594:d=2021032312:ICEG:10 m above mean sea level:24 hour fcst: -855:14643241:d=2021032312:CPOFP:surface:24 hour fcst: -856:14663348:d=2021032312:CPRAT:surface:24 hour fcst: -857:14684649:d=2021032312:PRATE:surface:24 hour fcst: -858:14711666:d=2021032312:CPRAT:surface:18-24 hour ave fcst: -859:14735720:d=2021032312:PRATE:surface:18-24 hour ave fcst: -860:14760778:d=2021032312:APCP:surface:18-24 hour acc fcst: -861:14770996:d=2021032312:APCP:surface:0-1 day acc fcst: -862:14783965:d=2021032312:ACPCP:surface:18-24 hour acc fcst: -863:14792304:d=2021032312:ACPCP:surface:0-1 day acc fcst: -866:14818649:d=2021032312:FROZR:surface:0-1 day acc fcst: -867:14824950:d=2021032312:FRZR:surface:0-1 day acc fcst: -868:14828053:d=2021032312:TSNOWP:surface:0-1 day acc fcst: -869:14835599:d=2021032312:WATR:surface:18-24 hour acc fcst: -870:14848010:d=2021032312:FROZR:surface:18-24 hour acc fcst: -871:14851970:d=2021032312:FRZR:surface:18-24 hour acc fcst: -872:14853743:d=2021032312:TSNOWP:surface:18-24 hour acc fcst: -873:14861669:d=2021032312:CSNOW:surface:24 hour fcst: -874:14863449:d=2021032312:CICEP:surface:24 hour fcst: -875:14863628:d=2021032312:CFRZR:surface:24 hour fcst: -876:14863848:d=2021032312:CRAIN:surface:24 hour fcst: -877:14867383:d=2021032312:LHTFL:surface:18-24 hour ave fcst: -878:14889027:d=2021032312:SHTFL:surface:18-24 hour ave fcst: -879:14909688:d=2021032312:GFLUX:surface:18-24 hour ave fcst: -880:14921746:d=2021032312:UFLX:surface:18-24 hour ave fcst: -881:14937989:d=2021032312:VFLX:surface:18-24 hour ave fcst: -882:14954352:d=2021032312:SFCR:surface:24 hour fcst: -883:14972818:d=2021032312:FRICV:surface:24 hour fcst: -884:14994522:d=2021032312:U-GWD:surface:18-24 hour ave fcst: -885:15009281:d=2021032312:V-GWD:surface:18-24 hour ave fcst: -886:15024064:d=2021032312:VEG:surface:24 hour fcst: -887:15032088:d=2021032312:SOTYP:surface:24 hour fcst: -888:15043137:d=2021032312:WILT:surface:24 hour fcst: -889:15053047:d=2021032312:FLDCP:surface:24 hour fcst: -890:15063253:d=2021032312:SUNSD:surface:24 hour fcst: -891:15080351:d=2021032312:LFTX:surface:24 hour fcst: -892:15094145:d=2021032312:CAPE:surface:24 hour fcst: -893:15109237:d=2021032312:CIN:surface:24 hour fcst: -894:15122223:d=2021032312:PWAT:entire atmosphere (considered as a single layer):24 hour fcst: -895:15149124:d=2021032312:CWAT:entire atmosphere (considered as a single layer):24 hour fcst: -896:15175589:d=2021032312:RH:entire atmosphere (considered as a single layer):24 hour fcst: -897:15190732:d=2021032312:TOZNE:entire atmosphere (considered as a single layer):24 hour fcst: -898:15205083:d=2021032312:LCDC:low cloud layer:24 hour fcst: -899:15222792:d=2021032312:LCDC:low cloud layer:18-24 hour ave fcst: -900:15242746:d=2021032312:MCDC:middle cloud layer:24 hour fcst: -901:15258582:d=2021032312:MCDC:middle cloud layer:18-24 hour ave fcst: -902:15275850:d=2021032312:HCDC:high cloud layer:24 hour fcst: -903:15290909:d=2021032312:HCDC:high cloud layer:18-24 hour ave fcst: -904:15309008:d=2021032312:TCDC:entire atmosphere (considered as a single layer):24 hour fcst: -905:15326511:d=2021032312:TCDC:entire atmosphere (considered as a single layer):18-24 hour ave fcst: -906:15346594:d=2021032312:HGT:cloud ceiling:24 hour fcst: -907:15379112:d=2021032312:PRES:convective cloud bottom level:24 hour fcst: -908:15392766:d=2021032312:PRES:low cloud bottom level:18-24 hour ave fcst: -909:15415606:d=2021032312:PRES:middle cloud bottom level:18-24 hour ave fcst: -910:15433333:d=2021032312:PRES:high cloud bottom level:18-24 hour ave fcst: -912:15472788:d=2021032312:PRES:convective cloud top level:24 hour fcst: -913:15488123:d=2021032312:PRES:low cloud top level:18-24 hour ave fcst: -914:15511137:d=2021032312:PRES:middle cloud top level:18-24 hour ave fcst: -915:15528848:d=2021032312:PRES:high cloud top level:18-24 hour ave fcst: -916:15548070:d=2021032312:TMP:low cloud top level:18-24 hour ave fcst: -917:15565226:d=2021032312:TMP:middle cloud top level:18-24 hour ave fcst: -918:15579272:d=2021032312:TMP:high cloud top level:18-24 hour ave fcst: -919:15595233:d=2021032312:TCDC:convective cloud layer:24 hour fcst: -920:15616051:d=2021032312:TCDC:boundary layer cloud layer:18-24 hour ave fcst: -921:15633966:d=2021032312:CWORK:entire atmosphere (considered as a single layer):18-24 hour ave fcst: -922:15647665:d=2021032312:DSWRF:surface:18-24 hour ave fcst: -925:15706482:d=2021032312:DLWRF:surface:18-24 hour ave fcst: -926:15727752:d=2021032312:USWRF:surface:18-24 hour ave fcst: -927:15747536:d=2021032312:ULWRF:surface:18-24 hour ave fcst: -928:15766290:d=2021032312:USWRF:top of atmosphere:18-24 hour ave fcst: -929:15788879:d=2021032312:ULWRF:top of atmosphere:18-24 hour ave fcst: -930:15809642:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Total Aerosol:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -931:15837184:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Dust Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -932:15855325:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Sea Salt Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -933:15882744:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Sulphate Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -934:15908405:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Particulate Organic Matter Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -935:15933974:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Black Carbon Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -936:15956253:d=2021032312:HLCY:3000-0 m above ground:24 hour fcst: -937:15972145:d=2021032312:HLCY:1000-0 m above ground:24 hour fcst: -938:15994425:d=2021032312:USTM:6000-0 m above ground:24 hour fcst: -939:16017651:d=2021032312:VSTM:6000-0 m above ground:24 hour fcst: -940:16040470:d=2021032312:PRES:tropopause:24 hour fcst: -941:16070262:d=2021032312:ICAHT:tropopause:24 hour fcst: -942:16101038:d=2021032312:HGT:tropopause:24 hour fcst: -943:16129233:d=2021032312:TMP:tropopause:24 hour fcst: -944:16150799:d=2021032312:UGRD:tropopause:24 hour fcst: -945:16166786:d=2021032312:VGRD:tropopause:24 hour fcst: -946:16182630:d=2021032312:VWSH:tropopause:24 hour fcst: -947:16201304:d=2021032312:PRES:max wind:24 hour fcst: -948:16235590:d=2021032312:ICAHT:max wind:24 hour fcst: -949:16270614:d=2021032312:HGT:max wind:24 hour fcst: -950:16305811:d=2021032312:UGRD:max wind:24 hour fcst: -951:16322297:d=2021032312:VGRD:max wind:24 hour fcst: -952:16339640:d=2021032312:TMP:max wind:24 hour fcst: -953:16366375:d=2021032312:UGRD:20 m above ground:24 hour fcst: -954:16386899:d=2021032312:VGRD:20 m above ground:24 hour fcst: -955:16407997:d=2021032312:UGRD:30 m above ground:24 hour fcst: -956:16428883:d=2021032312:VGRD:30 m above ground:24 hour fcst: -957:16450176:d=2021032312:UGRD:40 m above ground:24 hour fcst: -958:16471162:d=2021032312:VGRD:40 m above ground:24 hour fcst: -959:16492591:d=2021032312:UGRD:50 m above ground:24 hour fcst: -960:16513685:d=2021032312:VGRD:50 m above ground:24 hour fcst: -961:16535212:d=2021032312:TMP:80 m above ground:24 hour fcst: -962:16555991:d=2021032312:SPFH:80 m above ground:24 hour fcst: -963:16584638:d=2021032312:PRES:80 m above ground:24 hour fcst: -964:16610293:d=2021032312:UGRD:80 m above ground:24 hour fcst: -965:16631550:d=2021032312:VGRD:80 m above ground:24 hour fcst: -966:16653256:d=2021032312:TMP:100 m above ground:24 hour fcst: -967:16674024:d=2021032312:UGRD:100 m above ground:24 hour fcst: -968:16695351:d=2021032312:VGRD:100 m above ground:24 hour fcst: -981:16960917:d=2021032312:TMP:1829 m above mean sea level:24 hour fcst: -982:16982571:d=2021032312:UGRD:1829 m above mean sea level:24 hour fcst: -983:17004788:d=2021032312:VGRD:1829 m above mean sea level:24 hour fcst: -984:17027631:d=2021032312:TMP:2743 m above mean sea level:24 hour fcst: -985:17049101:d=2021032312:UGRD:2743 m above mean sea level:24 hour fcst: -986:17071775:d=2021032312:VGRD:2743 m above mean sea level:24 hour fcst: -987:17095096:d=2021032312:TMP:3658 m above mean sea level:24 hour fcst: -988:17116393:d=2021032312:UGRD:3658 m above mean sea level:24 hour fcst: -989:17139586:d=2021032312:VGRD:3658 m above mean sea level:24 hour fcst: -993:17231902:d=2021032312:HGT:0C isotherm:24 hour fcst: -994:17253627:d=2021032312:RH:0C isotherm:24 hour fcst: -995:17272906:d=2021032312:HGT:highest tropospheric freezing level:24 hour fcst: -996:17294266:d=2021032312:RH:highest tropospheric freezing level:24 hour fcst: -997:17313505:d=2021032312:TMP:30-0 mb above ground:24 hour fcst: -998:17334320:d=2021032312:RH:30-0 mb above ground:24 hour fcst: -1000:17373720:d=2021032312:SPFH:30-0 mb above ground:24 hour fcst: -1001:17402485:d=2021032312:UGRD:30-0 mb above ground:24 hour fcst: -1002:17423605:d=2021032312:VGRD:30-0 mb above ground:24 hour fcst: -1030:18048648:d=2021032312:4LFTX:surface:24 hour fcst: -1031:18062145:d=2021032312:CAPE:180-0 mb above ground:24 hour fcst: -1032:18076145:d=2021032312:CIN:180-0 mb above ground:24 hour fcst: -1033:18090257:d=2021032312:HPBL:surface:24 hour fcst: -1034:18121411:d=2021032312:RH:0.33-1 sigma layer:24 hour fcst: -1035:18139787:d=2021032312:RH:0.44-1 sigma layer:24 hour fcst: -1036:18157789:d=2021032312:RH:0.72-0.94 sigma layer:24 hour fcst: -1037:18176266:d=2021032312:RH:0.44-0.72 sigma layer:24 hour fcst: -1038:18195574:d=2021032312:TMP:0.995 sigma level:24 hour fcst: -1039:18216412:d=2021032312:POT:0.995 sigma level:24 hour fcst: -1040:18237233:d=2021032312:RH:0.995 sigma level:24 hour fcst: -1041:18255030:d=2021032312:UGRD:0.995 sigma level:24 hour fcst: -1042:18276189:d=2021032312:VGRD:0.995 sigma level:24 hour fcst: -1043:18297706:d=2021032312:VVEL:0.995 sigma level:24 hour fcst: -1044:18318589:d=2021032312:CAPE:90-0 mb above ground:24 hour fcst: -1045:18331400:d=2021032312:CIN:90-0 mb above ground:24 hour fcst: -1046:18345766:d=2021032312:CAPE:255-0 mb above ground:24 hour fcst: -1047:18358884:d=2021032312:CIN:255-0 mb above ground:24 hour fcst: -1048:18370243:d=2021032312:PLPL:255-0 mb above ground:24 hour fcst: -1049:18393964:d=2021032312:LAND:surface:24 hour fcst: -1050:18395905:d=2021032312:ICEC:surface:24 hour fcst: -1051:18399070:d=2021032312:ALBDO:surface:18-24 hour ave fcst: -1052:18413832:d=2021032312:ICETMP:surface:24 hour fcst: -1094:19022015:d=2021032312:UGRD:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -1095:19032738:d=2021032312:VGRD:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -1096:19043451:d=2021032312:TMP:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -1097:19057826:d=2021032312:HGT:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -1098:19075199:d=2021032312:PRES:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -1099:19093633:d=2021032312:VWSH:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -1100:19105590:d=2021032312:UGRD:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -1101:19116465:d=2021032312:VGRD:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -1102:19131275:d=2021032312:TMP:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -1103:19145691:d=2021032312:HGT:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -1104:19163193:d=2021032312:PRES:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -1105:19180166:d=2021032312:VWSH:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -+ exglobal_atmos_products.sh[45]true -+ exglobal_atmos_products.sh[46]export err=0 -+ exglobal_atmos_products.sh[46]err=0 -+ exglobal_atmos_products.sh[47][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[52][[ 2 -eq 2 ]] -+ exglobal_atmos_products.sh[54]wgrib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.master.grb2f024 -+ exglobal_atmos_products.sh[54]wgrib2 -i -grib tmpfileb_f024 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.master.grb2f024 -+ exglobal_atmos_products.sh[54]grep -F -f /work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.fFFF.paramlist.b.txt -95:1805811:d=2021032312:HGT:1 mb:24 hour fcst: -96:1824058:d=2021032312:TMP:1 mb:24 hour fcst: -97:1839923:d=2021032312:RH:1 mb:24 hour fcst: -101:1921096:d=2021032312:UGRD:1 mb:24 hour fcst: -102:1939697:d=2021032312:VGRD:1 mb:24 hour fcst: -103:1957398:d=2021032312:ABSV:1 mb:24 hour fcst: -104:1973752:d=2021032312:O3MR:1 mb:24 hour fcst: -105:1994820:d=2021032312:HGT:2 mb:24 hour fcst: -106:2015132:d=2021032312:TMP:2 mb:24 hour fcst: -107:2031736:d=2021032312:RH:2 mb:24 hour fcst: -111:2114485:d=2021032312:UGRD:2 mb:24 hour fcst: -112:2126070:d=2021032312:VGRD:2 mb:24 hour fcst: -113:2144416:d=2021032312:ABSV:2 mb:24 hour fcst: -114:2161499:d=2021032312:O3MR:2 mb:24 hour fcst: -115:2184317:d=2021032312:HGT:3 mb:24 hour fcst: -116:2202153:d=2021032312:TMP:3 mb:24 hour fcst: -117:2218626:d=2021032312:RH:3 mb:24 hour fcst: -121:2302386:d=2021032312:UGRD:3 mb:24 hour fcst: -122:2313769:d=2021032312:VGRD:3 mb:24 hour fcst: -123:2331906:d=2021032312:ABSV:3 mb:24 hour fcst: -124:2348920:d=2021032312:O3MR:3 mb:24 hour fcst: -125:2372293:d=2021032312:HGT:5 mb:24 hour fcst: -126:2389968:d=2021032312:TMP:5 mb:24 hour fcst: -127:2406597:d=2021032312:RH:5 mb:24 hour fcst: -131:2491636:d=2021032312:UGRD:5 mb:24 hour fcst: -132:2503228:d=2021032312:VGRD:5 mb:24 hour fcst: -133:2521798:d=2021032312:ABSV:5 mb:24 hour fcst: -134:2539419:d=2021032312:O3MR:5 mb:24 hour fcst: -135:2562079:d=2021032312:HGT:7 mb:24 hour fcst: -136:2579838:d=2021032312:TMP:7 mb:24 hour fcst: -137:2596027:d=2021032312:RH:7 mb:24 hour fcst: -141:2671752:d=2021032312:UGRD:7 mb:24 hour fcst: -142:2683166:d=2021032312:VGRD:7 mb:24 hour fcst: -143:2701858:d=2021032312:ABSV:7 mb:24 hour fcst: -144:2719667:d=2021032312:O3MR:7 mb:24 hour fcst: -214:3947743:d=2021032312:TCDC:70 mb:24 hour fcst: -221:4087576:d=2021032312:CLMR:70 mb:24 hour fcst: -222:4087755:d=2021032312:ICMR:70 mb:24 hour fcst: -223:4087934:d=2021032312:RWMR:70 mb:24 hour fcst: -224:4088113:d=2021032312:SNMR:70 mb:24 hour fcst: -225:4088292:d=2021032312:GRLE:70 mb:24 hour fcst: -243:4321033:d=2021032312:HGT:125 mb:24 hour fcst: -244:4341121:d=2021032312:TMP:125 mb:24 hour fcst: -245:4358162:d=2021032312:RH:125 mb:24 hour fcst: -246:4368807:d=2021032312:TCDC:125 mb:24 hour fcst: -248:4397253:d=2021032312:VVEL:125 mb:24 hour fcst: -249:4423697:d=2021032312:DZDT:125 mb:24 hour fcst: -250:4448851:d=2021032312:UGRD:125 mb:24 hour fcst: -251:4468386:d=2021032312:VGRD:125 mb:24 hour fcst: -252:4488038:d=2021032312:ABSV:125 mb:24 hour fcst: -253:4506365:d=2021032312:CLMR:125 mb:24 hour fcst: -254:4506544:d=2021032312:ICMR:125 mb:24 hour fcst: -255:4510938:d=2021032312:RWMR:125 mb:24 hour fcst: -256:4511117:d=2021032312:SNMR:125 mb:24 hour fcst: -257:4514598:d=2021032312:GRLE:125 mb:24 hour fcst: -275:4758433:d=2021032312:HGT:175 mb:24 hour fcst: -276:4779079:d=2021032312:TMP:175 mb:24 hour fcst: -277:4796607:d=2021032312:RH:175 mb:24 hour fcst: -278:4809762:d=2021032312:TCDC:175 mb:24 hour fcst: -280:4841063:d=2021032312:VVEL:175 mb:24 hour fcst: -281:4869548:d=2021032312:DZDT:175 mb:24 hour fcst: -282:4895682:d=2021032312:UGRD:175 mb:24 hour fcst: -283:4908643:d=2021032312:VGRD:175 mb:24 hour fcst: -284:4929413:d=2021032312:ABSV:175 mb:24 hour fcst: -285:4948812:d=2021032312:CLMR:175 mb:24 hour fcst: -286:4948991:d=2021032312:ICMR:175 mb:24 hour fcst: -287:4955509:d=2021032312:RWMR:175 mb:24 hour fcst: -288:4955688:d=2021032312:SNMR:175 mb:24 hour fcst: -289:4959397:d=2021032312:GRLE:175 mb:24 hour fcst: -307:5210538:d=2021032312:HGT:225 mb:24 hour fcst: -308:5231994:d=2021032312:TMP:225 mb:24 hour fcst: -309:5250433:d=2021032312:RH:225 mb:24 hour fcst: -310:5267061:d=2021032312:TCDC:225 mb:24 hour fcst: -312:5303753:d=2021032312:VVEL:225 mb:24 hour fcst: -313:5331318:d=2021032312:DZDT:225 mb:24 hour fcst: -314:5358118:d=2021032312:UGRD:225 mb:24 hour fcst: -315:5372098:d=2021032312:VGRD:225 mb:24 hour fcst: -316:5386470:d=2021032312:ABSV:225 mb:24 hour fcst: -317:5407187:d=2021032312:CLMR:225 mb:24 hour fcst: -318:5407366:d=2021032312:ICMR:225 mb:24 hour fcst: -319:5418838:d=2021032312:RWMR:225 mb:24 hour fcst: -320:5419017:d=2021032312:SNMR:225 mb:24 hour fcst: -321:5425942:d=2021032312:GRLE:225 mb:24 hour fcst: -339:5699249:d=2021032312:HGT:275 mb:24 hour fcst: -340:5721428:d=2021032312:TMP:275 mb:24 hour fcst: -341:5740048:d=2021032312:RH:275 mb:24 hour fcst: -342:5758548:d=2021032312:TCDC:275 mb:24 hour fcst: -344:5797948:d=2021032312:VVEL:275 mb:24 hour fcst: -345:5821238:d=2021032312:DZDT:275 mb:24 hour fcst: -346:5848833:d=2021032312:UGRD:275 mb:24 hour fcst: -347:5863161:d=2021032312:VGRD:275 mb:24 hour fcst: -348:5878318:d=2021032312:ABSV:275 mb:24 hour fcst: -349:5900376:d=2021032312:CLMR:275 mb:24 hour fcst: -350:5900555:d=2021032312:ICMR:275 mb:24 hour fcst: -351:5915080:d=2021032312:RWMR:275 mb:24 hour fcst: -352:5915259:d=2021032312:SNMR:275 mb:24 hour fcst: -353:5925255:d=2021032312:GRLE:275 mb:24 hour fcst: -371:6209111:d=2021032312:HGT:325 mb:24 hour fcst: -372:6231300:d=2021032312:TMP:325 mb:24 hour fcst: -373:6249666:d=2021032312:RH:325 mb:24 hour fcst: -374:6268689:d=2021032312:TCDC:325 mb:24 hour fcst: -376:6308875:d=2021032312:VVEL:325 mb:24 hour fcst: -377:6333133:d=2021032312:DZDT:325 mb:24 hour fcst: -378:6361377:d=2021032312:UGRD:325 mb:24 hour fcst: -379:6376016:d=2021032312:VGRD:325 mb:24 hour fcst: -380:6391304:d=2021032312:ABSV:325 mb:24 hour fcst: -381:6413626:d=2021032312:CLMR:325 mb:24 hour fcst: -382:6413805:d=2021032312:ICMR:325 mb:24 hour fcst: -383:6430308:d=2021032312:RWMR:325 mb:24 hour fcst: -384:6430487:d=2021032312:SNMR:325 mb:24 hour fcst: -385:6442274:d=2021032312:GRLE:325 mb:24 hour fcst: -403:6731274:d=2021032312:HGT:375 mb:24 hour fcst: -404:6753166:d=2021032312:TMP:375 mb:24 hour fcst: -405:6771466:d=2021032312:RH:375 mb:24 hour fcst: -406:6790630:d=2021032312:TCDC:375 mb:24 hour fcst: -408:6830861:d=2021032312:VVEL:375 mb:24 hour fcst: -409:6855688:d=2021032312:DZDT:375 mb:24 hour fcst: -410:6883831:d=2021032312:UGRD:375 mb:24 hour fcst: -411:6898278:d=2021032312:VGRD:375 mb:24 hour fcst: -412:6921216:d=2021032312:ABSV:375 mb:24 hour fcst: -413:6943432:d=2021032312:CLMR:375 mb:24 hour fcst: -414:6943611:d=2021032312:ICMR:375 mb:24 hour fcst: -415:6960539:d=2021032312:RWMR:375 mb:24 hour fcst: -416:6960718:d=2021032312:SNMR:375 mb:24 hour fcst: -417:6973751:d=2021032312:GRLE:375 mb:24 hour fcst: -435:7268558:d=2021032312:HGT:425 mb:24 hour fcst: -436:7292608:d=2021032312:TMP:425 mb:24 hour fcst: -437:7311137:d=2021032312:RH:425 mb:24 hour fcst: -438:7330506:d=2021032312:TCDC:425 mb:24 hour fcst: -440:7371921:d=2021032312:VVEL:425 mb:24 hour fcst: -441:7397000:d=2021032312:DZDT:425 mb:24 hour fcst: -442:7425196:d=2021032312:UGRD:425 mb:24 hour fcst: -443:7447029:d=2021032312:VGRD:425 mb:24 hour fcst: -444:7469479:d=2021032312:ABSV:425 mb:24 hour fcst: -445:7491090:d=2021032312:CLMR:425 mb:24 hour fcst: -446:7491746:d=2021032312:ICMR:425 mb:24 hour fcst: -447:7508548:d=2021032312:RWMR:425 mb:24 hour fcst: -448:7508902:d=2021032312:SNMR:425 mb:24 hour fcst: -449:7523254:d=2021032312:GRLE:425 mb:24 hour fcst: -467:7829463:d=2021032312:HGT:475 mb:24 hour fcst: -468:7853311:d=2021032312:TMP:475 mb:24 hour fcst: -469:7871785:d=2021032312:RH:475 mb:24 hour fcst: -470:7891283:d=2021032312:TCDC:475 mb:24 hour fcst: -472:7931383:d=2021032312:VVEL:475 mb:24 hour fcst: -473:7956362:d=2021032312:DZDT:475 mb:24 hour fcst: -474:7984367:d=2021032312:UGRD:475 mb:24 hour fcst: -475:8005969:d=2021032312:VGRD:475 mb:24 hour fcst: -476:8028125:d=2021032312:ABSV:475 mb:24 hour fcst: -477:8049435:d=2021032312:CLMR:475 mb:24 hour fcst: -478:8050809:d=2021032312:ICMR:475 mb:24 hour fcst: -479:8066993:d=2021032312:RWMR:475 mb:24 hour fcst: -480:8067690:d=2021032312:SNMR:475 mb:24 hour fcst: -481:8081645:d=2021032312:GRLE:475 mb:24 hour fcst: -499:8386292:d=2021032312:HGT:525 mb:24 hour fcst: -500:8409953:d=2021032312:TMP:525 mb:24 hour fcst: -501:8428583:d=2021032312:RH:525 mb:24 hour fcst: -502:8448298:d=2021032312:TCDC:525 mb:24 hour fcst: -504:8489169:d=2021032312:VVEL:525 mb:24 hour fcst: -505:8514245:d=2021032312:DZDT:525 mb:24 hour fcst: -506:8541867:d=2021032312:UGRD:525 mb:24 hour fcst: -507:8562998:d=2021032312:VGRD:525 mb:24 hour fcst: -508:8584900:d=2021032312:ABSV:525 mb:24 hour fcst: -509:8605931:d=2021032312:CLMR:525 mb:24 hour fcst: -510:8607992:d=2021032312:ICMR:525 mb:24 hour fcst: -511:8624980:d=2021032312:RWMR:525 mb:24 hour fcst: -512:8626145:d=2021032312:SNMR:525 mb:24 hour fcst: -513:8640498:d=2021032312:GRLE:525 mb:24 hour fcst: -531:8946286:d=2021032312:HGT:575 mb:24 hour fcst: -532:8969817:d=2021032312:TMP:575 mb:24 hour fcst: -533:8988705:d=2021032312:RH:575 mb:24 hour fcst: -534:9008640:d=2021032312:TCDC:575 mb:24 hour fcst: -536:9050748:d=2021032312:VVEL:575 mb:24 hour fcst: -537:9075877:d=2021032312:DZDT:575 mb:24 hour fcst: -538:9103287:d=2021032312:UGRD:575 mb:24 hour fcst: -539:9124215:d=2021032312:VGRD:575 mb:24 hour fcst: -540:9145864:d=2021032312:ABSV:575 mb:24 hour fcst: -541:9166735:d=2021032312:CLMR:575 mb:24 hour fcst: -542:9170871:d=2021032312:ICMR:575 mb:24 hour fcst: -543:9186808:d=2021032312:RWMR:575 mb:24 hour fcst: -544:9189266:d=2021032312:SNMR:575 mb:24 hour fcst: -545:9203808:d=2021032312:GRLE:575 mb:24 hour fcst: -563:9519574:d=2021032312:HGT:625 mb:24 hour fcst: -564:9543743:d=2021032312:TMP:625 mb:24 hour fcst: -565:9562873:d=2021032312:RH:625 mb:24 hour fcst: -566:9583048:d=2021032312:TCDC:625 mb:24 hour fcst: -568:9623213:d=2021032312:VVEL:625 mb:24 hour fcst: -569:9648699:d=2021032312:DZDT:625 mb:24 hour fcst: -570:9676115:d=2021032312:UGRD:625 mb:24 hour fcst: -571:9697144:d=2021032312:VGRD:625 mb:24 hour fcst: -572:9718592:d=2021032312:ABSV:625 mb:24 hour fcst: -573:9739402:d=2021032312:CLMR:625 mb:24 hour fcst: -574:9744460:d=2021032312:ICMR:625 mb:24 hour fcst: -575:9757661:d=2021032312:RWMR:625 mb:24 hour fcst: -576:9761915:d=2021032312:SNMR:625 mb:24 hour fcst: -577:9777391:d=2021032312:GRLE:625 mb:24 hour fcst: -595:10095633:d=2021032312:HGT:675 mb:24 hour fcst: -596:10119940:d=2021032312:TMP:675 mb:24 hour fcst: -597:10139388:d=2021032312:RH:675 mb:24 hour fcst: -598:10159680:d=2021032312:TCDC:675 mb:24 hour fcst: -600:10202122:d=2021032312:VVEL:675 mb:24 hour fcst: -601:10227764:d=2021032312:DZDT:675 mb:24 hour fcst: -602:10255176:d=2021032312:UGRD:675 mb:24 hour fcst: -603:10276264:d=2021032312:VGRD:675 mb:24 hour fcst: -604:10297943:d=2021032312:ABSV:675 mb:24 hour fcst: -605:10318983:d=2021032312:CLMR:675 mb:24 hour fcst: -606:10324675:d=2021032312:ICMR:675 mb:24 hour fcst: -607:10336570:d=2021032312:RWMR:675 mb:24 hour fcst: -608:10341500:d=2021032312:SNMR:675 mb:24 hour fcst: -609:10356124:d=2021032312:GRLE:675 mb:24 hour fcst: -627:10673893:d=2021032312:HGT:725 mb:24 hour fcst: -628:10698061:d=2021032312:TMP:725 mb:24 hour fcst: -629:10717781:d=2021032312:RH:725 mb:24 hour fcst: -630:10738123:d=2021032312:TCDC:725 mb:24 hour fcst: -632:10778746:d=2021032312:VVEL:725 mb:24 hour fcst: -633:10804434:d=2021032312:DZDT:725 mb:24 hour fcst: -634:10831699:d=2021032312:UGRD:725 mb:24 hour fcst: -635:10852522:d=2021032312:VGRD:725 mb:24 hour fcst: -636:10874152:d=2021032312:ABSV:725 mb:24 hour fcst: -637:10895037:d=2021032312:CLMR:725 mb:24 hour fcst: -638:10902973:d=2021032312:ICMR:725 mb:24 hour fcst: -639:10914857:d=2021032312:RWMR:725 mb:24 hour fcst: -640:10920589:d=2021032312:SNMR:725 mb:24 hour fcst: -641:10934404:d=2021032312:GRLE:725 mb:24 hour fcst: -659:11256089:d=2021032312:HGT:775 mb:24 hour fcst: -660:11280561:d=2021032312:TMP:775 mb:24 hour fcst: -661:11300654:d=2021032312:RH:775 mb:24 hour fcst: -662:11321187:d=2021032312:TCDC:775 mb:24 hour fcst: -664:11363453:d=2021032312:VVEL:775 mb:24 hour fcst: -665:11389143:d=2021032312:DZDT:775 mb:24 hour fcst: -666:11416085:d=2021032312:UGRD:775 mb:24 hour fcst: -667:11436945:d=2021032312:VGRD:775 mb:24 hour fcst: -668:11458630:d=2021032312:ABSV:775 mb:24 hour fcst: -669:11479589:d=2021032312:CLMR:775 mb:24 hour fcst: -670:11490099:d=2021032312:ICMR:775 mb:24 hour fcst: -671:11501578:d=2021032312:RWMR:775 mb:24 hour fcst: -672:11508486:d=2021032312:SNMR:775 mb:24 hour fcst: -673:11521630:d=2021032312:GRLE:775 mb:24 hour fcst: -691:11848192:d=2021032312:HGT:825 mb:24 hour fcst: -692:11872677:d=2021032312:TMP:825 mb:24 hour fcst: -693:11893496:d=2021032312:RH:825 mb:24 hour fcst: -694:11914142:d=2021032312:TCDC:825 mb:24 hour fcst: -696:11957854:d=2021032312:VVEL:825 mb:24 hour fcst: -697:11983437:d=2021032312:DZDT:825 mb:24 hour fcst: -698:12010186:d=2021032312:UGRD:825 mb:24 hour fcst: -699:12031344:d=2021032312:VGRD:825 mb:24 hour fcst: -700:12053157:d=2021032312:ABSV:825 mb:24 hour fcst: -701:12074187:d=2021032312:CLMR:825 mb:24 hour fcst: -702:12087572:d=2021032312:ICMR:825 mb:24 hour fcst: -703:12097994:d=2021032312:RWMR:825 mb:24 hour fcst: -704:12107653:d=2021032312:SNMR:825 mb:24 hour fcst: -705:12119949:d=2021032312:GRLE:825 mb:24 hour fcst: -723:12451770:d=2021032312:HGT:875 mb:24 hour fcst: -724:12476701:d=2021032312:TMP:875 mb:24 hour fcst: -725:12497466:d=2021032312:RH:875 mb:24 hour fcst: -726:12517467:d=2021032312:TCDC:875 mb:24 hour fcst: -728:12560980:d=2021032312:VVEL:875 mb:24 hour fcst: -729:12586101:d=2021032312:DZDT:875 mb:24 hour fcst: -730:12612494:d=2021032312:UGRD:875 mb:24 hour fcst: -731:12633722:d=2021032312:VGRD:875 mb:24 hour fcst: -732:12655522:d=2021032312:ABSV:875 mb:24 hour fcst: -733:12676731:d=2021032312:CLMR:875 mb:24 hour fcst: -734:12689334:d=2021032312:ICMR:875 mb:24 hour fcst: -735:12699233:d=2021032312:RWMR:875 mb:24 hour fcst: -736:12710993:d=2021032312:SNMR:875 mb:24 hour fcst: -737:12722017:d=2021032312:GRLE:875 mb:24 hour fcst: -829:14289049:d=2021032312:SOILL:0-0.1 m below ground:24 hour fcst: -832:14322323:d=2021032312:SOILL:0.1-0.4 m below ground:24 hour fcst: -835:14354789:d=2021032312:SOILL:0.4-1 m below ground:24 hour fcst: -838:14387015:d=2021032312:SOILL:1-2 m below ground:24 hour fcst: -839:14396091:d=2021032312:CNWAT:surface:24 hour fcst: -843:14444469:d=2021032312:ICETK:surface:24 hour fcst: -923:15669944:d=2021032312:DUVB:surface:18-24 hour ave fcst: -924:15690634:d=2021032312:CDUVB:surface:18-24 hour ave fcst: -969:16717113:d=2021032312:TMP:305 m above mean sea level:24 hour fcst: -970:16735447:d=2021032312:UGRD:305 m above mean sea level:24 hour fcst: -971:16754619:d=2021032312:VGRD:305 m above mean sea level:24 hour fcst: -972:16774205:d=2021032312:TMP:457 m above mean sea level:24 hour fcst: -973:16793711:d=2021032312:UGRD:457 m above mean sea level:24 hour fcst: -974:16813941:d=2021032312:VGRD:457 m above mean sea level:24 hour fcst: -975:16834496:d=2021032312:TMP:610 m above mean sea level:24 hour fcst: -976:16854759:d=2021032312:UGRD:610 m above mean sea level:24 hour fcst: -977:16875611:d=2021032312:VGRD:610 m above mean sea level:24 hour fcst: -978:16896802:d=2021032312:TMP:914 m above mean sea level:24 hour fcst: -979:16917692:d=2021032312:UGRD:914 m above mean sea level:24 hour fcst: -980:16939065:d=2021032312:VGRD:914 m above mean sea level:24 hour fcst: -990:17163364:d=2021032312:TMP:4572 m above mean sea level:24 hour fcst: -991:17184538:d=2021032312:UGRD:4572 m above mean sea level:24 hour fcst: -992:17207932:d=2021032312:VGRD:4572 m above mean sea level:24 hour fcst: -1005:17486635:d=2021032312:TMP:60-30 mb above ground:24 hour fcst: -1006:17507456:d=2021032312:RH:60-30 mb above ground:24 hour fcst: -1007:17526154:d=2021032312:SPFH:60-30 mb above ground:24 hour fcst: -1008:17554901:d=2021032312:UGRD:60-30 mb above ground:24 hour fcst: -1009:17576528:d=2021032312:VGRD:60-30 mb above ground:24 hour fcst: -1010:17598592:d=2021032312:TMP:90-60 mb above ground:24 hour fcst: -1011:17619438:d=2021032312:RH:90-60 mb above ground:24 hour fcst: -1012:17638551:d=2021032312:SPFH:90-60 mb above ground:24 hour fcst: -1013:17667466:d=2021032312:UGRD:90-60 mb above ground:24 hour fcst: -1014:17688852:d=2021032312:VGRD:90-60 mb above ground:24 hour fcst: -1015:17710863:d=2021032312:TMP:120-90 mb above ground:24 hour fcst: -1016:17731807:d=2021032312:RH:120-90 mb above ground:24 hour fcst: -1017:17751204:d=2021032312:SPFH:120-90 mb above ground:24 hour fcst: -1018:17780102:d=2021032312:UGRD:120-90 mb above ground:24 hour fcst: -1019:17801231:d=2021032312:VGRD:120-90 mb above ground:24 hour fcst: -1020:17823044:d=2021032312:TMP:150-120 mb above ground:24 hour fcst: -1021:17843984:d=2021032312:RH:150-120 mb above ground:24 hour fcst: -1022:17863678:d=2021032312:SPFH:150-120 mb above ground:24 hour fcst: -1023:17892812:d=2021032312:UGRD:150-120 mb above ground:24 hour fcst: -1024:17913847:d=2021032312:VGRD:150-120 mb above ground:24 hour fcst: -1025:17935550:d=2021032312:TMP:180-150 mb above ground:24 hour fcst: -1026:17956662:d=2021032312:RH:180-150 mb above ground:24 hour fcst: -1027:17976676:d=2021032312:SPFH:180-150 mb above ground:24 hour fcst: -1028:18005910:d=2021032312:UGRD:180-150 mb above ground:24 hour fcst: -1029:18026976:d=2021032312:VGRD:180-150 mb above ground:24 hour fcst: -1058:18509389:d=2021032312:UGRD:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -1059:18520071:d=2021032312:VGRD:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -1060:18530613:d=2021032312:TMP:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -1061:18545189:d=2021032312:HGT:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -1062:18561905:d=2021032312:PRES:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -1063:18578747:d=2021032312:VWSH:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -1064:18589650:d=2021032312:UGRD:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -1065:18600714:d=2021032312:VGRD:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -1066:18615213:d=2021032312:TMP:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -1067:18630578:d=2021032312:HGT:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -1068:18648282:d=2021032312:PRES:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -1069:18665972:d=2021032312:VWSH:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -1070:18677338:d=2021032312:UGRD:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -1071:18688682:d=2021032312:VGRD:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -1072:18699996:d=2021032312:TMP:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -1073:18715895:d=2021032312:HGT:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -1074:18734631:d=2021032312:PRES:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -1075:18753389:d=2021032312:VWSH:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -1076:18765678:d=2021032312:UGRD:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -1077:18775835:d=2021032312:VGRD:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -1078:18787126:d=2021032312:TMP:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -1079:18802920:d=2021032312:HGT:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -1080:18821446:d=2021032312:PRES:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -1081:18839899:d=2021032312:VWSH:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -1082:18852207:d=2021032312:UGRD:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -1083:18863311:d=2021032312:VGRD:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -1084:18874274:d=2021032312:TMP:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -1085:18889241:d=2021032312:HGT:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -1086:18907097:d=2021032312:PRES:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -1087:18924911:d=2021032312:VWSH:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -1088:18937051:d=2021032312:UGRD:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -1089:18948116:d=2021032312:VGRD:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -1090:18959144:d=2021032312:TMP:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -1091:18974201:d=2021032312:HGT:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -1092:18992035:d=2021032312:PRES:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -1093:19009711:d=2021032312:VWSH:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -+ exglobal_atmos_products.sh[54]true -+ exglobal_atmos_products.sh[55]export err=0 -+ exglobal_atmos_products.sh[55]err=0 -+ exglobal_atmos_products.sh[56][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[62]grid_string=0p25 -+ exglobal_atmos_products.sh[63][[ YES == \Y\E\S ]] -+ exglobal_atmos_products.sh[64]grid_string=0p25:0p50:1p00 -+ exglobal_atmos_products.sh[70]IFS=: -+ exglobal_atmos_products.sh[70]read -ra grids -+ exglobal_atmos_products.sh[154](( nset=1 )) -+ exglobal_atmos_products.sh[154](( nset <= downset )) -+ exglobal_atmos_products.sh[74]echo 'Begin processing nset = 1' -Begin processing nset = 1 -+ exglobal_atmos_products.sh[77]nproc=24 -+ exglobal_atmos_products.sh[80][[ 1 == 1 ]] -+ exglobal_atmos_products.sh[81]grp= -+ exglobal_atmos_products.sh[87]tmpfile=tmpfile_f024 -++ exglobal_atmos_products.sh[90]wgrib2 tmpfile_f024 -++ exglobal_atmos_products.sh[90]wc -l -+ exglobal_atmos_products.sh[90]ncount=753 -+ exglobal_atmos_products.sh[91][[ 24 -gt 753 ]] -+ exglobal_atmos_products.sh[95]inv=31 -+ exglobal_atmos_products.sh[96]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601/poescript -+ exglobal_atmos_products.sh[98]last=0 -+ exglobal_atmos_products.sh[134](( iproc = 1 )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=1 -+ exglobal_atmos_products.sh[101]last=31 -+ exglobal_atmos_products.sh[102][[ 31 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 31 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -31:527345:d=2021032312:UGRD:0.02 mb:24 hour fcst: -+ exglobal_atmos_products.sh[112]rc=0 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 0 == 0 ]] -+ exglobal_atmos_products.sh[115]last=32 -+ exglobal_atmos_products.sh[117][[ 1 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 1:32 -grib tmpfile_f024_1 -1:0:d=2021032312:PRMSL:mean sea level:24 hour fcst: -2:25731:d=2021032312:CLMR:1 hybrid level:24 hour fcst: -3:29524:d=2021032312:ICMR:1 hybrid level:24 hour fcst: -4:37100:d=2021032312:RWMR:1 hybrid level:24 hour fcst: -5:48509:d=2021032312:SNMR:1 hybrid level:24 hour fcst: -6:59635:d=2021032312:GRLE:1 hybrid level:24 hour fcst: -7:62405:d=2021032312:REFD:1 hybrid level:24 hour fcst: -8:88221:d=2021032312:REFD:2 hybrid level:24 hour fcst: -9:114004:d=2021032312:REFC:entire atmosphere (considered as a single layer):24 hour fcst: -10:141439:d=2021032312:VIS:surface:24 hour fcst: -11:173907:d=2021032312:UGRD:planetary boundary layer:24 hour fcst: -12:187864:d=2021032312:VGRD:planetary boundary layer:24 hour fcst: -13:202409:d=2021032312:VRATE:planetary boundary layer:24 hour fcst: -14:217517:d=2021032312:GUST:surface:24 hour fcst: -15:232221:d=2021032312:HGT:0.01 mb:24 hour fcst: -16:257938:d=2021032312:TMP:0.01 mb:24 hour fcst: -17:268909:d=2021032312:RH:0.01 mb:24 hour fcst: -18:275412:d=2021032312:SPFH:0.01 mb:24 hour fcst: -19:298176:d=2021032312:VVEL:0.01 mb:24 hour fcst: -20:325074:d=2021032312:DZDT:0.01 mb:24 hour fcst: -21:352568:d=2021032312:UGRD:0.01 mb:24 hour fcst: -22:363684:d=2021032312:VGRD:0.01 mb:24 hour fcst: -23:374056:d=2021032312:ABSV:0.01 mb:24 hour fcst: -24:389577:d=2021032312:O3MR:0.01 mb:24 hour fcst: -25:403656:d=2021032312:HGT:0.02 mb:24 hour fcst: -26:429080:d=2021032312:TMP:0.02 mb:24 hour fcst: -27:443922:d=2021032312:RH:0.02 mb:24 hour fcst: -28:451652:d=2021032312:SPFH:0.02 mb:24 hour fcst: -29:475844:d=2021032312:VVEL:0.02 mb:24 hour fcst: -30:499814:d=2021032312:DZDT:0.02 mb:24 hour fcst: -31:527345:d=2021032312:UGRD:0.02 mb:24 hour fcst: -32:539841:d=2021032312:VGRD:0.02 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_1 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_1 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_1 pgb2file_f024_1 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 32 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=33 -+ exglobal_atmos_products.sh[101]last=63 -+ exglobal_atmos_products.sh[102][[ 63 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 63 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 2 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 33:63 -grib tmpfile_f024_2 -33:551546:d=2021032312:ABSV:0.02 mb:24 hour fcst: -34:568507:d=2021032312:O3MR:0.02 mb:24 hour fcst: -35:584083:d=2021032312:HGT:0.04 mb:24 hour fcst: -36:608394:d=2021032312:TMP:0.04 mb:24 hour fcst: -37:625667:d=2021032312:RH:0.04 mb:24 hour fcst: -38:631654:d=2021032312:SPFH:0.04 mb:24 hour fcst: -39:657290:d=2021032312:VVEL:0.04 mb:24 hour fcst: -40:684752:d=2021032312:DZDT:0.04 mb:24 hour fcst: -41:709597:d=2021032312:UGRD:0.04 mb:24 hour fcst: -42:729530:d=2021032312:VGRD:0.04 mb:24 hour fcst: -43:749243:d=2021032312:ABSV:0.04 mb:24 hour fcst: -44:767626:d=2021032312:O3MR:0.04 mb:24 hour fcst: -45:791532:d=2021032312:HGT:0.07 mb:24 hour fcst: -46:815629:d=2021032312:TMP:0.07 mb:24 hour fcst: -47:832541:d=2021032312:RH:0.07 mb:24 hour fcst: -48:840370:d=2021032312:SPFH:0.07 mb:24 hour fcst: -49:864128:d=2021032312:VVEL:0.07 mb:24 hour fcst: -50:893051:d=2021032312:DZDT:0.07 mb:24 hour fcst: -51:918481:d=2021032312:UGRD:0.07 mb:24 hour fcst: -52:938837:d=2021032312:VGRD:0.07 mb:24 hour fcst: -53:959146:d=2021032312:ABSV:0.07 mb:24 hour fcst: -54:978056:d=2021032312:O3MR:0.07 mb:24 hour fcst: -55:997869:d=2021032312:HGT:0.1 mb:24 hour fcst: -56:1019441:d=2021032312:TMP:0.1 mb:24 hour fcst: -57:1036109:d=2021032312:RH:0.1 mb:24 hour fcst: -58:1044096:d=2021032312:SPFH:0.1 mb:24 hour fcst: -59:1065442:d=2021032312:VVEL:0.1 mb:24 hour fcst: -60:1095289:d=2021032312:DZDT:0.1 mb:24 hour fcst: -61:1121161:d=2021032312:UGRD:0.1 mb:24 hour fcst: -62:1141154:d=2021032312:VGRD:0.1 mb:24 hour fcst: -63:1161201:d=2021032312:ABSV:0.1 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_2 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_2 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_2 pgb2file_f024_2 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 63 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=64 -+ exglobal_atmos_products.sh[101]last=94 -+ exglobal_atmos_products.sh[102][[ 94 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 94 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 3 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 64:94 -grib tmpfile_f024_3 -64:1179757:d=2021032312:O3MR:0.1 mb:24 hour fcst: -65:1203381:d=2021032312:HGT:0.2 mb:24 hour fcst: -66:1224532:d=2021032312:TMP:0.2 mb:24 hour fcst: -67:1240846:d=2021032312:RH:0.2 mb:24 hour fcst: -68:1251778:d=2021032312:SPFH:0.2 mb:24 hour fcst: -69:1271448:d=2021032312:VVEL:0.2 mb:24 hour fcst: -70:1300400:d=2021032312:DZDT:0.2 mb:24 hour fcst: -71:1327937:d=2021032312:UGRD:0.2 mb:24 hour fcst: -72:1347357:d=2021032312:VGRD:0.2 mb:24 hour fcst: -73:1366271:d=2021032312:ABSV:0.2 mb:24 hour fcst: -74:1383935:d=2021032312:O3MR:0.2 mb:24 hour fcst: -75:1409717:d=2021032312:HGT:0.4 mb:24 hour fcst: -76:1430479:d=2021032312:TMP:0.4 mb:24 hour fcst: -77:1446494:d=2021032312:RH:0.4 mb:24 hour fcst: -78:1455603:d=2021032312:SPFH:0.4 mb:24 hour fcst: -79:1475567:d=2021032312:VVEL:0.4 mb:24 hour fcst: -80:1500487:d=2021032312:DZDT:0.4 mb:24 hour fcst: -81:1531015:d=2021032312:UGRD:0.4 mb:24 hour fcst: -82:1549911:d=2021032312:VGRD:0.4 mb:24 hour fcst: -83:1568182:d=2021032312:ABSV:0.4 mb:24 hour fcst: -84:1584980:d=2021032312:O3MR:0.4 mb:24 hour fcst: -85:1609600:d=2021032312:HGT:0.7 mb:24 hour fcst: -86:1627732:d=2021032312:TMP:0.7 mb:24 hour fcst: -87:1643578:d=2021032312:RH:0.7 mb:24 hour fcst: -88:1652546:d=2021032312:SPFH:0.7 mb:24 hour fcst: -89:1671792:d=2021032312:VVEL:0.7 mb:24 hour fcst: -90:1697702:d=2021032312:DZDT:0.7 mb:24 hour fcst: -91:1726842:d=2021032312:UGRD:0.7 mb:24 hour fcst: -92:1745287:d=2021032312:VGRD:0.7 mb:24 hour fcst: -93:1763146:d=2021032312:ABSV:0.7 mb:24 hour fcst: -94:1779740:d=2021032312:O3MR:0.7 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_3 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_3 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_3 pgb2file_f024_3 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 94 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=95 -+ exglobal_atmos_products.sh[101]last=125 -+ exglobal_atmos_products.sh[102][[ 125 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 125 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 4 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 95:125 -grib tmpfile_f024_4 -95:1805811:d=2021032312:HGT:1 mb:24 hour fcst: -96:1824058:d=2021032312:TMP:1 mb:24 hour fcst: -97:1839923:d=2021032312:RH:1 mb:24 hour fcst: -98:1849749:d=2021032312:SPFH:1 mb:24 hour fcst: -99:1868149:d=2021032312:VVEL:1 mb:24 hour fcst: -100:1895003:d=2021032312:DZDT:1 mb:24 hour fcst: -101:1921096:d=2021032312:UGRD:1 mb:24 hour fcst: -102:1939697:d=2021032312:VGRD:1 mb:24 hour fcst: -103:1957398:d=2021032312:ABSV:1 mb:24 hour fcst: -104:1973752:d=2021032312:O3MR:1 mb:24 hour fcst: -105:1994820:d=2021032312:HGT:2 mb:24 hour fcst: -106:2015132:d=2021032312:TMP:2 mb:24 hour fcst: -107:2031736:d=2021032312:RH:2 mb:24 hour fcst: -108:2040317:d=2021032312:SPFH:2 mb:24 hour fcst: -109:2059312:d=2021032312:VVEL:2 mb:24 hour fcst: -110:2087999:d=2021032312:DZDT:2 mb:24 hour fcst: -111:2114485:d=2021032312:UGRD:2 mb:24 hour fcst: -112:2126070:d=2021032312:VGRD:2 mb:24 hour fcst: -113:2144416:d=2021032312:ABSV:2 mb:24 hour fcst: -114:2161499:d=2021032312:O3MR:2 mb:24 hour fcst: -115:2184317:d=2021032312:HGT:3 mb:24 hour fcst: -116:2202153:d=2021032312:TMP:3 mb:24 hour fcst: -117:2218626:d=2021032312:RH:3 mb:24 hour fcst: -118:2224656:d=2021032312:SPFH:3 mb:24 hour fcst: -119:2245914:d=2021032312:VVEL:3 mb:24 hour fcst: -120:2275970:d=2021032312:DZDT:3 mb:24 hour fcst: -121:2302386:d=2021032312:UGRD:3 mb:24 hour fcst: -122:2313769:d=2021032312:VGRD:3 mb:24 hour fcst: -123:2331906:d=2021032312:ABSV:3 mb:24 hour fcst: -124:2348920:d=2021032312:O3MR:3 mb:24 hour fcst: -125:2372293:d=2021032312:HGT:5 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_4 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_4 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_4 pgb2file_f024_4 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 125 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=126 -+ exglobal_atmos_products.sh[101]last=156 -+ exglobal_atmos_products.sh[102][[ 156 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 156 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 5 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 126:156 -grib tmpfile_f024_5 -126:2389968:d=2021032312:TMP:5 mb:24 hour fcst: -127:2406597:d=2021032312:RH:5 mb:24 hour fcst: -128:2415639:d=2021032312:SPFH:5 mb:24 hour fcst: -129:2436387:d=2021032312:VVEL:5 mb:24 hour fcst: -130:2465571:d=2021032312:DZDT:5 mb:24 hour fcst: -131:2491636:d=2021032312:UGRD:5 mb:24 hour fcst: -132:2503228:d=2021032312:VGRD:5 mb:24 hour fcst: -133:2521798:d=2021032312:ABSV:5 mb:24 hour fcst: -134:2539419:d=2021032312:O3MR:5 mb:24 hour fcst: -135:2562079:d=2021032312:HGT:7 mb:24 hour fcst: -136:2579838:d=2021032312:TMP:7 mb:24 hour fcst: -137:2596027:d=2021032312:RH:7 mb:24 hour fcst: -138:2602146:d=2021032312:SPFH:7 mb:24 hour fcst: -139:2621569:d=2021032312:VVEL:7 mb:24 hour fcst: -140:2646243:d=2021032312:DZDT:7 mb:24 hour fcst: -141:2671752:d=2021032312:UGRD:7 mb:24 hour fcst: -142:2683166:d=2021032312:VGRD:7 mb:24 hour fcst: -143:2701858:d=2021032312:ABSV:7 mb:24 hour fcst: -144:2719667:d=2021032312:O3MR:7 mb:24 hour fcst: -145:2743441:d=2021032312:HGT:10 mb:24 hour fcst: -146:2763289:d=2021032312:TMP:10 mb:24 hour fcst: -147:2779445:d=2021032312:RH:10 mb:24 hour fcst: -148:2787087:d=2021032312:SPFH:10 mb:24 hour fcst: -149:2805333:d=2021032312:VVEL:10 mb:24 hour fcst: -150:2830855:d=2021032312:DZDT:10 mb:24 hour fcst: -151:2856265:d=2021032312:UGRD:10 mb:24 hour fcst: -152:2867606:d=2021032312:VGRD:10 mb:24 hour fcst: -153:2886269:d=2021032312:ABSV:10 mb:24 hour fcst: -154:2904103:d=2021032312:O3MR:10 mb:24 hour fcst: -155:2922881:d=2021032312:HGT:15 mb:24 hour fcst: -156:2942392:d=2021032312:TMP:15 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_5 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_5 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_5 pgb2file_f024_5 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 156 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=157 -+ exglobal_atmos_products.sh[101]last=187 -+ exglobal_atmos_products.sh[102][[ 187 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 187 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 6 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 157:187 -grib tmpfile_f024_6 -157:2958577:d=2021032312:RH:15 mb:24 hour fcst: -158:2963889:d=2021032312:SPFH:15 mb:24 hour fcst: -159:2979362:d=2021032312:VVEL:15 mb:24 hour fcst: -160:3005936:d=2021032312:DZDT:15 mb:24 hour fcst: -161:3036315:d=2021032312:UGRD:15 mb:24 hour fcst: -162:3047581:d=2021032312:VGRD:15 mb:24 hour fcst: -163:3065874:d=2021032312:ABSV:15 mb:24 hour fcst: -164:3083430:d=2021032312:O3MR:15 mb:24 hour fcst: -165:3102481:d=2021032312:HGT:20 mb:24 hour fcst: -166:3121899:d=2021032312:TMP:20 mb:24 hour fcst: -167:3137954:d=2021032312:RH:20 mb:24 hour fcst: -168:3144158:d=2021032312:SPFH:20 mb:24 hour fcst: -169:3158568:d=2021032312:VVEL:20 mb:24 hour fcst: -170:3185926:d=2021032312:DZDT:20 mb:24 hour fcst: -171:3216210:d=2021032312:UGRD:20 mb:24 hour fcst: -172:3234848:d=2021032312:VGRD:20 mb:24 hour fcst: -173:3253386:d=2021032312:ABSV:20 mb:24 hour fcst: -174:3270995:d=2021032312:O3MR:20 mb:24 hour fcst: -175:3295921:d=2021032312:HGT:30 mb:24 hour fcst: -176:3315236:d=2021032312:TMP:30 mb:24 hour fcst: -177:3331242:d=2021032312:RH:30 mb:24 hour fcst: -178:3338836:d=2021032312:SPFH:30 mb:24 hour fcst: -179:3354139:d=2021032312:VVEL:30 mb:24 hour fcst: -180:3382646:d=2021032312:DZDT:30 mb:24 hour fcst: -181:3412625:d=2021032312:UGRD:30 mb:24 hour fcst: -182:3431328:d=2021032312:VGRD:30 mb:24 hour fcst: -183:3449908:d=2021032312:ABSV:30 mb:24 hour fcst: -184:3467608:d=2021032312:O3MR:30 mb:24 hour fcst: -185:3494438:d=2021032312:HGT:40 mb:24 hour fcst: -186:3513288:d=2021032312:TMP:40 mb:24 hour fcst: -187:3529623:d=2021032312:RH:40 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_6 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_6 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_6 pgb2file_f024_6 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 187 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=188 -+ exglobal_atmos_products.sh[101]last=218 -+ exglobal_atmos_products.sh[102][[ 218 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 218 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 7 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 188:218 -grib tmpfile_f024_7 -188:3539149:d=2021032312:SPFH:40 mb:24 hour fcst: -189:3556233:d=2021032312:VVEL:40 mb:24 hour fcst: -190:3585634:d=2021032312:DZDT:40 mb:24 hour fcst: -191:3615576:d=2021032312:UGRD:40 mb:24 hour fcst: -192:3634336:d=2021032312:VGRD:40 mb:24 hour fcst: -193:3653066:d=2021032312:ABSV:40 mb:24 hour fcst: -194:3670911:d=2021032312:O3MR:40 mb:24 hour fcst: -195:3697855:d=2021032312:HGT:50 mb:24 hour fcst: -196:3716760:d=2021032312:TMP:50 mb:24 hour fcst: -197:3732896:d=2021032312:RH:50 mb:24 hour fcst: -198:3743536:d=2021032312:TCDC:50 mb:24 hour fcst: -199:3743715:d=2021032312:SPFH:50 mb:24 hour fcst: -200:3762260:d=2021032312:VVEL:50 mb:24 hour fcst: -201:3792355:d=2021032312:DZDT:50 mb:24 hour fcst: -202:3821970:d=2021032312:UGRD:50 mb:24 hour fcst: -203:3840451:d=2021032312:VGRD:50 mb:24 hour fcst: -204:3858977:d=2021032312:ABSV:50 mb:24 hour fcst: -205:3876672:d=2021032312:CLMR:50 mb:24 hour fcst: -206:3876851:d=2021032312:ICMR:50 mb:24 hour fcst: -207:3877030:d=2021032312:RWMR:50 mb:24 hour fcst: -208:3877209:d=2021032312:SNMR:50 mb:24 hour fcst: -209:3877388:d=2021032312:GRLE:50 mb:24 hour fcst: -210:3877567:d=2021032312:O3MR:50 mb:24 hour fcst: -211:3903927:d=2021032312:HGT:70 mb:24 hour fcst: -212:3923007:d=2021032312:TMP:70 mb:24 hour fcst: -213:3939512:d=2021032312:RH:70 mb:24 hour fcst: -214:3947743:d=2021032312:SPFH:70 mb:24 hour fcst: -215:3968720:d=2021032312:VVEL:70 mb:24 hour fcst: -216:4000141:d=2021032312:DZDT:70 mb:24 hour fcst: -217:4032244:d=2021032312:UGRD:70 mb:24 hour fcst: -218:4050937:d=2021032312:VGRD:70 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_7 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_7 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_7 pgb2file_f024_7 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 218 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=219 -+ exglobal_atmos_products.sh[101]last=249 -+ exglobal_atmos_products.sh[102][[ 249 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 249 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 8 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 219:249 -grib tmpfile_f024_8 -219:4069588:d=2021032312:ABSV:70 mb:24 hour fcst: -220:4087397:d=2021032312:O3MR:70 mb:24 hour fcst: -221:4113201:d=2021032312:HGT:100 mb:24 hour fcst: -222:4132850:d=2021032312:TMP:100 mb:24 hour fcst: -223:4149768:d=2021032312:RH:100 mb:24 hour fcst: -224:4159917:d=2021032312:TCDC:100 mb:24 hour fcst: -225:4160096:d=2021032312:SPFH:100 mb:24 hour fcst: -226:4185930:d=2021032312:VVEL:100 mb:24 hour fcst: -227:4210987:d=2021032312:DZDT:100 mb:24 hour fcst: -228:4235599:d=2021032312:UGRD:100 mb:24 hour fcst: -229:4254965:d=2021032312:VGRD:100 mb:24 hour fcst: -230:4274397:d=2021032312:ABSV:100 mb:24 hour fcst: -231:4292586:d=2021032312:CLMR:100 mb:24 hour fcst: -232:4292765:d=2021032312:ICMR:100 mb:24 hour fcst: -233:4293571:d=2021032312:RWMR:100 mb:24 hour fcst: -234:4293750:d=2021032312:SNMR:100 mb:24 hour fcst: -235:4295717:d=2021032312:GRLE:100 mb:24 hour fcst: -236:4295896:d=2021032312:O3MR:100 mb:24 hour fcst: -237:4319959:d=2021032312:HGT:150 mb:24 hour fcst: -238:4340359:d=2021032312:TMP:150 mb:24 hour fcst: -239:4357328:d=2021032312:RH:150 mb:24 hour fcst: -240:4368787:d=2021032312:TCDC:150 mb:24 hour fcst: -241:4372818:d=2021032312:SPFH:150 mb:24 hour fcst: -242:4397195:d=2021032312:VVEL:150 mb:24 hour fcst: -243:4424507:d=2021032312:DZDT:150 mb:24 hour fcst: -244:4450000:d=2021032312:UGRD:150 mb:24 hour fcst: -245:4469973:d=2021032312:VGRD:150 mb:24 hour fcst: -246:4489999:d=2021032312:ABSV:150 mb:24 hour fcst: -247:4508877:d=2021032312:CLMR:150 mb:24 hour fcst: -248:4509056:d=2021032312:ICMR:150 mb:24 hour fcst: -249:4514328:d=2021032312:RWMR:150 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_8 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_8 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_8 pgb2file_f024_8 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 249 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=250 -+ exglobal_atmos_products.sh[101]last=280 -+ exglobal_atmos_products.sh[102][[ 280 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 280 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 9 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 250:280 -grib tmpfile_f024_9 -250:4514507:d=2021032312:SNMR:150 mb:24 hour fcst: -251:4518026:d=2021032312:GRLE:150 mb:24 hour fcst: -252:4518578:d=2021032312:O3MR:150 mb:24 hour fcst: -253:4540488:d=2021032312:HGT:200 mb:24 hour fcst: -254:4561526:d=2021032312:TMP:200 mb:24 hour fcst: -255:4579631:d=2021032312:RH:200 mb:24 hour fcst: -256:4594717:d=2021032312:TCDC:200 mb:24 hour fcst: -257:4601209:d=2021032312:SPFH:200 mb:24 hour fcst: -258:4626978:d=2021032312:VVEL:200 mb:24 hour fcst: -259:4653713:d=2021032312:DZDT:200 mb:24 hour fcst: -260:4680062:d=2021032312:UGRD:200 mb:24 hour fcst: -261:4693465:d=2021032312:VGRD:200 mb:24 hour fcst: -262:4714742:d=2021032312:ABSV:200 mb:24 hour fcst: -263:4734690:d=2021032312:CLMR:200 mb:24 hour fcst: -264:4734869:d=2021032312:ICMR:200 mb:24 hour fcst: -265:4743280:d=2021032312:RWMR:200 mb:24 hour fcst: -266:4743459:d=2021032312:SNMR:200 mb:24 hour fcst: -267:4748767:d=2021032312:GRLE:200 mb:24 hour fcst: -268:4749186:d=2021032312:O3MR:200 mb:24 hour fcst: -269:4769840:d=2021032312:HGT:250 mb:24 hour fcst: -270:4791765:d=2021032312:TMP:250 mb:24 hour fcst: -271:4810651:d=2021032312:RH:250 mb:24 hour fcst: -272:4828442:d=2021032312:TCDC:250 mb:24 hour fcst: -273:4838980:d=2021032312:SPFH:250 mb:24 hour fcst: -274:4867283:d=2021032312:VVEL:250 mb:24 hour fcst: -275:4895668:d=2021032312:DZDT:250 mb:24 hour fcst: -276:4922815:d=2021032312:UGRD:250 mb:24 hour fcst: -277:4937115:d=2021032312:VGRD:250 mb:24 hour fcst: -278:4951932:d=2021032312:ABSV:250 mb:24 hour fcst: -279:4973381:d=2021032312:CLMR:250 mb:24 hour fcst: -280:4973560:d=2021032312:ICMR:250 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_9 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_9 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_9 pgb2file_f024_9 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 280 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=281 -+ exglobal_atmos_products.sh[101]last=311 -+ exglobal_atmos_products.sh[102][[ 311 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 311 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 10 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 281:311 -grib tmpfile_f024_10 -281:4987823:d=2021032312:RWMR:250 mb:24 hour fcst: -282:4988002:d=2021032312:SNMR:250 mb:24 hour fcst: -283:4996807:d=2021032312:GRLE:250 mb:24 hour fcst: -284:4997151:d=2021032312:O3MR:250 mb:24 hour fcst: -285:5022587:d=2021032312:HGT:300 mb:24 hour fcst: -286:5044897:d=2021032312:TMP:300 mb:24 hour fcst: -287:5063465:d=2021032312:RH:300 mb:24 hour fcst: -288:5082464:d=2021032312:TCDC:300 mb:24 hour fcst: -289:5095196:d=2021032312:SPFH:300 mb:24 hour fcst: -290:5121900:d=2021032312:VVEL:300 mb:24 hour fcst: -291:5145815:d=2021032312:DZDT:300 mb:24 hour fcst: -292:5173963:d=2021032312:UGRD:300 mb:24 hour fcst: -293:5188468:d=2021032312:VGRD:300 mb:24 hour fcst: -294:5203790:d=2021032312:ABSV:300 mb:24 hour fcst: -295:5226099:d=2021032312:CLMR:300 mb:24 hour fcst: -296:5226278:d=2021032312:ICMR:300 mb:24 hour fcst: -297:5242097:d=2021032312:RWMR:300 mb:24 hour fcst: -298:5242276:d=2021032312:SNMR:300 mb:24 hour fcst: -299:5253677:d=2021032312:GRLE:300 mb:24 hour fcst: -300:5254088:d=2021032312:O3MR:300 mb:24 hour fcst: -301:5281087:d=2021032312:HGT:350 mb:24 hour fcst: -302:5303224:d=2021032312:TMP:350 mb:24 hour fcst: -303:5321438:d=2021032312:RH:350 mb:24 hour fcst: -304:5340551:d=2021032312:TCDC:350 mb:24 hour fcst: -305:5353643:d=2021032312:SPFH:350 mb:24 hour fcst: -306:5381826:d=2021032312:VVEL:350 mb:24 hour fcst: -307:5406353:d=2021032312:DZDT:350 mb:24 hour fcst: -308:5434673:d=2021032312:UGRD:350 mb:24 hour fcst: -309:5449263:d=2021032312:VGRD:350 mb:24 hour fcst: -310:5464513:d=2021032312:ABSV:350 mb:24 hour fcst: -311:5486782:d=2021032312:CLMR:350 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_10 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_10 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_10 pgb2file_f024_10 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 311 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=312 -+ exglobal_atmos_products.sh[101]last=342 -+ exglobal_atmos_products.sh[102][[ 342 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 342 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 11 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 312:342 -grib tmpfile_f024_11 -312:5486961:d=2021032312:ICMR:350 mb:24 hour fcst: -313:5503775:d=2021032312:RWMR:350 mb:24 hour fcst: -314:5503954:d=2021032312:SNMR:350 mb:24 hour fcst: -315:5517233:d=2021032312:GRLE:350 mb:24 hour fcst: -316:5517740:d=2021032312:O3MR:350 mb:24 hour fcst: -317:5543316:d=2021032312:HGT:400 mb:24 hour fcst: -318:5565069:d=2021032312:TMP:400 mb:24 hour fcst: -319:5583253:d=2021032312:RH:400 mb:24 hour fcst: -320:5602454:d=2021032312:TCDC:400 mb:24 hour fcst: -321:5615571:d=2021032312:SPFH:400 mb:24 hour fcst: -322:5643439:d=2021032312:VVEL:400 mb:24 hour fcst: -323:5668395:d=2021032312:DZDT:400 mb:24 hour fcst: -324:5696788:d=2021032312:UGRD:400 mb:24 hour fcst: -325:5710863:d=2021032312:VGRD:400 mb:24 hour fcst: -326:5733573:d=2021032312:ABSV:400 mb:24 hour fcst: -327:5755535:d=2021032312:CLMR:400 mb:24 hour fcst: -328:5755909:d=2021032312:ICMR:400 mb:24 hour fcst: -329:5772894:d=2021032312:RWMR:400 mb:24 hour fcst: -330:5773182:d=2021032312:SNMR:400 mb:24 hour fcst: -331:5787105:d=2021032312:GRLE:400 mb:24 hour fcst: -332:5787917:d=2021032312:O3MR:400 mb:24 hour fcst: -333:5812501:d=2021032312:HGT:450 mb:24 hour fcst: -334:5836463:d=2021032312:TMP:450 mb:24 hour fcst: -335:5854839:d=2021032312:RH:450 mb:24 hour fcst: -336:5874249:d=2021032312:TCDC:450 mb:24 hour fcst: -337:5886976:d=2021032312:SPFH:450 mb:24 hour fcst: -338:5916453:d=2021032312:VVEL:450 mb:24 hour fcst: -339:5941611:d=2021032312:DZDT:450 mb:24 hour fcst: -340:5969760:d=2021032312:UGRD:450 mb:24 hour fcst: -341:5991444:d=2021032312:VGRD:450 mb:24 hour fcst: -342:6013772:d=2021032312:ABSV:450 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_11 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_11 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_11 pgb2file_f024_11 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 342 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=343 -+ exglobal_atmos_products.sh[101]last=373 -+ exglobal_atmos_products.sh[102][[ 373 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 373 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 12 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 343:373 -grib tmpfile_f024_12 -343:6035214:d=2021032312:CLMR:450 mb:24 hour fcst: -344:6036153:d=2021032312:ICMR:450 mb:24 hour fcst: -345:6052714:d=2021032312:RWMR:450 mb:24 hour fcst: -346:6053233:d=2021032312:SNMR:450 mb:24 hour fcst: -347:6068111:d=2021032312:GRLE:450 mb:24 hour fcst: -348:6069380:d=2021032312:O3MR:450 mb:24 hour fcst: -349:6093477:d=2021032312:HGT:500 mb:24 hour fcst: -350:6117204:d=2021032312:TMP:500 mb:24 hour fcst: -351:6135745:d=2021032312:RH:500 mb:24 hour fcst: -352:6155320:d=2021032312:TCDC:500 mb:24 hour fcst: -353:6167522:d=2021032312:SPFH:500 mb:24 hour fcst: -354:6195869:d=2021032312:VVEL:500 mb:24 hour fcst: -355:6220942:d=2021032312:DZDT:500 mb:24 hour fcst: -356:6248665:d=2021032312:UGRD:500 mb:24 hour fcst: -357:6270102:d=2021032312:VGRD:500 mb:24 hour fcst: -358:6292142:d=2021032312:ABSV:500 mb:24 hour fcst: -359:6313311:d=2021032312:CLMR:500 mb:24 hour fcst: -360:6314955:d=2021032312:ICMR:500 mb:24 hour fcst: -361:6332343:d=2021032312:RWMR:500 mb:24 hour fcst: -362:6333284:d=2021032312:SNMR:500 mb:24 hour fcst: -363:6347558:d=2021032312:GRLE:500 mb:24 hour fcst: -364:6349466:d=2021032312:O3MR:500 mb:24 hour fcst: -365:6372884:d=2021032312:HGT:550 mb:24 hour fcst: -366:6396451:d=2021032312:TMP:550 mb:24 hour fcst: -367:6415124:d=2021032312:RH:550 mb:24 hour fcst: -368:6434972:d=2021032312:TCDC:550 mb:24 hour fcst: -369:6447151:d=2021032312:SPFH:550 mb:24 hour fcst: -370:6476742:d=2021032312:VVEL:550 mb:24 hour fcst: -371:6501989:d=2021032312:DZDT:550 mb:24 hour fcst: -372:6529526:d=2021032312:UGRD:550 mb:24 hour fcst: -373:6550699:d=2021032312:VGRD:550 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_12 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_12 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_12 pgb2file_f024_12 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 373 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=374 -+ exglobal_atmos_products.sh[101]last=404 -+ exglobal_atmos_products.sh[102][[ 404 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 404 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -404:7101926:d=2021032312:UGRD:650 mb:24 hour fcst: -+ exglobal_atmos_products.sh[112]rc=0 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 0 == 0 ]] -+ exglobal_atmos_products.sh[115]last=405 -+ exglobal_atmos_products.sh[117][[ 13 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 374:405 -grib tmpfile_f024_13 -374:6572462:d=2021032312:ABSV:550 mb:24 hour fcst: -375:6593416:d=2021032312:CLMR:550 mb:24 hour fcst: -376:6596611:d=2021032312:ICMR:550 mb:24 hour fcst: -377:6613183:d=2021032312:RWMR:550 mb:24 hour fcst: -378:6614702:d=2021032312:SNMR:550 mb:24 hour fcst: -379:6629097:d=2021032312:GRLE:550 mb:24 hour fcst: -380:6630938:d=2021032312:O3MR:550 mb:24 hour fcst: -381:6653860:d=2021032312:HGT:600 mb:24 hour fcst: -382:6678085:d=2021032312:TMP:600 mb:24 hour fcst: -383:6697106:d=2021032312:RH:600 mb:24 hour fcst: -384:6717183:d=2021032312:TCDC:600 mb:24 hour fcst: -385:6729299:d=2021032312:SPFH:600 mb:24 hour fcst: -386:6759744:d=2021032312:VVEL:600 mb:24 hour fcst: -387:6785157:d=2021032312:DZDT:600 mb:24 hour fcst: -388:6812716:d=2021032312:UGRD:600 mb:24 hour fcst: -389:6833761:d=2021032312:VGRD:600 mb:24 hour fcst: -390:6855383:d=2021032312:ABSV:600 mb:24 hour fcst: -391:6876143:d=2021032312:CLMR:600 mb:24 hour fcst: -392:6880918:d=2021032312:ICMR:600 mb:24 hour fcst: -393:6894851:d=2021032312:RWMR:600 mb:24 hour fcst: -394:6898392:d=2021032312:SNMR:600 mb:24 hour fcst: -395:6914213:d=2021032312:GRLE:600 mb:24 hour fcst: -396:6916611:d=2021032312:O3MR:600 mb:24 hour fcst: -397:6944822:d=2021032312:HGT:650 mb:24 hour fcst: -398:6968998:d=2021032312:TMP:650 mb:24 hour fcst: -399:6988336:d=2021032312:RH:650 mb:24 hour fcst: -400:7008586:d=2021032312:TCDC:650 mb:24 hour fcst: -401:7020303:d=2021032312:SPFH:650 mb:24 hour fcst: -402:7048978:d=2021032312:VVEL:650 mb:24 hour fcst: -403:7074530:d=2021032312:DZDT:650 mb:24 hour fcst: -404:7101926:d=2021032312:UGRD:650 mb:24 hour fcst: -405:7122803:d=2021032312:VGRD:650 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_13 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_13 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_13 pgb2file_f024_13 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 405 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=406 -+ exglobal_atmos_products.sh[101]last=436 -+ exglobal_atmos_products.sh[102][[ 436 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 436 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -436:7679832:d=2021032312:UGRD:750 mb:24 hour fcst: -+ exglobal_atmos_products.sh[112]rc=0 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 0 == 0 ]] -+ exglobal_atmos_products.sh[115]last=437 -+ exglobal_atmos_products.sh[117][[ 14 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 406:437 -grib tmpfile_f024_14 -406:7144498:d=2021032312:ABSV:650 mb:24 hour fcst: -407:7165402:d=2021032312:CLMR:650 mb:24 hour fcst: -408:7170624:d=2021032312:ICMR:650 mb:24 hour fcst: -409:7183060:d=2021032312:RWMR:650 mb:24 hour fcst: -410:7187630:d=2021032312:SNMR:650 mb:24 hour fcst: -411:7202697:d=2021032312:GRLE:650 mb:24 hour fcst: -412:7204975:d=2021032312:O3MR:650 mb:24 hour fcst: -413:7232736:d=2021032312:HGT:700 mb:24 hour fcst: -414:7257074:d=2021032312:TMP:700 mb:24 hour fcst: -415:7276681:d=2021032312:RH:700 mb:24 hour fcst: -416:7296972:d=2021032312:TCDC:700 mb:24 hour fcst: -417:7308504:d=2021032312:SPFH:700 mb:24 hour fcst: -418:7337116:d=2021032312:VVEL:700 mb:24 hour fcst: -419:7362634:d=2021032312:DZDT:700 mb:24 hour fcst: -420:7389799:d=2021032312:UGRD:700 mb:24 hour fcst: -421:7410634:d=2021032312:VGRD:700 mb:24 hour fcst: -422:7432251:d=2021032312:ABSV:700 mb:24 hour fcst: -423:7453286:d=2021032312:CLMR:700 mb:24 hour fcst: -424:7460235:d=2021032312:ICMR:700 mb:24 hour fcst: -425:7471377:d=2021032312:RWMR:700 mb:24 hour fcst: -426:7476896:d=2021032312:SNMR:700 mb:24 hour fcst: -427:7491124:d=2021032312:GRLE:700 mb:24 hour fcst: -428:7493373:d=2021032312:O3MR:700 mb:24 hour fcst: -429:7520748:d=2021032312:HGT:750 mb:24 hour fcst: -430:7545153:d=2021032312:TMP:750 mb:24 hour fcst: -431:7565045:d=2021032312:RH:750 mb:24 hour fcst: -432:7585497:d=2021032312:TCDC:750 mb:24 hour fcst: -433:7597830:d=2021032312:SPFH:750 mb:24 hour fcst: -434:7626958:d=2021032312:VVEL:750 mb:24 hour fcst: -435:7652661:d=2021032312:DZDT:750 mb:24 hour fcst: -436:7679832:d=2021032312:UGRD:750 mb:24 hour fcst: -437:7700676:d=2021032312:VGRD:750 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_14 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_14 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_14 pgb2file_f024_14 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 437 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=438 -+ exglobal_atmos_products.sh[101]last=468 -+ exglobal_atmos_products.sh[102][[ 468 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 468 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -468:8272175:d=2021032312:UGRD:850 mb:24 hour fcst: -+ exglobal_atmos_products.sh[112]rc=0 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 0 == 0 ]] -+ exglobal_atmos_products.sh[115]last=469 -+ exglobal_atmos_products.sh[117][[ 15 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 438:469 -grib tmpfile_f024_15 -438:7722360:d=2021032312:ABSV:750 mb:24 hour fcst: -439:7743471:d=2021032312:CLMR:750 mb:24 hour fcst: -440:7752847:d=2021032312:ICMR:750 mb:24 hour fcst: -441:7764539:d=2021032312:RWMR:750 mb:24 hour fcst: -442:7770469:d=2021032312:SNMR:750 mb:24 hour fcst: -443:7783985:d=2021032312:GRLE:750 mb:24 hour fcst: -444:7785919:d=2021032312:O3MR:750 mb:24 hour fcst: -445:7813182:d=2021032312:HGT:800 mb:24 hour fcst: -446:7837702:d=2021032312:TMP:800 mb:24 hour fcst: -447:7858104:d=2021032312:RH:800 mb:24 hour fcst: -448:7878581:d=2021032312:TCDC:800 mb:24 hour fcst: -449:7891846:d=2021032312:SPFH:800 mb:24 hour fcst: -450:7921270:d=2021032312:VVEL:800 mb:24 hour fcst: -451:7946970:d=2021032312:DZDT:800 mb:24 hour fcst: -452:7974014:d=2021032312:UGRD:800 mb:24 hour fcst: -453:7994946:d=2021032312:VGRD:800 mb:24 hour fcst: -454:8016698:d=2021032312:ABSV:800 mb:24 hour fcst: -455:8037915:d=2021032312:CLMR:800 mb:24 hour fcst: -456:8049264:d=2021032312:ICMR:800 mb:24 hour fcst: -457:8060217:d=2021032312:RWMR:800 mb:24 hour fcst: -458:8067929:d=2021032312:SNMR:800 mb:24 hour fcst: -459:8080848:d=2021032312:GRLE:800 mb:24 hour fcst: -460:8083160:d=2021032312:O3MR:800 mb:24 hour fcst: -461:8110171:d=2021032312:HGT:850 mb:24 hour fcst: -462:8134965:d=2021032312:TMP:850 mb:24 hour fcst: -463:8155653:d=2021032312:RH:850 mb:24 hour fcst: -464:8175943:d=2021032312:TCDC:850 mb:24 hour fcst: -465:8190661:d=2021032312:SPFH:850 mb:24 hour fcst: -466:8220335:d=2021032312:VVEL:850 mb:24 hour fcst: -467:8245615:d=2021032312:DZDT:850 mb:24 hour fcst: -468:8272175:d=2021032312:UGRD:850 mb:24 hour fcst: -469:8293330:d=2021032312:VGRD:850 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_15 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_15 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_15 pgb2file_f024_15 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 469 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=470 -+ exglobal_atmos_products.sh[101]last=500 -+ exglobal_atmos_products.sh[102][[ 500 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 500 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -500:8863902:d=2021032312:UGRD:925 mb:24 hour fcst: -+ exglobal_atmos_products.sh[112]rc=0 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 0 == 0 ]] -+ exglobal_atmos_products.sh[115]last=501 -+ exglobal_atmos_products.sh[117][[ 16 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 470:501 -grib tmpfile_f024_16 -470:8315148:d=2021032312:ABSV:850 mb:24 hour fcst: -471:8336365:d=2021032312:CLMR:850 mb:24 hour fcst: -472:8350153:d=2021032312:ICMR:850 mb:24 hour fcst: -473:8360776:d=2021032312:RWMR:850 mb:24 hour fcst: -474:8371572:d=2021032312:SNMR:850 mb:24 hour fcst: -475:8383270:d=2021032312:GRLE:850 mb:24 hour fcst: -476:8385729:d=2021032312:O3MR:850 mb:24 hour fcst: -477:8412306:d=2021032312:HGT:900 mb:24 hour fcst: -478:8437342:d=2021032312:TMP:900 mb:24 hour fcst: -479:8458149:d=2021032312:RH:900 mb:24 hour fcst: -480:8478058:d=2021032312:TCDC:900 mb:24 hour fcst: -481:8491037:d=2021032312:SPFH:900 mb:24 hour fcst: -482:8520453:d=2021032312:VVEL:900 mb:24 hour fcst: -483:8545288:d=2021032312:DZDT:900 mb:24 hour fcst: -484:8571350:d=2021032312:UGRD:900 mb:24 hour fcst: -485:8592524:d=2021032312:VGRD:900 mb:24 hour fcst: -486:8614428:d=2021032312:ABSV:900 mb:24 hour fcst: -487:8635624:d=2021032312:CLMR:900 mb:24 hour fcst: -488:8645994:d=2021032312:ICMR:900 mb:24 hour fcst: -489:8654469:d=2021032312:RWMR:900 mb:24 hour fcst: -490:8666813:d=2021032312:SNMR:900 mb:24 hour fcst: -491:8677094:d=2021032312:GRLE:900 mb:24 hour fcst: -492:8680153:d=2021032312:O3MR:900 mb:24 hour fcst: -493:8706255:d=2021032312:HGT:925 mb:24 hour fcst: -494:8731426:d=2021032312:TMP:925 mb:24 hour fcst: -495:8752573:d=2021032312:RH:925 mb:24 hour fcst: -496:8772300:d=2021032312:TCDC:925 mb:24 hour fcst: -497:8784130:d=2021032312:SPFH:925 mb:24 hour fcst: -498:8813869:d=2021032312:VVEL:925 mb:24 hour fcst: -499:8838306:d=2021032312:DZDT:925 mb:24 hour fcst: -500:8863902:d=2021032312:UGRD:925 mb:24 hour fcst: -501:8885331:d=2021032312:VGRD:925 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_16 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_16 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_16 pgb2file_f024_16 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 501 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=502 -+ exglobal_atmos_products.sh[101]last=532 -+ exglobal_atmos_products.sh[102][[ 532 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 532 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 17 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 502:532 -grib tmpfile_f024_17 -502:8907235:d=2021032312:ABSV:925 mb:24 hour fcst: -503:8928527:d=2021032312:CLMR:925 mb:24 hour fcst: -504:8937230:d=2021032312:ICMR:925 mb:24 hour fcst: -505:8944858:d=2021032312:RWMR:925 mb:24 hour fcst: -506:8957364:d=2021032312:SNMR:925 mb:24 hour fcst: -507:8966844:d=2021032312:GRLE:925 mb:24 hour fcst: -508:8970028:d=2021032312:O3MR:925 mb:24 hour fcst: -509:8996059:d=2021032312:HGT:950 mb:24 hour fcst: -510:9021412:d=2021032312:TMP:950 mb:24 hour fcst: -511:9042531:d=2021032312:RH:950 mb:24 hour fcst: -512:9062059:d=2021032312:TCDC:950 mb:24 hour fcst: -513:9072224:d=2021032312:SPFH:950 mb:24 hour fcst: -514:9099426:d=2021032312:VVEL:950 mb:24 hour fcst: -515:9123249:d=2021032312:DZDT:950 mb:24 hour fcst: -516:9147928:d=2021032312:UGRD:950 mb:24 hour fcst: -517:9169366:d=2021032312:VGRD:950 mb:24 hour fcst: -518:9191266:d=2021032312:ABSV:950 mb:24 hour fcst: -519:9212493:d=2021032312:CLMR:950 mb:24 hour fcst: -520:9218918:d=2021032312:ICMR:950 mb:24 hour fcst: -521:9226093:d=2021032312:RWMR:950 mb:24 hour fcst: -522:9237409:d=2021032312:SNMR:950 mb:24 hour fcst: -523:9246161:d=2021032312:GRLE:950 mb:24 hour fcst: -524:9249046:d=2021032312:O3MR:950 mb:24 hour fcst: -525:9275040:d=2021032312:HINDEX:surface:24 hour fcst: -526:9280133:d=2021032312:HGT:975 mb:24 hour fcst: -527:9305645:d=2021032312:TMP:975 mb:24 hour fcst: -528:9326550:d=2021032312:RH:975 mb:24 hour fcst: -529:9345669:d=2021032312:TCDC:975 mb:24 hour fcst: -530:9353946:d=2021032312:SPFH:975 mb:24 hour fcst: -531:9381012:d=2021032312:VVEL:975 mb:24 hour fcst: -532:9403758:d=2021032312:DZDT:975 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_17 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_17 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_17 pgb2file_f024_17 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 532 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=533 -+ exglobal_atmos_products.sh[101]last=563 -+ exglobal_atmos_products.sh[102][[ 563 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 563 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 18 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 533:563 -grib tmpfile_f024_18 -533:9427318:d=2021032312:UGRD:975 mb:24 hour fcst: -534:9448712:d=2021032312:VGRD:975 mb:24 hour fcst: -535:9470544:d=2021032312:ABSV:975 mb:24 hour fcst: -536:9491538:d=2021032312:CLMR:975 mb:24 hour fcst: -537:9496240:d=2021032312:ICMR:975 mb:24 hour fcst: -538:9501488:d=2021032312:RWMR:975 mb:24 hour fcst: -539:9513299:d=2021032312:SNMR:975 mb:24 hour fcst: -540:9520971:d=2021032312:GRLE:975 mb:24 hour fcst: -541:9523666:d=2021032312:O3MR:975 mb:24 hour fcst: -542:9549556:d=2021032312:TMP:1000 mb:24 hour fcst: -543:9570331:d=2021032312:RH:1000 mb:24 hour fcst: -544:9589217:d=2021032312:TCDC:1000 mb:24 hour fcst: -545:9593796:d=2021032312:SPFH:1000 mb:24 hour fcst: -546:9620886:d=2021032312:VVEL:1000 mb:24 hour fcst: -547:9642185:d=2021032312:DZDT:1000 mb:24 hour fcst: -548:9663533:d=2021032312:UGRD:1000 mb:24 hour fcst: -549:9684322:d=2021032312:VGRD:1000 mb:24 hour fcst: -550:9705713:d=2021032312:ABSV:1000 mb:24 hour fcst: -551:9726245:d=2021032312:CLMR:1000 mb:24 hour fcst: -552:9728160:d=2021032312:ICMR:1000 mb:24 hour fcst: -553:9730680:d=2021032312:RWMR:1000 mb:24 hour fcst: -554:9740389:d=2021032312:SNMR:1000 mb:24 hour fcst: -555:9743708:d=2021032312:GRLE:1000 mb:24 hour fcst: -556:9744743:d=2021032312:O3MR:1000 mb:24 hour fcst: -557:9770675:d=2021032312:MSLET:mean sea level:24 hour fcst: -558:9797283:d=2021032312:HGT:1000 mb:24 hour fcst: -559:9823000:d=2021032312:SDEN:surface:24 hour fcst: -560:9833960:d=2021032312:REFD:4000 m above ground:24 hour fcst: -561:9846988:d=2021032312:REFD:1000 m above ground:24 hour fcst: -562:9871191:d=2021032312:PRES:surface:24 hour fcst: -563:9896840:d=2021032312:HGT:surface:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_18 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_18 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_18 pgb2file_f024_18 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 563 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=564 -+ exglobal_atmos_products.sh[101]last=594 -+ exglobal_atmos_products.sh[102][[ 594 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 594 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 19 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 564:594 -grib tmpfile_f024_19 -564:9912910:d=2021032312:TMP:surface:24 hour fcst: -565:9929540:d=2021032312:TSOIL:0-0.1 m below ground:24 hour fcst: -566:9941471:d=2021032312:SOILW:0-0.1 m below ground:24 hour fcst: -567:9950382:d=2021032312:SOILL:0-0.1 m below ground:24 hour fcst: -568:9959179:d=2021032312:TSOIL:0.1-0.4 m below ground:24 hour fcst: -569:9974639:d=2021032312:SOILW:0.1-0.4 m below ground:24 hour fcst: -570:9983656:d=2021032312:SOILL:0.1-0.4 m below ground:24 hour fcst: -571:9992550:d=2021032312:TSOIL:0.4-1 m below ground:24 hour fcst: -572:10006989:d=2021032312:SOILW:0.4-1 m below ground:24 hour fcst: -573:10016122:d=2021032312:SOILL:0.4-1 m below ground:24 hour fcst: -574:10025026:d=2021032312:TSOIL:1-2 m below ground:24 hour fcst: -575:10039288:d=2021032312:SOILW:1-2 m below ground:24 hour fcst: -576:10048348:d=2021032312:SOILL:1-2 m below ground:24 hour fcst: -577:10057424:d=2021032312:CNWAT:surface:24 hour fcst: -578:10063009:d=2021032312:WEASD:surface:24 hour fcst: -579:10075806:d=2021032312:SNOD:surface:24 hour fcst: -580:10090152:d=2021032312:PEVPR:surface:24 hour fcst: -581:10105802:d=2021032312:ICETK:surface:24 hour fcst: -582:10109216:d=2021032312:TMP:2 m above ground:24 hour fcst: -583:10129976:d=2021032312:SPFH:2 m above ground:24 hour fcst: -584:10158843:d=2021032312:DPT:2 m above ground:24 hour fcst: -585:10180082:d=2021032312:RH:2 m above ground:24 hour fcst: -586:10198619:d=2021032312:APTMP:2 m above ground:24 hour fcst: -587:10212997:d=2021032312:TMAX:2 m above ground:18-24 hour max fcst: -588:10233693:d=2021032312:TMIN:2 m above ground:18-24 hour min fcst: -589:10254350:d=2021032312:UGRD:10 m above ground:24 hour fcst: -590:10274865:d=2021032312:VGRD:10 m above ground:24 hour fcst: -591:10295906:d=2021032312:ICEG:10 m above mean sea level:24 hour fcst: -592:10296553:d=2021032312:CPOFP:surface:24 hour fcst: -593:10316660:d=2021032312:CPRAT:surface:24 hour fcst: -594:10337961:d=2021032312:PRATE:surface:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_19 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_19 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_19 pgb2file_f024_19 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 594 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=595 -+ exglobal_atmos_products.sh[101]last=625 -+ exglobal_atmos_products.sh[102][[ 625 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 625 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 20 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 595:625 -grib tmpfile_f024_20 -595:10364978:d=2021032312:CPRAT:surface:18-24 hour ave fcst: -596:10389032:d=2021032312:PRATE:surface:18-24 hour ave fcst: -597:10414090:d=2021032312:APCP:surface:18-24 hour acc fcst: -598:10424308:d=2021032312:APCP:surface:0-1 day acc fcst: -599:10437277:d=2021032312:ACPCP:surface:18-24 hour acc fcst: -600:10445616:d=2021032312:ACPCP:surface:0-1 day acc fcst: -601:10456608:d=2021032312:FROZR:surface:0-1 day acc fcst: -602:10462909:d=2021032312:FRZR:surface:0-1 day acc fcst: -603:10466012:d=2021032312:TSNOWP:surface:0-1 day acc fcst: -604:10473558:d=2021032312:WATR:surface:18-24 hour acc fcst: -605:10485969:d=2021032312:FROZR:surface:18-24 hour acc fcst: -606:10489929:d=2021032312:FRZR:surface:18-24 hour acc fcst: -607:10491702:d=2021032312:TSNOWP:surface:18-24 hour acc fcst: -608:10499628:d=2021032312:CSNOW:surface:24 hour fcst: -609:10501408:d=2021032312:CICEP:surface:24 hour fcst: -610:10501587:d=2021032312:CFRZR:surface:24 hour fcst: -611:10501807:d=2021032312:CRAIN:surface:24 hour fcst: -612:10505342:d=2021032312:LHTFL:surface:18-24 hour ave fcst: -613:10526986:d=2021032312:SHTFL:surface:18-24 hour ave fcst: -614:10547647:d=2021032312:GFLUX:surface:18-24 hour ave fcst: -615:10559705:d=2021032312:UFLX:surface:18-24 hour ave fcst: -616:10575948:d=2021032312:VFLX:surface:18-24 hour ave fcst: -617:10592311:d=2021032312:SFCR:surface:24 hour fcst: -618:10610777:d=2021032312:FRICV:surface:24 hour fcst: -619:10632481:d=2021032312:U-GWD:surface:18-24 hour ave fcst: -620:10647240:d=2021032312:V-GWD:surface:18-24 hour ave fcst: -621:10662023:d=2021032312:VEG:surface:24 hour fcst: -622:10670047:d=2021032312:SOTYP:surface:24 hour fcst: -623:10681096:d=2021032312:WILT:surface:24 hour fcst: -624:10691006:d=2021032312:FLDCP:surface:24 hour fcst: -625:10701212:d=2021032312:SUNSD:surface:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_20 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_20 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_20 pgb2file_f024_20 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 625 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=626 -+ exglobal_atmos_products.sh[101]last=656 -+ exglobal_atmos_products.sh[102][[ 656 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 656 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 21 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 626:656 -grib tmpfile_f024_21 -626:10718310:d=2021032312:LFTX:surface:24 hour fcst: -627:10732104:d=2021032312:CAPE:surface:24 hour fcst: -628:10747196:d=2021032312:CIN:surface:24 hour fcst: -629:10760182:d=2021032312:PWAT:entire atmosphere (considered as a single layer):24 hour fcst: -630:10787083:d=2021032312:CWAT:entire atmosphere (considered as a single layer):24 hour fcst: -631:10813548:d=2021032312:RH:entire atmosphere (considered as a single layer):24 hour fcst: -632:10828691:d=2021032312:TOZNE:entire atmosphere (considered as a single layer):24 hour fcst: -633:10843042:d=2021032312:LCDC:low cloud layer:24 hour fcst: -634:10860751:d=2021032312:LCDC:low cloud layer:18-24 hour ave fcst: -635:10880705:d=2021032312:MCDC:middle cloud layer:24 hour fcst: -636:10896541:d=2021032312:MCDC:middle cloud layer:18-24 hour ave fcst: -637:10913809:d=2021032312:HCDC:high cloud layer:24 hour fcst: -638:10928868:d=2021032312:HCDC:high cloud layer:18-24 hour ave fcst: -639:10946967:d=2021032312:TCDC:entire atmosphere (considered as a single layer):24 hour fcst: -640:10964470:d=2021032312:TCDC:entire atmosphere (considered as a single layer):18-24 hour ave fcst: -641:10984553:d=2021032312:HGT:cloud ceiling:24 hour fcst: -642:11017071:d=2021032312:PRES:convective cloud bottom level:24 hour fcst: -643:11030725:d=2021032312:PRES:low cloud bottom level:18-24 hour ave fcst: -644:11053565:d=2021032312:PRES:middle cloud bottom level:18-24 hour ave fcst: -645:11071292:d=2021032312:PRES:high cloud bottom level:18-24 hour ave fcst: -646:11090184:d=2021032312:PRES:convective cloud top level:24 hour fcst: -647:11105519:d=2021032312:PRES:low cloud top level:18-24 hour ave fcst: -648:11128533:d=2021032312:PRES:middle cloud top level:18-24 hour ave fcst: -649:11146244:d=2021032312:PRES:high cloud top level:18-24 hour ave fcst: -650:11165466:d=2021032312:TMP:low cloud top level:18-24 hour ave fcst: -651:11182622:d=2021032312:TMP:middle cloud top level:18-24 hour ave fcst: -652:11196668:d=2021032312:TMP:high cloud top level:18-24 hour ave fcst: -653:11212629:d=2021032312:TCDC:convective cloud layer:24 hour fcst: -654:11233447:d=2021032312:TCDC:boundary layer cloud layer:18-24 hour ave fcst: -655:11251362:d=2021032312:CWORK:entire atmosphere (considered as a single layer):18-24 hour ave fcst: -656:11265061:d=2021032312:DSWRF:surface:18-24 hour ave fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_21 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_21 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_21 pgb2file_f024_21 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 656 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=657 -+ exglobal_atmos_products.sh[101]last=687 -+ exglobal_atmos_products.sh[102][[ 687 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 687 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -687:11988855:d=2021032312:UGRD:30 m above ground:24 hour fcst: -+ exglobal_atmos_products.sh[112]rc=0 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 0 == 0 ]] -+ exglobal_atmos_products.sh[115]last=688 -+ exglobal_atmos_products.sh[117][[ 22 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 657:688 -grib tmpfile_f024_22 -657:11287340:d=2021032312:DLWRF:surface:18-24 hour ave fcst: -658:11308610:d=2021032312:USWRF:surface:18-24 hour ave fcst: -659:11328394:d=2021032312:ULWRF:surface:18-24 hour ave fcst: -660:11347148:d=2021032312:USWRF:top of atmosphere:18-24 hour ave fcst: -661:11369737:d=2021032312:ULWRF:top of atmosphere:18-24 hour ave fcst: -662:11390500:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Total Aerosol:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -663:11418042:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Dust Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -664:11436183:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Sea Salt Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -665:11463602:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Sulphate Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -666:11489263:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Particulate Organic Matter Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -667:11514832:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Black Carbon Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -668:11537111:d=2021032312:HLCY:3000-0 m above ground:24 hour fcst: -669:11553003:d=2021032312:HLCY:1000-0 m above ground:24 hour fcst: -670:11575283:d=2021032312:USTM:6000-0 m above ground:24 hour fcst: -671:11598509:d=2021032312:VSTM:6000-0 m above ground:24 hour fcst: -672:11621328:d=2021032312:PRES:tropopause:24 hour fcst: -673:11651120:d=2021032312:ICAHT:tropopause:24 hour fcst: -674:11681896:d=2021032312:HGT:tropopause:24 hour fcst: -675:11710091:d=2021032312:TMP:tropopause:24 hour fcst: -676:11731657:d=2021032312:UGRD:tropopause:24 hour fcst: -677:11747644:d=2021032312:VGRD:tropopause:24 hour fcst: -678:11763488:d=2021032312:VWSH:tropopause:24 hour fcst: -679:11782162:d=2021032312:PRES:max wind:24 hour fcst: -680:11816448:d=2021032312:ICAHT:max wind:24 hour fcst: -681:11851472:d=2021032312:HGT:max wind:24 hour fcst: -682:11886669:d=2021032312:UGRD:max wind:24 hour fcst: -683:11903155:d=2021032312:VGRD:max wind:24 hour fcst: -684:11920498:d=2021032312:TMP:max wind:24 hour fcst: -685:11947233:d=2021032312:UGRD:20 m above ground:24 hour fcst: -686:11967757:d=2021032312:VGRD:20 m above ground:24 hour fcst: -687:11988855:d=2021032312:UGRD:30 m above ground:24 hour fcst: -688:12009741:d=2021032312:VGRD:30 m above ground:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_22 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_22 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_22 pgb2file_f024_22 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 688 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=689 -+ exglobal_atmos_products.sh[101]last=719 -+ exglobal_atmos_products.sh[102][[ 719 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 719 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 23 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 689:719 -grib tmpfile_f024_23 -689:12031034:d=2021032312:UGRD:40 m above ground:24 hour fcst: -690:12052020:d=2021032312:VGRD:40 m above ground:24 hour fcst: -691:12073449:d=2021032312:UGRD:50 m above ground:24 hour fcst: -692:12094543:d=2021032312:VGRD:50 m above ground:24 hour fcst: -693:12116070:d=2021032312:TMP:80 m above ground:24 hour fcst: -694:12136849:d=2021032312:SPFH:80 m above ground:24 hour fcst: -695:12165496:d=2021032312:PRES:80 m above ground:24 hour fcst: -696:12191151:d=2021032312:UGRD:80 m above ground:24 hour fcst: -697:12212408:d=2021032312:VGRD:80 m above ground:24 hour fcst: -698:12234114:d=2021032312:TMP:100 m above ground:24 hour fcst: -699:12254882:d=2021032312:UGRD:100 m above ground:24 hour fcst: -700:12276209:d=2021032312:VGRD:100 m above ground:24 hour fcst: -701:12297971:d=2021032312:TMP:1829 m above mean sea level:24 hour fcst: -702:12319625:d=2021032312:UGRD:1829 m above mean sea level:24 hour fcst: -703:12341842:d=2021032312:VGRD:1829 m above mean sea level:24 hour fcst: -704:12364685:d=2021032312:TMP:2743 m above mean sea level:24 hour fcst: -705:12386155:d=2021032312:UGRD:2743 m above mean sea level:24 hour fcst: -706:12408829:d=2021032312:VGRD:2743 m above mean sea level:24 hour fcst: -707:12432150:d=2021032312:TMP:3658 m above mean sea level:24 hour fcst: -708:12453447:d=2021032312:UGRD:3658 m above mean sea level:24 hour fcst: -709:12476640:d=2021032312:VGRD:3658 m above mean sea level:24 hour fcst: -710:12500418:d=2021032312:HGT:0C isotherm:24 hour fcst: -711:12522143:d=2021032312:RH:0C isotherm:24 hour fcst: -712:12541422:d=2021032312:HGT:highest tropospheric freezing level:24 hour fcst: -713:12562782:d=2021032312:RH:highest tropospheric freezing level:24 hour fcst: -714:12582021:d=2021032312:TMP:30-0 mb above ground:24 hour fcst: -715:12602836:d=2021032312:RH:30-0 mb above ground:24 hour fcst: -716:12621052:d=2021032312:SPFH:30-0 mb above ground:24 hour fcst: -717:12649817:d=2021032312:UGRD:30-0 mb above ground:24 hour fcst: -718:12670937:d=2021032312:VGRD:30-0 mb above ground:24 hour fcst: -719:12692727:d=2021032312:4LFTX:surface:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_23 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_23 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_23 pgb2file_f024_23 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 719 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=720 -+ exglobal_atmos_products.sh[101]last=750 -+ exglobal_atmos_products.sh[102][[ 750 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 750 tmpfile_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 24 -eq 24 ]] -+ exglobal_atmos_products.sh[118]last=753 -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f024 -for 720:753 -grib tmpfile_f024_24 -720:12706224:d=2021032312:CAPE:180-0 mb above ground:24 hour fcst: -721:12720224:d=2021032312:CIN:180-0 mb above ground:24 hour fcst: -722:12734336:d=2021032312:HPBL:surface:24 hour fcst: -723:12765490:d=2021032312:RH:0.33-1 sigma layer:24 hour fcst: -724:12783866:d=2021032312:RH:0.44-1 sigma layer:24 hour fcst: -725:12801868:d=2021032312:RH:0.72-0.94 sigma layer:24 hour fcst: -726:12820345:d=2021032312:RH:0.44-0.72 sigma layer:24 hour fcst: -727:12839653:d=2021032312:TMP:0.995 sigma level:24 hour fcst: -728:12860491:d=2021032312:POT:0.995 sigma level:24 hour fcst: -729:12881312:d=2021032312:RH:0.995 sigma level:24 hour fcst: -730:12899109:d=2021032312:UGRD:0.995 sigma level:24 hour fcst: -731:12920268:d=2021032312:VGRD:0.995 sigma level:24 hour fcst: -732:12941785:d=2021032312:VVEL:0.995 sigma level:24 hour fcst: -733:12962668:d=2021032312:CAPE:90-0 mb above ground:24 hour fcst: -734:12975479:d=2021032312:CIN:90-0 mb above ground:24 hour fcst: -735:12989845:d=2021032312:CAPE:255-0 mb above ground:24 hour fcst: -736:13002963:d=2021032312:CIN:255-0 mb above ground:24 hour fcst: -737:13014322:d=2021032312:PLPL:255-0 mb above ground:24 hour fcst: -738:13038043:d=2021032312:LAND:surface:24 hour fcst: -739:13039984:d=2021032312:ICEC:surface:24 hour fcst: -740:13043149:d=2021032312:ALBDO:surface:18-24 hour ave fcst: -741:13057911:d=2021032312:ICETMP:surface:24 hour fcst: -742:13064641:d=2021032312:UGRD:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -743:13075364:d=2021032312:VGRD:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -744:13086077:d=2021032312:TMP:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -745:13100452:d=2021032312:HGT:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -746:13117825:d=2021032312:PRES:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -747:13136259:d=2021032312:VWSH:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -748:13148216:d=2021032312:UGRD:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -749:13159091:d=2021032312:VGRD:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -750:13173901:d=2021032312:TMP:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -751:13188317:d=2021032312:HGT:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -752:13205819:d=2021032312:PRES:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -753:13222792:d=2021032312:VWSH:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f024_24 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f024_24 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_24 pgb2file_f024_24 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 753 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( pproc = iproc+1 )) -+ exglobal_atmos_products.sh[134](( pproc < nproc )) -+ exglobal_atmos_products.sh[137]break -+ exglobal_atmos_products.sh[142]/work2/noaa/global/mterry/global-workflow_forked/ush/run_mpmd.sh /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601/poescript -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ run_mpmd.sh[31]source /work2/noaa/global/mterry/global-workflow_forked/ush/preamble.sh -++ preamble.sh[20]set +x -Begin run_mpmd.sh at Tue Jul 29 02:24:45 UTC 2025 -+ run_mpmd.sh[33]cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601/poescript -+ run_mpmd.sh[36][[ YES != \Y\E\S ]] -+ run_mpmd.sh[46]export OMP_NUM_THREADS=1 -+ run_mpmd.sh[46]OMP_NUM_THREADS=1 -++ run_mpmd.sh[49]wc -l -+ run_mpmd.sh[49]nprocs=24 -+ run_mpmd.sh[52]mpmd_cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601/mpmd_cmdfile -+ run_mpmd.sh[53][[ -s /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601/mpmd_cmdfile ]] -+ run_mpmd.sh[55]cat - INFO: Executing MPMD job, STDOUT redirected for each process separately - INFO: On failure, logs for each job will be available in /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601/mpmd.proc_num.out - INFO: The proc_num corresponds to the line in '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601/mpmd_cmdfile' -+ run_mpmd.sh[61][[ srun -l --export=ALL --hint=nomultithread =~ ^srun.* ]] -+ run_mpmd.sh[65]nm=0 -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '0 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_1 pgb2file_f024_1 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '1 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_2 pgb2file_f024_2 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '2 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_3 pgb2file_f024_3 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '3 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_4 pgb2file_f024_4 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '4 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_5 pgb2file_f024_5 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '5 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_6 pgb2file_f024_6 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '6 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_7 pgb2file_f024_7 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '7 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_8 pgb2file_f024_8 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '8 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_9 pgb2file_f024_9 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '9 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_10 pgb2file_f024_10 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '10 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_11 pgb2file_f024_11 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '11 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_12 pgb2file_f024_12 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '12 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_13 pgb2file_f024_13 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '13 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_14 pgb2file_f024_14 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '14 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_15 pgb2file_f024_15 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '15 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_16 pgb2file_f024_16 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '16 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_17 pgb2file_f024_17 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '17 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_18 pgb2file_f024_18 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '18 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_19 pgb2file_f024_19 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '19 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_20 pgb2file_f024_20 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '20 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_21 pgb2file_f024_21 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '21 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_22 pgb2file_f024_22 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '22 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_23 pgb2file_f024_23 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '23 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f024_24 pgb2file_f024_24 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[72]set +e -+ run_mpmd.sh[74]srun -l --export=ALL --hint=nomultithread --multi-prog --output=mpmd.%j.%t.out -n 24 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601/mpmd_cmdfile -+ run_mpmd.sh[75]err=0 -+ run_mpmd.sh[76]set_strict -+ preamble.sh[35][[ YES == \Y\E\S ]] -+ preamble.sh[37]set -eu -+ run_mpmd.sh[103][[ 0 -eq 0 ]] -+ run_mpmd.sh[104]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601/mpmd_cmdfile -++ run_mpmd.sh[105]find . -name 'mpmd.*.out' -+ run_mpmd.sh[105]out_files='./mpmd.5951061.17.out -./mpmd.5951061.0.out -./mpmd.5951061.6.out -./mpmd.5951061.15.out -./mpmd.5951061.9.out -./mpmd.5951061.2.out -./mpmd.5951061.3.out -./mpmd.5951061.23.out -./mpmd.5951061.4.out -./mpmd.5951061.22.out -./mpmd.5951061.1.out -./mpmd.5951061.7.out -./mpmd.5951061.21.out -./mpmd.5951061.12.out -./mpmd.5951061.19.out -./mpmd.5951061.8.out -./mpmd.5951061.16.out -./mpmd.5951061.18.out -./mpmd.5951061.13.out -./mpmd.5951061.10.out -./mpmd.5951061.20.out -./mpmd.5951061.14.out -./mpmd.5951061.5.out -./mpmd.5951061.11.out' -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.17.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.17.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.0.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.0.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.6.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.6.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.15.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.15.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.9.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.9.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.2.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.2.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.3.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.3.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.23.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.23.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.4.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.4.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.22.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.22.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.1.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.1.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.7.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.7.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.21.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.21.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.12.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.12.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.19.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.19.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.8.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.8.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.16.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.16.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.18.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.18.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.13.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.13.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.10.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.10.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.20.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.20.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.14.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.14.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.5.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.5.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.11.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.11.out -+ run_mpmd.sh[110]cat mpmd.out -17: + bash[8]'[' -z '' ']' -17: + bash[9]case "$-" in -17: + bash[12]__lmod_vx=x -17: + bash[16]'[' -n x ']' -17: + bash[16]set +x -17: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -17: Shell debugging restarted -17: + bash[224]unset __lmod_vx -17: + interp_atmos_master.sh[7]input_file=tmpfile_f024_18 -17: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_18 -17: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -17: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -17: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -17: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -17: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -17: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -17: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -17: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -17: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -17: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -17: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -17: + interp_atmos_master.sh[31]IFS=: -17: + interp_atmos_master.sh[31]read -ra grids -17: + interp_atmos_master.sh[33]output_grids= -17: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -17: + interp_atmos_master.sh[35]gridopt=grid0p25 -17: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_18_0p25' -17: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -17: + interp_atmos_master.sh[35]gridopt=grid0p50 -17: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_18_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_18_0p50' -17: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -17: + interp_atmos_master.sh[35]gridopt=grid1p00 -17: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_18_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_18_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_18_1p00' -17: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_18 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_18_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_18_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_18_1p00 -17: 1:0:d=2021032312:UGRD:975 mb:24 hour fcst: -17: 2:21394:d=2021032312:VGRD:975 mb:24 hour fcst: -17: 3:43226:d=2021032312:ABSV:975 mb:24 hour fcst: -17: 4:64220:d=2021032312:CLMR:975 mb:24 hour fcst: -17: 5:68922:d=2021032312:ICMR:975 mb:24 hour fcst: -17: 6:74170:d=2021032312:RWMR:975 mb:24 hour fcst: -17: 7:85981:d=2021032312:SNMR:975 mb:24 hour fcst: -17: 8:93653:d=2021032312:GRLE:975 mb:24 hour fcst: -17: 9:96348:d=2021032312:O3MR:975 mb:24 hour fcst: -17: 10:122238:d=2021032312:TMP:1000 mb:24 hour fcst: -17: 11:143013:d=2021032312:RH:1000 mb:24 hour fcst: -17: 12:161899:d=2021032312:TCDC:1000 mb:24 hour fcst: -17: 13:166478:d=2021032312:SPFH:1000 mb:24 hour fcst: -17: 14:193568:d=2021032312:VVEL:1000 mb:24 hour fcst: -17: 15:214867:d=2021032312:DZDT:1000 mb:24 hour fcst: -17: 16:236215:d=2021032312:UGRD:1000 mb:24 hour fcst: -17: 17:257004:d=2021032312:VGRD:1000 mb:24 hour fcst: -17: 18:278395:d=2021032312:ABSV:1000 mb:24 hour fcst: -17: 19:298927:d=2021032312:CLMR:1000 mb:24 hour fcst: -17: 20:300842:d=2021032312:ICMR:1000 mb:24 hour fcst: -17: 21:303362:d=2021032312:RWMR:1000 mb:24 hour fcst: -17: 22:313071:d=2021032312:SNMR:1000 mb:24 hour fcst: -17: 23:316390:d=2021032312:GRLE:1000 mb:24 hour fcst: -17: 24:317425:d=2021032312:O3MR:1000 mb:24 hour fcst: -17: 25:343357:d=2021032312:MSLET:mean sea level:24 hour fcst: -17: 26:369965:d=2021032312:HGT:1000 mb:24 hour fcst: -17: 27:395682:d=2021032312:SDEN:surface:24 hour fcst: -17: 28:406642:d=2021032312:REFD:4000 m above ground:24 hour fcst: -17: 29:419670:d=2021032312:REFD:1000 m above ground:24 hour fcst: -17: 30:443873:d=2021032312:PRES:surface:24 hour fcst: -17: 31:469522:d=2021032312:HGT:surface:24 hour fcst: -17: + interp_atmos_master.sh[47]export err=0 -17: + interp_atmos_master.sh[47]err=0 -17: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -17: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -17: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_18_0p25 -17: + product_functions.sh[5]local filename=pgb2file_f024_18_0p25 -17: + product_functions.sh[6]wgrib2 pgb2file_f024_18_0p25 -not_if :RH: -grib pgb2file_f024_18_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_18_0p25.new -17: 1:0:d=2021032312:UGRD:975 mb:24 hour fcst: -17: 2:518253:d=2021032312:VGRD:975 mb:24 hour fcst: -17: 3:1060698:d=2021032312:ABSV:975 mb:24 hour fcst: -17: 4:1560755:d=2021032312:CLMR:975 mb:24 hour fcst: -17: 5:1662491:d=2021032312:ICMR:975 mb:24 hour fcst: -17: 6:1780478:d=2021032312:RWMR:975 mb:24 hour fcst: -17: 7:2024762:d=2021032312:SNMR:975 mb:24 hour fcst: -17: 8:2201776:d=2021032312:GRLE:975 mb:24 hour fcst: -17: 9:2253784:d=2021032312:O3MR:975 mb:24 hour fcst: -17: 10:2952305:d=2021032312:TMP:1000 mb:24 hour fcst: -17: 11:3435040:d=2021032312:RH:1000 mb:24 hour fcst: -17: 12:3855513:d=2021032312:TCDC:1000 mb:24 hour fcst: -17: 13:3935696:d=2021032312:SPFH:1000 mb:24 hour fcst: -17: 14:4667142:d=2021032312:VVEL:1000 mb:24 hour fcst: -17: 15:5182686:d=2021032312:DZDT:1000 mb:24 hour fcst: -17: 16:5697038:d=2021032312:UGRD:1000 mb:24 hour fcst: -17: 17:6195284:d=2021032312:VGRD:1000 mb:24 hour fcst: -17: 18:6717037:d=2021032312:ABSV:1000 mb:24 hour fcst: -17: 19:7200402:d=2021032312:CLMR:1000 mb:24 hour fcst: -17: 20:7241998:d=2021032312:ICMR:1000 mb:24 hour fcst: -17: 21:7295014:d=2021032312:RWMR:1000 mb:24 hour fcst: -17: 22:7493234:d=2021032312:SNMR:1000 mb:24 hour fcst: -17: 23:7561708:d=2021032312:GRLE:1000 mb:24 hour fcst: -17: 24:7578341:d=2021032312:O3MR:1000 mb:24 hour fcst: -17: 25:8275183:d=2021032312:MSLET:mean sea level:24 hour fcst: -17: 26:8996164:d=2021032312:HGT:1000 mb:24 hour fcst: -17: 27:9682938:d=2021032312:SDEN:surface:24 hour fcst: -17: 28:10021269:d=2021032312:REFD:4000 m above ground:24 hour fcst: -17: 29:10299312:d=2021032312:REFD:1000 m above ground:24 hour fcst: -17: 30:10866074:d=2021032312:PRES:surface:24 hour fcst: -17: 31:11510399:d=2021032312:HGT:surface:24 hour fcst: -17: + product_functions.sh[10]rc=0 -17: + product_functions.sh[11](( rc == 0 )) -17: + product_functions.sh[11]mv pgb2file_f024_18_0p25.new pgb2file_f024_18_0p25 -17: + product_functions.sh[12]return 0 -17: + interp_atmos_master.sh[56]export err=0 -17: + interp_atmos_master.sh[56]err=0 -17: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -17: ++ interp_atmos_master.sh[62]wc -l -17: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_18_0p25 -match 'LAND|ICEC' -17: + interp_atmos_master.sh[62]var_count=0 -17: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -17: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -17: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_18_0p50 -17: + product_functions.sh[5]local filename=pgb2file_f024_18_0p50 -17: + product_functions.sh[6]wgrib2 pgb2file_f024_18_0p50 -not_if :RH: -grib pgb2file_f024_18_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_18_0p50.new -17: 1:0:d=2021032312:UGRD:975 mb:24 hour fcst: -17: 2:178421:d=2021032312:VGRD:975 mb:24 hour fcst: -17: 3:362543:d=2021032312:ABSV:975 mb:24 hour fcst: -17: 4:535325:d=2021032312:CLMR:975 mb:24 hour fcst: -17: 5:572838:d=2021032312:ICMR:975 mb:24 hour fcst: -17: 6:615082:d=2021032312:RWMR:975 mb:24 hour fcst: -17: 7:703635:d=2021032312:SNMR:975 mb:24 hour fcst: -17: 8:765954:d=2021032312:GRLE:975 mb:24 hour fcst: -17: 9:784905:d=2021032312:O3MR:975 mb:24 hour fcst: -17: 10:1019916:d=2021032312:TMP:1000 mb:24 hour fcst: -17: 11:1187393:d=2021032312:RH:1000 mb:24 hour fcst: -17: 12:1335195:d=2021032312:TCDC:1000 mb:24 hour fcst: -17: 13:1365478:d=2021032312:SPFH:1000 mb:24 hour fcst: -17: 14:1611960:d=2021032312:VVEL:1000 mb:24 hour fcst: -17: 15:1788684:d=2021032312:DZDT:1000 mb:24 hour fcst: -17: 16:1962260:d=2021032312:UGRD:1000 mb:24 hour fcst: -17: 17:2134710:d=2021032312:VGRD:1000 mb:24 hour fcst: -17: 18:2313955:d=2021032312:ABSV:1000 mb:24 hour fcst: -17: 19:2479822:d=2021032312:CLMR:1000 mb:24 hour fcst: -17: 20:2495038:d=2021032312:ICMR:1000 mb:24 hour fcst: -17: 21:2513824:d=2021032312:RWMR:1000 mb:24 hour fcst: -17: 22:2585253:d=2021032312:SNMR:1000 mb:24 hour fcst: -17: 23:2609942:d=2021032312:GRLE:1000 mb:24 hour fcst: -17: 24:2616331:d=2021032312:O3MR:1000 mb:24 hour fcst: -17: 25:2851065:d=2021032312:MSLET:mean sea level:24 hour fcst: -17: 26:3092643:d=2021032312:HGT:1000 mb:24 hour fcst: -17: 27:3321994:d=2021032312:SDEN:surface:24 hour fcst: -17: 28:3424490:d=2021032312:REFD:4000 m above ground:24 hour fcst: -17: 29:3530321:d=2021032312:REFD:1000 m above ground:24 hour fcst: -17: 30:3746310:d=2021032312:PRES:surface:24 hour fcst: -17: 31:3967680:d=2021032312:HGT:surface:24 hour fcst: -17: + product_functions.sh[10]rc=0 -17: + product_functions.sh[11](( rc == 0 )) -17: + product_functions.sh[11]mv pgb2file_f024_18_0p50.new pgb2file_f024_18_0p50 -17: + product_functions.sh[12]return 0 -17: + interp_atmos_master.sh[56]export err=0 -17: + interp_atmos_master.sh[56]err=0 -17: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -17: ++ interp_atmos_master.sh[62]wc -l -17: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_18_0p50 -match 'LAND|ICEC' -17: + interp_atmos_master.sh[62]var_count=0 -17: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -17: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -17: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_18_1p00 -17: + product_functions.sh[5]local filename=pgb2file_f024_18_1p00 -17: + product_functions.sh[6]wgrib2 pgb2file_f024_18_1p00 -not_if :RH: -grib pgb2file_f024_18_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_18_1p00.new -17: 1:0:d=2021032312:UGRD:975 mb:24 hour fcst: -17: 2:57185:d=2021032312:VGRD:975 mb:24 hour fcst: -17: 3:115804:d=2021032312:ABSV:975 mb:24 hour fcst: -17: 4:171758:d=2021032312:CLMR:975 mb:24 hour fcst: -17: 5:184761:d=2021032312:ICMR:975 mb:24 hour fcst: -17: 6:199260:d=2021032312:RWMR:975 mb:24 hour fcst: -17: 7:229835:d=2021032312:SNMR:975 mb:24 hour fcst: -17: 8:250785:d=2021032312:GRLE:975 mb:24 hour fcst: -17: 9:257732:d=2021032312:O3MR:975 mb:24 hour fcst: -17: 10:329701:d=2021032312:TMP:1000 mb:24 hour fcst: -17: 11:384172:d=2021032312:RH:1000 mb:24 hour fcst: -17: 12:432814:d=2021032312:TCDC:1000 mb:24 hour fcst: -17: 13:444557:d=2021032312:SPFH:1000 mb:24 hour fcst: -17: 14:519263:d=2021032312:VVEL:1000 mb:24 hour fcst: -17: 15:575550:d=2021032312:DZDT:1000 mb:24 hour fcst: -17: 16:630775:d=2021032312:UGRD:1000 mb:24 hour fcst: -17: 17:686305:d=2021032312:VGRD:1000 mb:24 hour fcst: -17: 18:743386:d=2021032312:ABSV:1000 mb:24 hour fcst: -17: 19:797483:d=2021032312:CLMR:1000 mb:24 hour fcst: -17: 20:802946:d=2021032312:ICMR:1000 mb:24 hour fcst: -17: 21:809482:d=2021032312:RWMR:1000 mb:24 hour fcst: -17: 22:834483:d=2021032312:SNMR:1000 mb:24 hour fcst: -17: 23:843506:d=2021032312:GRLE:1000 mb:24 hour fcst: -17: 24:846063:d=2021032312:O3MR:1000 mb:24 hour fcst: -17: 25:917987:d=2021032312:MSLET:mean sea level:24 hour fcst: -17: 26:991912:d=2021032312:HGT:1000 mb:24 hour fcst: -17: 27:1063184:d=2021032312:SDEN:surface:24 hour fcst: -17: 28:1093653:d=2021032312:REFD:4000 m above ground:24 hour fcst: -17: 29:1128415:d=2021032312:REFD:1000 m above ground:24 hour fcst: -17: 30:1194203:d=2021032312:PRES:surface:24 hour fcst: -17: 31:1263716:d=2021032312:HGT:surface:24 hour fcst: -17: + product_functions.sh[10]rc=0 -17: + product_functions.sh[11](( rc == 0 )) -17: + product_functions.sh[11]mv pgb2file_f024_18_1p00.new pgb2file_f024_18_1p00 -17: + product_functions.sh[12]return 0 -17: + interp_atmos_master.sh[56]export err=0 -17: + interp_atmos_master.sh[56]err=0 -17: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -17: ++ interp_atmos_master.sh[62]wc -l -17: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_18_1p00 -match 'LAND|ICEC' -17: + interp_atmos_master.sh[62]var_count=0 -17: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -17: + interp_atmos_master.sh[73]exit 0 - 0: + bash[8]'[' -z '' ']' - 0: + bash[9]case "$-" in - 0: + bash[12]__lmod_vx=x - 0: + bash[16]'[' -n x ']' - 0: + bash[16]set +x - 0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 0: Shell debugging restarted - 0: + bash[224]unset __lmod_vx - 0: + interp_atmos_master.sh[7]input_file=tmpfile_f024_1 - 0: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_1 - 0: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 - 0: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 0: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 0: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 0: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 0: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 0: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 0: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 0: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 0: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 0: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 0: + interp_atmos_master.sh[31]IFS=: - 0: + interp_atmos_master.sh[31]read -ra grids - 0: + interp_atmos_master.sh[33]output_grids= - 0: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 0: + interp_atmos_master.sh[35]gridopt=grid0p25 - 0: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_1_0p25' - 0: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 0: + interp_atmos_master.sh[35]gridopt=grid0p50 - 0: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_1_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_1_0p50' - 0: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 0: + interp_atmos_master.sh[35]gridopt=grid1p00 - 0: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_1_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_1_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_1_1p00' - 0: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_1 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_1_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_1_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_1_1p00 - 0: 1:0:d=2021032312:PRMSL:mean sea level:24 hour fcst: - 0: 2:25731:d=2021032312:CLMR:1 hybrid level:24 hour fcst: - 0: 3:29524:d=2021032312:ICMR:1 hybrid level:24 hour fcst: - 0: 4:37100:d=2021032312:RWMR:1 hybrid level:24 hour fcst: - 0: 5:48509:d=2021032312:SNMR:1 hybrid level:24 hour fcst: - 0: 6:59635:d=2021032312:GRLE:1 hybrid level:24 hour fcst: - 0: 7:62405:d=2021032312:REFD:1 hybrid level:24 hour fcst: - 0: 8:88221:d=2021032312:REFD:2 hybrid level:24 hour fcst: - 0: 9:114004:d=2021032312:REFC:entire atmosphere (considered as a single layer):24 hour fcst: - 0: 10:141439:d=2021032312:VIS:surface:24 hour fcst: - 0: 11:173907:d=2021032312:UGRD:planetary boundary layer:24 hour fcst: - 0: 12:187864:d=2021032312:VGRD:planetary boundary layer:24 hour fcst: - 0: 13:202409:d=2021032312:VRATE:planetary boundary layer:24 hour fcst: - 0: 14:217517:d=2021032312:GUST:surface:24 hour fcst: - 0: 15:232221:d=2021032312:HGT:0.01 mb:24 hour fcst: - 0: 16:257938:d=2021032312:TMP:0.01 mb:24 hour fcst: - 0: 17:268909:d=2021032312:RH:0.01 mb:24 hour fcst: - 0: 18:275412:d=2021032312:SPFH:0.01 mb:24 hour fcst: - 0: 19:298176:d=2021032312:VVEL:0.01 mb:24 hour fcst: - 0: 20:325074:d=2021032312:DZDT:0.01 mb:24 hour fcst: - 0: 21:352568:d=2021032312:UGRD:0.01 mb:24 hour fcst: - 0: 22:363684:d=2021032312:VGRD:0.01 mb:24 hour fcst: - 0: 23:374056:d=2021032312:ABSV:0.01 mb:24 hour fcst: - 0: 24:389577:d=2021032312:O3MR:0.01 mb:24 hour fcst: - 0: 25:403656:d=2021032312:HGT:0.02 mb:24 hour fcst: - 0: 26:429080:d=2021032312:TMP:0.02 mb:24 hour fcst: - 0: 27:443922:d=2021032312:RH:0.02 mb:24 hour fcst: - 0: 28:451652:d=2021032312:SPFH:0.02 mb:24 hour fcst: - 0: 29:475844:d=2021032312:VVEL:0.02 mb:24 hour fcst: - 0: 30:499814:d=2021032312:DZDT:0.02 mb:24 hour fcst: - 0: 31:527345:d=2021032312:UGRD:0.02 mb:24 hour fcst: - 0: 32:539841:d=2021032312:VGRD:0.02 mb:24 hour fcst: - 0: + interp_atmos_master.sh[47]export err=0 - 0: + interp_atmos_master.sh[47]err=0 - 0: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 0: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 0: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_1_0p25 - 0: + product_functions.sh[5]local filename=pgb2file_f024_1_0p25 - 0: + product_functions.sh[6]wgrib2 pgb2file_f024_1_0p25 -not_if :RH: -grib pgb2file_f024_1_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_1_0p25.new - 0: 1:0:d=2021032312:PRMSL:mean sea level:24 hour fcst: - 0: 2:687973:d=2021032312:CLMR:1 hybrid level:24 hour fcst: - 0: 3:771962:d=2021032312:ICMR:1 hybrid level:24 hour fcst: - 0: 4:943801:d=2021032312:RWMR:1 hybrid level:24 hour fcst: - 0: 5:1167969:d=2021032312:SNMR:1 hybrid level:24 hour fcst: - 0: 6:1427132:d=2021032312:GRLE:1 hybrid level:24 hour fcst: - 0: 7:1478938:d=2021032312:REFD:1 hybrid level:24 hour fcst: - 0: 8:2077828:d=2021032312:REFD:2 hybrid level:24 hour fcst: - 0: 9:2676683:d=2021032312:REFC:entire atmosphere (considered as a single layer):24 hour fcst: - 0: 10:3330627:d=2021032312:VIS:surface:24 hour fcst: - 0: 11:4211026:d=2021032312:UGRD:planetary boundary layer:24 hour fcst: - 0: 12:4509991:d=2021032312:VGRD:planetary boundary layer:24 hour fcst: - 0: 13:4819996:d=2021032312:VRATE:planetary boundary layer:24 hour fcst: - 0: 14:5134445:d=2021032312:GUST:surface:24 hour fcst: - 0: 15:5443360:d=2021032312:HGT:0.01 mb:24 hour fcst: - 0: 16:6145924:d=2021032312:TMP:0.01 mb:24 hour fcst: - 0: 17:6415808:d=2021032312:RH:0.01 mb:24 hour fcst: - 0: 18:6604460:d=2021032312:SPFH:0.01 mb:24 hour fcst: - 0: 19:7200974:d=2021032312:VVEL:0.01 mb:24 hour fcst: - 0: 20:7942500:d=2021032312:DZDT:0.01 mb:24 hour fcst: - 0: 21:8735261:d=2021032312:UGRD:0.01 mb:24 hour fcst: - 0: 22:9011040:d=2021032312:VGRD:0.01 mb:24 hour fcst: - 0: 23:9278529:d=2021032312:ABSV:0.01 mb:24 hour fcst: - 0: 24:9603568:d=2021032312:O3MR:0.01 mb:24 hour fcst: - 0: 25:9904096:d=2021032312:HGT:0.02 mb:24 hour fcst: - 0: 26:10599745:d=2021032312:TMP:0.02 mb:24 hour fcst: - 0: 27:10944033:d=2021032312:RH:0.02 mb:24 hour fcst: - 0: 28:11130783:d=2021032312:SPFH:0.02 mb:24 hour fcst: - 0: 29:11778266:d=2021032312:VVEL:0.02 mb:24 hour fcst: - 0: 30:12413179:d=2021032312:DZDT:0.02 mb:24 hour fcst: - 0: 31:13199765:d=2021032312:UGRD:0.02 mb:24 hour fcst: - 0: 32:13489974:d=2021032312:VGRD:0.02 mb:24 hour fcst: - 0: + product_functions.sh[10]rc=0 - 0: + product_functions.sh[11](( rc == 0 )) - 0: + product_functions.sh[11]mv pgb2file_f024_1_0p25.new pgb2file_f024_1_0p25 - 0: + product_functions.sh[12]return 0 - 0: + interp_atmos_master.sh[56]export err=0 - 0: + interp_atmos_master.sh[56]err=0 - 0: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 0: ++ interp_atmos_master.sh[62]wc -l - 0: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_1_0p25 -match 'LAND|ICEC' - 0: + interp_atmos_master.sh[62]var_count=0 - 0: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 0: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 0: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_1_0p50 - 0: + product_functions.sh[5]local filename=pgb2file_f024_1_0p50 - 0: + product_functions.sh[6]wgrib2 pgb2file_f024_1_0p50 -not_if :RH: -grib pgb2file_f024_1_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_1_0p50.new - 0: 1:0:d=2021032312:PRMSL:mean sea level:24 hour fcst: - 0: 2:229976:d=2021032312:CLMR:1 hybrid level:24 hour fcst: - 0: 3:260722:d=2021032312:ICMR:1 hybrid level:24 hour fcst: - 0: 4:320197:d=2021032312:RWMR:1 hybrid level:24 hour fcst: - 0: 5:400546:d=2021032312:SNMR:1 hybrid level:24 hour fcst: - 0: 6:491036:d=2021032312:GRLE:1 hybrid level:24 hour fcst: - 0: 7:509941:d=2021032312:REFD:1 hybrid level:24 hour fcst: - 0: 8:741433:d=2021032312:REFD:2 hybrid level:24 hour fcst: - 0: 9:972950:d=2021032312:REFC:entire atmosphere (considered as a single layer):24 hour fcst: - 0: 10:1220897:d=2021032312:VIS:surface:24 hour fcst: - 0: 11:1530972:d=2021032312:UGRD:planetary boundary layer:24 hour fcst: - 0: 12:1627049:d=2021032312:VGRD:planetary boundary layer:24 hour fcst: - 0: 13:1727920:d=2021032312:VRATE:planetary boundary layer:24 hour fcst: - 0: 14:1834516:d=2021032312:GUST:surface:24 hour fcst: - 0: 15:1935565:d=2021032312:HGT:0.01 mb:24 hour fcst: - 0: 16:2173221:d=2021032312:TMP:0.01 mb:24 hour fcst: - 0: 17:2253277:d=2021032312:RH:0.01 mb:24 hour fcst: - 0: 18:2311970:d=2021032312:SPFH:0.01 mb:24 hour fcst: - 0: 19:2512408:d=2021032312:VVEL:0.01 mb:24 hour fcst: - 0: 20:2765346:d=2021032312:DZDT:0.01 mb:24 hour fcst: - 0: 21:3022747:d=2021032312:UGRD:0.01 mb:24 hour fcst: - 0: 22:3105969:d=2021032312:VGRD:0.01 mb:24 hour fcst: - 0: 23:3183017:d=2021032312:ABSV:0.01 mb:24 hour fcst: - 0: 24:3292808:d=2021032312:O3MR:0.01 mb:24 hour fcst: - 0: 25:3388175:d=2021032312:HGT:0.02 mb:24 hour fcst: - 0: 26:3623003:d=2021032312:TMP:0.02 mb:24 hour fcst: - 0: 27:3737926:d=2021032312:RH:0.02 mb:24 hour fcst: - 0: 28:3795959:d=2021032312:SPFH:0.02 mb:24 hour fcst: - 0: 29:4014300:d=2021032312:VVEL:0.02 mb:24 hour fcst: - 0: 30:4231751:d=2021032312:DZDT:0.02 mb:24 hour fcst: - 0: 31:4490843:d=2021032312:UGRD:0.02 mb:24 hour fcst: - 0: 32:4581718:d=2021032312:VGRD:0.02 mb:24 hour fcst: - 0: + product_functions.sh[10]rc=0 - 0: + product_functions.sh[11](( rc == 0 )) - 0: + product_functions.sh[11]mv pgb2file_f024_1_0p50.new pgb2file_f024_1_0p50 - 0: + product_functions.sh[12]return 0 - 0: + interp_atmos_master.sh[56]export err=0 - 0: + interp_atmos_master.sh[56]err=0 - 0: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 0: ++ interp_atmos_master.sh[62]wc -l - 0: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_1_0p50 -match 'LAND|ICEC' - 0: + interp_atmos_master.sh[62]var_count=0 - 0: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 0: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 0: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_1_1p00 - 0: + product_functions.sh[5]local filename=pgb2file_f024_1_1p00 - 0: + product_functions.sh[6]wgrib2 pgb2file_f024_1_1p00 -not_if :RH: -grib pgb2file_f024_1_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_1_1p00.new - 0: 1:0:d=2021032312:PRMSL:mean sea level:24 hour fcst: - 0: 2:71472:d=2021032312:CLMR:1 hybrid level:24 hour fcst: - 0: 3:82123:d=2021032312:ICMR:1 hybrid level:24 hour fcst: - 0: 4:102118:d=2021032312:RWMR:1 hybrid level:24 hour fcst: - 0: 5:130461:d=2021032312:SNMR:1 hybrid level:24 hour fcst: - 0: 6:160194:d=2021032312:GRLE:1 hybrid level:24 hour fcst: - 0: 7:167053:d=2021032312:REFD:1 hybrid level:24 hour fcst: - 0: 8:238136:d=2021032312:REFD:2 hybrid level:24 hour fcst: - 0: 9:309200:d=2021032312:REFC:entire atmosphere (considered as a single layer):24 hour fcst: - 0: 10:385293:d=2021032312:VIS:surface:24 hour fcst: - 0: 11:478315:d=2021032312:UGRD:planetary boundary layer:24 hour fcst: - 0: 12:511693:d=2021032312:VGRD:planetary boundary layer:24 hour fcst: - 0: 13:546571:d=2021032312:VRATE:planetary boundary layer:24 hour fcst: - 0: 14:582971:d=2021032312:GUST:surface:24 hour fcst: - 0: 15:618067:d=2021032312:HGT:0.01 mb:24 hour fcst: - 0: 16:690284:d=2021032312:TMP:0.01 mb:24 hour fcst: - 0: 17:716292:d=2021032312:RH:0.01 mb:24 hour fcst: - 0: 18:733410:d=2021032312:SPFH:0.01 mb:24 hour fcst: - 0: 19:795630:d=2021032312:VVEL:0.01 mb:24 hour fcst: - 0: 20:871579:d=2021032312:DZDT:0.01 mb:24 hour fcst: - 0: 21:948296:d=2021032312:UGRD:0.01 mb:24 hour fcst: - 0: 22:975657:d=2021032312:VGRD:0.01 mb:24 hour fcst: - 0: 23:1000617:d=2021032312:ABSV:0.01 mb:24 hour fcst: - 0: 24:1038289:d=2021032312:O3MR:0.01 mb:24 hour fcst: - 0: 25:1071071:d=2021032312:HGT:0.02 mb:24 hour fcst: - 0: 26:1142400:d=2021032312:TMP:0.02 mb:24 hour fcst: - 0: 27:1179872:d=2021032312:RH:0.02 mb:24 hour fcst: - 0: 28:1198221:d=2021032312:SPFH:0.02 mb:24 hour fcst: - 0: 29:1264617:d=2021032312:VVEL:0.02 mb:24 hour fcst: - 0: 30:1331414:d=2021032312:DZDT:0.02 mb:24 hour fcst: - 0: 31:1408565:d=2021032312:UGRD:0.02 mb:24 hour fcst: - 0: 32:1439149:d=2021032312:VGRD:0.02 mb:24 hour fcst: - 0: + product_functions.sh[10]rc=0 - 0: + product_functions.sh[11](( rc == 0 )) - 0: + product_functions.sh[11]mv pgb2file_f024_1_1p00.new pgb2file_f024_1_1p00 - 0: + product_functions.sh[12]return 0 - 0: + interp_atmos_master.sh[56]export err=0 - 0: + interp_atmos_master.sh[56]err=0 - 0: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 0: ++ interp_atmos_master.sh[62]wc -l - 0: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_1_1p00 -match 'LAND|ICEC' - 0: + interp_atmos_master.sh[62]var_count=0 - 0: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 0: + interp_atmos_master.sh[73]exit 0 - 6: + bash[8]'[' -z '' ']' - 6: + bash[9]case "$-" in - 6: + bash[12]__lmod_vx=x - 6: + bash[16]'[' -n x ']' - 6: + bash[16]set +x - 6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 6: Shell debugging restarted - 6: + bash[224]unset __lmod_vx - 6: + interp_atmos_master.sh[7]input_file=tmpfile_f024_7 - 6: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_7 - 6: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 - 6: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 6: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 6: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 6: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 6: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 6: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 6: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 6: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 6: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 6: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 6: + interp_atmos_master.sh[31]IFS=: - 6: + interp_atmos_master.sh[31]read -ra grids - 6: + interp_atmos_master.sh[33]output_grids= - 6: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 6: + interp_atmos_master.sh[35]gridopt=grid0p25 - 6: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_7_0p25' - 6: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 6: + interp_atmos_master.sh[35]gridopt=grid0p50 - 6: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_7_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_7_0p50' - 6: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 6: + interp_atmos_master.sh[35]gridopt=grid1p00 - 6: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_7_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_7_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_7_1p00' - 6: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_7 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_7_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_7_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_7_1p00 - 6: 1:0:d=2021032312:SPFH:40 mb:24 hour fcst: - 6: 2:17084:d=2021032312:VVEL:40 mb:24 hour fcst: - 6: 3:46485:d=2021032312:DZDT:40 mb:24 hour fcst: - 6: 4:76427:d=2021032312:UGRD:40 mb:24 hour fcst: - 6: 5:95187:d=2021032312:VGRD:40 mb:24 hour fcst: - 6: 6:113917:d=2021032312:ABSV:40 mb:24 hour fcst: - 6: 7:131762:d=2021032312:O3MR:40 mb:24 hour fcst: - 6: 8:158706:d=2021032312:HGT:50 mb:24 hour fcst: - 6: 9:177611:d=2021032312:TMP:50 mb:24 hour fcst: - 6: 10:193747:d=2021032312:RH:50 mb:24 hour fcst: - 6: 11:204387:d=2021032312:TCDC:50 mb:24 hour fcst: - 6: 12:204566:d=2021032312:SPFH:50 mb:24 hour fcst: - 6: 13:223111:d=2021032312:VVEL:50 mb:24 hour fcst: - 6: 14:253206:d=2021032312:DZDT:50 mb:24 hour fcst: - 6: 15:282821:d=2021032312:UGRD:50 mb:24 hour fcst: - 6: 16:301302:d=2021032312:VGRD:50 mb:24 hour fcst: - 6: 17:319828:d=2021032312:ABSV:50 mb:24 hour fcst: - 6: 18:337523:d=2021032312:CLMR:50 mb:24 hour fcst: - 6: 19:337702:d=2021032312:ICMR:50 mb:24 hour fcst: - 6: 20:337881:d=2021032312:RWMR:50 mb:24 hour fcst: - 6: 21:338060:d=2021032312:SNMR:50 mb:24 hour fcst: - 6: 22:338239:d=2021032312:GRLE:50 mb:24 hour fcst: - 6: 23:338418:d=2021032312:O3MR:50 mb:24 hour fcst: - 6: 24:364778:d=2021032312:HGT:70 mb:24 hour fcst: - 6: 25:383858:d=2021032312:TMP:70 mb:24 hour fcst: - 6: 26:400363:d=2021032312:RH:70 mb:24 hour fcst: - 6: 27:408594:d=2021032312:SPFH:70 mb:24 hour fcst: - 6: 28:429571:d=2021032312:VVEL:70 mb:24 hour fcst: - 6: 29:460992:d=2021032312:DZDT:70 mb:24 hour fcst: - 6: 30:493095:d=2021032312:UGRD:70 mb:24 hour fcst: - 6: 31:511788:d=2021032312:VGRD:70 mb:24 hour fcst: - 6: + interp_atmos_master.sh[47]export err=0 - 6: + interp_atmos_master.sh[47]err=0 - 6: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 6: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 6: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_7_0p25 - 6: + product_functions.sh[5]local filename=pgb2file_f024_7_0p25 - 6: + product_functions.sh[6]wgrib2 pgb2file_f024_7_0p25 -not_if :RH: -grib pgb2file_f024_7_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_7_0p25.new - 6: 1:0:d=2021032312:SPFH:40 mb:24 hour fcst: - 6: 2:416261:d=2021032312:VVEL:40 mb:24 hour fcst: - 6: 3:1239618:d=2021032312:DZDT:40 mb:24 hour fcst: - 6: 4:2128119:d=2021032312:UGRD:40 mb:24 hour fcst: - 6: 5:2569963:d=2021032312:VGRD:40 mb:24 hour fcst: - 6: 6:3010058:d=2021032312:ABSV:40 mb:24 hour fcst: - 6: 7:3414942:d=2021032312:O3MR:40 mb:24 hour fcst: - 6: 8:4151470:d=2021032312:HGT:50 mb:24 hour fcst: - 6: 9:4591058:d=2021032312:TMP:50 mb:24 hour fcst: - 6: 10:4948399:d=2021032312:RH:50 mb:24 hour fcst: - 6: 11:5205739:d=2021032312:TCDC:50 mb:24 hour fcst: - 6: 12:5205918:d=2021032312:SPFH:50 mb:24 hour fcst: - 6: 13:5660935:d=2021032312:VVEL:50 mb:24 hour fcst: - 6: 14:6497674:d=2021032312:DZDT:50 mb:24 hour fcst: - 6: 15:7366167:d=2021032312:UGRD:50 mb:24 hour fcst: - 6: 16:7798668:d=2021032312:VGRD:50 mb:24 hour fcst: - 6: 17:8229010:d=2021032312:ABSV:50 mb:24 hour fcst: - 6: 18:8624249:d=2021032312:CLMR:50 mb:24 hour fcst: - 6: 19:8624428:d=2021032312:ICMR:50 mb:24 hour fcst: - 6: 20:8624607:d=2021032312:RWMR:50 mb:24 hour fcst: - 6: 21:8624786:d=2021032312:SNMR:50 mb:24 hour fcst: - 6: 22:8624965:d=2021032312:GRLE:50 mb:24 hour fcst: - 6: 23:8625144:d=2021032312:O3MR:50 mb:24 hour fcst: - 6: 24:9345575:d=2021032312:HGT:70 mb:24 hour fcst: - 6: 25:9790685:d=2021032312:TMP:70 mb:24 hour fcst: - 6: 26:10160618:d=2021032312:RH:70 mb:24 hour fcst: - 6: 27:10348825:d=2021032312:SPFH:70 mb:24 hour fcst: - 6: 28:10873266:d=2021032312:VVEL:70 mb:24 hour fcst: - 6: 29:11748490:d=2021032312:DZDT:70 mb:24 hour fcst: - 6: 30:12726604:d=2021032312:UGRD:70 mb:24 hour fcst: - 6: 31:13171691:d=2021032312:VGRD:70 mb:24 hour fcst: - 6: + product_functions.sh[10]rc=0 - 6: + product_functions.sh[11](( rc == 0 )) - 6: + product_functions.sh[11]mv pgb2file_f024_7_0p25.new pgb2file_f024_7_0p25 - 6: + product_functions.sh[12]return 0 - 6: + interp_atmos_master.sh[56]export err=0 - 6: + interp_atmos_master.sh[56]err=0 - 6: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 6: ++ interp_atmos_master.sh[62]wc -l - 6: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_7_0p25 -match 'LAND|ICEC' - 6: + interp_atmos_master.sh[62]var_count=0 - 6: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 6: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 6: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_7_0p50 - 6: + product_functions.sh[5]local filename=pgb2file_f024_7_0p50 - 6: + product_functions.sh[6]wgrib2 pgb2file_f024_7_0p50 -not_if :RH: -grib pgb2file_f024_7_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_7_0p50.new - 6: 1:0:d=2021032312:SPFH:40 mb:24 hour fcst: - 6: 2:137131:d=2021032312:VVEL:40 mb:24 hour fcst: - 6: 3:423795:d=2021032312:DZDT:40 mb:24 hour fcst: - 6: 4:710374:d=2021032312:UGRD:40 mb:24 hour fcst: - 6: 5:861485:d=2021032312:VGRD:40 mb:24 hour fcst: - 6: 6:1014363:d=2021032312:ABSV:40 mb:24 hour fcst: - 6: 7:1153899:d=2021032312:O3MR:40 mb:24 hour fcst: - 6: 8:1407932:d=2021032312:HGT:50 mb:24 hour fcst: - 6: 9:1557948:d=2021032312:TMP:50 mb:24 hour fcst: - 6: 10:1680388:d=2021032312:RH:50 mb:24 hour fcst: - 6: 11:1759944:d=2021032312:TCDC:50 mb:24 hour fcst: - 6: 12:1760123:d=2021032312:SPFH:50 mb:24 hour fcst: - 6: 13:1911572:d=2021032312:VVEL:50 mb:24 hour fcst: - 6: 14:2206334:d=2021032312:DZDT:50 mb:24 hour fcst: - 6: 15:2489813:d=2021032312:UGRD:50 mb:24 hour fcst: - 6: 16:2637744:d=2021032312:VGRD:50 mb:24 hour fcst: - 6: 17:2787406:d=2021032312:ABSV:50 mb:24 hour fcst: - 6: 18:2923705:d=2021032312:CLMR:50 mb:24 hour fcst: - 6: 19:2923884:d=2021032312:ICMR:50 mb:24 hour fcst: - 6: 20:2924063:d=2021032312:RWMR:50 mb:24 hour fcst: - 6: 21:2924242:d=2021032312:SNMR:50 mb:24 hour fcst: - 6: 22:2924421:d=2021032312:GRLE:50 mb:24 hour fcst: - 6: 23:2924600:d=2021032312:O3MR:50 mb:24 hour fcst: - 6: 24:3171174:d=2021032312:HGT:70 mb:24 hour fcst: - 6: 25:3322421:d=2021032312:TMP:70 mb:24 hour fcst: - 6: 26:3447872:d=2021032312:RH:70 mb:24 hour fcst: - 6: 27:3508437:d=2021032312:SPFH:70 mb:24 hour fcst: - 6: 28:3684041:d=2021032312:VVEL:70 mb:24 hour fcst: - 6: 29:3993147:d=2021032312:DZDT:70 mb:24 hour fcst: - 6: 30:4306896:d=2021032312:UGRD:70 mb:24 hour fcst: - 6: 31:4458910:d=2021032312:VGRD:70 mb:24 hour fcst: - 6: + product_functions.sh[10]rc=0 - 6: + product_functions.sh[11](( rc == 0 )) - 6: + product_functions.sh[11]mv pgb2file_f024_7_0p50.new pgb2file_f024_7_0p50 - 6: + product_functions.sh[12]return 0 - 6: + interp_atmos_master.sh[56]export err=0 - 6: + interp_atmos_master.sh[56]err=0 - 6: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 6: ++ interp_atmos_master.sh[62]wc -l - 6: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_7_0p50 -match 'LAND|ICEC' - 6: + interp_atmos_master.sh[62]var_count=0 - 6: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 6: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 6: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_7_1p00 - 6: + product_functions.sh[5]local filename=pgb2file_f024_7_1p00 - 6: + product_functions.sh[6]wgrib2 pgb2file_f024_7_1p00 -not_if :RH: -grib pgb2file_f024_7_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_7_1p00.new - 6: 1:0:d=2021032312:SPFH:40 mb:24 hour fcst: - 6: 2:43754:d=2021032312:VVEL:40 mb:24 hour fcst: - 6: 3:127847:d=2021032312:DZDT:40 mb:24 hour fcst: - 6: 4:211916:d=2021032312:UGRD:40 mb:24 hour fcst: - 6: 5:260772:d=2021032312:VGRD:40 mb:24 hour fcst: - 6: 6:310285:d=2021032312:ABSV:40 mb:24 hour fcst: - 6: 7:355877:d=2021032312:O3MR:40 mb:24 hour fcst: - 6: 8:432193:d=2021032312:HGT:50 mb:24 hour fcst: - 6: 9:480691:d=2021032312:TMP:50 mb:24 hour fcst: - 6: 10:521342:d=2021032312:RH:50 mb:24 hour fcst: - 6: 11:547248:d=2021032312:TCDC:50 mb:24 hour fcst: - 6: 12:547427:d=2021032312:SPFH:50 mb:24 hour fcst: - 6: 13:595642:d=2021032312:VVEL:50 mb:24 hour fcst: - 6: 14:682215:d=2021032312:DZDT:50 mb:24 hour fcst: - 6: 15:765645:d=2021032312:UGRD:50 mb:24 hour fcst: - 6: 16:813600:d=2021032312:VGRD:50 mb:24 hour fcst: - 6: 17:862251:d=2021032312:ABSV:50 mb:24 hour fcst: - 6: 18:907125:d=2021032312:CLMR:50 mb:24 hour fcst: - 6: 19:907304:d=2021032312:ICMR:50 mb:24 hour fcst: - 6: 20:907483:d=2021032312:RWMR:50 mb:24 hour fcst: - 6: 21:907662:d=2021032312:SNMR:50 mb:24 hour fcst: - 6: 22:907841:d=2021032312:GRLE:50 mb:24 hour fcst: - 6: 23:908020:d=2021032312:O3MR:50 mb:24 hour fcst: - 6: 24:982277:d=2021032312:HGT:70 mb:24 hour fcst: - 6: 25:1031441:d=2021032312:TMP:70 mb:24 hour fcst: - 6: 26:1073393:d=2021032312:RH:70 mb:24 hour fcst: - 6: 27:1093343:d=2021032312:SPFH:70 mb:24 hour fcst: - 6: 28:1149014:d=2021032312:VVEL:70 mb:24 hour fcst: - 6: 29:1239318:d=2021032312:DZDT:70 mb:24 hour fcst: - 6: 30:1330465:d=2021032312:UGRD:70 mb:24 hour fcst: - 6: 31:1379476:d=2021032312:VGRD:70 mb:24 hour fcst: - 6: + product_functions.sh[10]rc=0 - 6: + product_functions.sh[11](( rc == 0 )) - 6: + product_functions.sh[11]mv pgb2file_f024_7_1p00.new pgb2file_f024_7_1p00 - 6: + product_functions.sh[12]return 0 - 6: + interp_atmos_master.sh[56]export err=0 - 6: + interp_atmos_master.sh[56]err=0 - 6: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 6: ++ interp_atmos_master.sh[62]wc -l - 6: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_7_1p00 -match 'LAND|ICEC' - 6: + interp_atmos_master.sh[62]var_count=0 - 6: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 6: + interp_atmos_master.sh[73]exit 0 -15: + bash[8]'[' -z '' ']' -15: + bash[9]case "$-" in -15: + bash[12]__lmod_vx=x -15: + bash[16]'[' -n x ']' -15: + bash[16]set +x -15: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -15: Shell debugging restarted -15: + bash[224]unset __lmod_vx -15: + interp_atmos_master.sh[7]input_file=tmpfile_f024_16 -15: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_16 -15: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -15: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -15: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -15: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -15: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -15: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -15: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -15: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -15: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -15: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -15: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -15: + interp_atmos_master.sh[31]IFS=: -15: + interp_atmos_master.sh[31]read -ra grids -15: + interp_atmos_master.sh[33]output_grids= -15: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -15: + interp_atmos_master.sh[35]gridopt=grid0p25 -15: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_16_0p25' -15: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -15: + interp_atmos_master.sh[35]gridopt=grid0p50 -15: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_16_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_16_0p50' -15: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -15: + interp_atmos_master.sh[35]gridopt=grid1p00 -15: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_16_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_16_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_16_1p00' -15: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_16 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_16_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_16_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_16_1p00 -15: 1:0:d=2021032312:ABSV:850 mb:24 hour fcst: -15: 2:21217:d=2021032312:CLMR:850 mb:24 hour fcst: -15: 3:35005:d=2021032312:ICMR:850 mb:24 hour fcst: -15: 4:45628:d=2021032312:RWMR:850 mb:24 hour fcst: -15: 5:56424:d=2021032312:SNMR:850 mb:24 hour fcst: -15: 6:68122:d=2021032312:GRLE:850 mb:24 hour fcst: -15: 7:70581:d=2021032312:O3MR:850 mb:24 hour fcst: -15: 8:97158:d=2021032312:HGT:900 mb:24 hour fcst: -15: 9:122194:d=2021032312:TMP:900 mb:24 hour fcst: -15: 10:143001:d=2021032312:RH:900 mb:24 hour fcst: -15: 11:162910:d=2021032312:TCDC:900 mb:24 hour fcst: -15: 12:175889:d=2021032312:SPFH:900 mb:24 hour fcst: -15: 13:205305:d=2021032312:VVEL:900 mb:24 hour fcst: -15: 14:230140:d=2021032312:DZDT:900 mb:24 hour fcst: -15: 15:256202:d=2021032312:UGRD:900 mb:24 hour fcst: -15: 16:277376:d=2021032312:VGRD:900 mb:24 hour fcst: -15: 17:299280:d=2021032312:ABSV:900 mb:24 hour fcst: -15: 18:320476:d=2021032312:CLMR:900 mb:24 hour fcst: -15: 19:330846:d=2021032312:ICMR:900 mb:24 hour fcst: -15: 20:339321:d=2021032312:RWMR:900 mb:24 hour fcst: -15: 21:351665:d=2021032312:SNMR:900 mb:24 hour fcst: -15: 22:361946:d=2021032312:GRLE:900 mb:24 hour fcst: -15: 23:365005:d=2021032312:O3MR:900 mb:24 hour fcst: -15: 24:391107:d=2021032312:HGT:925 mb:24 hour fcst: -15: 25:416278:d=2021032312:TMP:925 mb:24 hour fcst: -15: 26:437425:d=2021032312:RH:925 mb:24 hour fcst: -15: 27:457152:d=2021032312:TCDC:925 mb:24 hour fcst: -15: 28:468982:d=2021032312:SPFH:925 mb:24 hour fcst: -15: 29:498721:d=2021032312:VVEL:925 mb:24 hour fcst: -15: 30:523158:d=2021032312:DZDT:925 mb:24 hour fcst: -15: 31:548754:d=2021032312:UGRD:925 mb:24 hour fcst: -15: 32:570183:d=2021032312:VGRD:925 mb:24 hour fcst: -15: + interp_atmos_master.sh[47]export err=0 -15: + interp_atmos_master.sh[47]err=0 -15: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -15: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -15: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_16_0p25 -15: + product_functions.sh[5]local filename=pgb2file_f024_16_0p25 -15: + product_functions.sh[6]wgrib2 pgb2file_f024_16_0p25 -not_if :RH: -grib pgb2file_f024_16_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_16_0p25.new -15: 1:0:d=2021032312:ABSV:850 mb:24 hour fcst: -15: 2:506105:d=2021032312:CLMR:850 mb:24 hour fcst: -15: 3:804238:d=2021032312:ICMR:850 mb:24 hour fcst: -15: 4:1066004:d=2021032312:RWMR:850 mb:24 hour fcst: -15: 5:1270928:d=2021032312:SNMR:850 mb:24 hour fcst: -15: 6:1535145:d=2021032312:GRLE:850 mb:24 hour fcst: -15: 7:1579692:d=2021032312:O3MR:850 mb:24 hour fcst: -15: 8:2304652:d=2021032312:HGT:900 mb:24 hour fcst: -15: 9:2954206:d=2021032312:TMP:900 mb:24 hour fcst: -15: 10:3448635:d=2021032312:RH:900 mb:24 hour fcst: -15: 11:3901057:d=2021032312:TCDC:900 mb:24 hour fcst: -15: 12:4166736:d=2021032312:SPFH:900 mb:24 hour fcst: -15: 13:4971327:d=2021032312:VVEL:900 mb:24 hour fcst: -15: 14:5634770:d=2021032312:DZDT:900 mb:24 hour fcst: -15: 15:6337438:d=2021032312:UGRD:900 mb:24 hour fcst: -15: 16:6852736:d=2021032312:VGRD:900 mb:24 hour fcst: -15: 17:7396459:d=2021032312:ABSV:900 mb:24 hour fcst: -15: 18:7900932:d=2021032312:CLMR:900 mb:24 hour fcst: -15: 19:8129170:d=2021032312:ICMR:900 mb:24 hour fcst: -15: 20:8325361:d=2021032312:RWMR:900 mb:24 hour fcst: -15: 21:8570557:d=2021032312:SNMR:900 mb:24 hour fcst: -15: 22:8806919:d=2021032312:GRLE:900 mb:24 hour fcst: -15: 23:8866716:d=2021032312:O3MR:900 mb:24 hour fcst: -15: 24:9575248:d=2021032312:HGT:925 mb:24 hour fcst: -15: 25:10241159:d=2021032312:TMP:925 mb:24 hour fcst: -15: 26:10741780:d=2021032312:RH:925 mb:24 hour fcst: -15: 27:11194480:d=2021032312:TCDC:925 mb:24 hour fcst: -15: 28:11429306:d=2021032312:SPFH:925 mb:24 hour fcst: -15: 29:12254099:d=2021032312:VVEL:925 mb:24 hour fcst: -15: 30:12901005:d=2021032312:DZDT:925 mb:24 hour fcst: -15: 31:13583455:d=2021032312:UGRD:925 mb:24 hour fcst: -15: 32:14101304:d=2021032312:VGRD:925 mb:24 hour fcst: -15: + product_functions.sh[10]rc=0 -15: + product_functions.sh[11](( rc == 0 )) -15: + product_functions.sh[11]mv pgb2file_f024_16_0p25.new pgb2file_f024_16_0p25 -15: + product_functions.sh[12]return 0 -15: + interp_atmos_master.sh[56]export err=0 -15: + interp_atmos_master.sh[56]err=0 -15: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -15: ++ interp_atmos_master.sh[62]wc -l -15: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_16_0p25 -match 'LAND|ICEC' -15: + interp_atmos_master.sh[62]var_count=0 -15: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -15: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -15: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_16_0p50 -15: + product_functions.sh[5]local filename=pgb2file_f024_16_0p50 -15: + product_functions.sh[6]wgrib2 pgb2file_f024_16_0p50 -not_if :RH: -grib pgb2file_f024_16_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_16_0p50.new -15: 1:0:d=2021032312:ABSV:850 mb:24 hour fcst: -15: 2:174602:d=2021032312:CLMR:850 mb:24 hour fcst: -15: 3:287523:d=2021032312:ICMR:850 mb:24 hour fcst: -15: 4:379644:d=2021032312:RWMR:850 mb:24 hour fcst: -15: 5:456280:d=2021032312:SNMR:850 mb:24 hour fcst: -15: 6:549674:d=2021032312:GRLE:850 mb:24 hour fcst: -15: 7:566406:d=2021032312:O3MR:850 mb:24 hour fcst: -15: 8:812454:d=2021032312:HGT:900 mb:24 hour fcst: -15: 9:1032238:d=2021032312:TMP:900 mb:24 hour fcst: -15: 10:1203218:d=2021032312:RH:900 mb:24 hour fcst: -15: 11:1362259:d=2021032312:TCDC:900 mb:24 hour fcst: -15: 12:1461184:d=2021032312:SPFH:900 mb:24 hour fcst: -15: 13:1740691:d=2021032312:VVEL:900 mb:24 hour fcst: -15: 14:1964291:d=2021032312:DZDT:900 mb:24 hour fcst: -15: 15:2197953:d=2021032312:UGRD:900 mb:24 hour fcst: -15: 16:2375311:d=2021032312:VGRD:900 mb:24 hour fcst: -15: 17:2560983:d=2021032312:ABSV:900 mb:24 hour fcst: -15: 18:2735353:d=2021032312:CLMR:900 mb:24 hour fcst: -15: 19:2819467:d=2021032312:ICMR:900 mb:24 hour fcst: -15: 20:2888566:d=2021032312:RWMR:900 mb:24 hour fcst: -15: 21:2978741:d=2021032312:SNMR:900 mb:24 hour fcst: -15: 22:3061226:d=2021032312:GRLE:900 mb:24 hour fcst: -15: 23:3082842:d=2021032312:O3MR:900 mb:24 hour fcst: -15: 24:3321233:d=2021032312:HGT:925 mb:24 hour fcst: -15: 25:3543163:d=2021032312:TMP:925 mb:24 hour fcst: -15: 26:3716116:d=2021032312:RH:925 mb:24 hour fcst: -15: 27:3874843:d=2021032312:TCDC:925 mb:24 hour fcst: -15: 28:3962404:d=2021032312:SPFH:925 mb:24 hour fcst: -15: 29:4242686:d=2021032312:VVEL:925 mb:24 hour fcst: -15: 30:4460496:d=2021032312:DZDT:925 mb:24 hour fcst: -15: 31:4686605:d=2021032312:UGRD:925 mb:24 hour fcst: -15: 32:4864862:d=2021032312:VGRD:925 mb:24 hour fcst: -15: + product_functions.sh[10]rc=0 -15: + product_functions.sh[11](( rc == 0 )) -15: + product_functions.sh[11]mv pgb2file_f024_16_0p50.new pgb2file_f024_16_0p50 -15: + product_functions.sh[12]return 0 -15: + interp_atmos_master.sh[56]export err=0 -15: + interp_atmos_master.sh[56]err=0 -15: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -15: ++ interp_atmos_master.sh[62]wc -l -15: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_16_0p50 -match 'LAND|ICEC' -15: + interp_atmos_master.sh[62]var_count=0 -15: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -15: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -15: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_16_1p00 -15: + product_functions.sh[5]local filename=pgb2file_f024_16_1p00 -15: + product_functions.sh[6]wgrib2 pgb2file_f024_16_1p00 -not_if :RH: -grib pgb2file_f024_16_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_16_1p00.new -15: 1:0:d=2021032312:ABSV:850 mb:24 hour fcst: -15: 2:56248:d=2021032312:CLMR:850 mb:24 hour fcst: -15: 3:93367:d=2021032312:ICMR:850 mb:24 hour fcst: -15: 4:122634:d=2021032312:RWMR:850 mb:24 hour fcst: -15: 5:149752:d=2021032312:SNMR:850 mb:24 hour fcst: -15: 6:180529:d=2021032312:GRLE:850 mb:24 hour fcst: -15: 7:186700:d=2021032312:O3MR:850 mb:24 hour fcst: -15: 8:261276:d=2021032312:HGT:900 mb:24 hour fcst: -15: 9:329722:d=2021032312:TMP:900 mb:24 hour fcst: -15: 10:384752:d=2021032312:RH:900 mb:24 hour fcst: -15: 11:436614:d=2021032312:TCDC:900 mb:24 hour fcst: -15: 12:471367:d=2021032312:SPFH:900 mb:24 hour fcst: -15: 13:554377:d=2021032312:VVEL:900 mb:24 hour fcst: -15: 14:623606:d=2021032312:DZDT:900 mb:24 hour fcst: -15: 15:694833:d=2021032312:UGRD:900 mb:24 hour fcst: -15: 16:751722:d=2021032312:VGRD:900 mb:24 hour fcst: -15: 17:810588:d=2021032312:ABSV:900 mb:24 hour fcst: -15: 18:866885:d=2021032312:CLMR:900 mb:24 hour fcst: -15: 19:895208:d=2021032312:ICMR:900 mb:24 hour fcst: -15: 20:917858:d=2021032312:RWMR:900 mb:24 hour fcst: -15: 21:949715:d=2021032312:SNMR:900 mb:24 hour fcst: -15: 22:976654:d=2021032312:GRLE:900 mb:24 hour fcst: -15: 23:984393:d=2021032312:O3MR:900 mb:24 hour fcst: -15: 24:1057546:d=2021032312:HGT:925 mb:24 hour fcst: -15: 25:1126605:d=2021032312:TMP:925 mb:24 hour fcst: -15: 26:1182552:d=2021032312:RH:925 mb:24 hour fcst: -15: 27:1234174:d=2021032312:TCDC:925 mb:24 hour fcst: -15: 28:1265773:d=2021032312:SPFH:925 mb:24 hour fcst: -15: 29:1349171:d=2021032312:VVEL:925 mb:24 hour fcst: -15: 30:1416706:d=2021032312:DZDT:925 mb:24 hour fcst: -15: 31:1485966:d=2021032312:UGRD:925 mb:24 hour fcst: -15: 32:1543210:d=2021032312:VGRD:925 mb:24 hour fcst: -15: + product_functions.sh[10]rc=0 -15: + product_functions.sh[11](( rc == 0 )) -15: + product_functions.sh[11]mv pgb2file_f024_16_1p00.new pgb2file_f024_16_1p00 -15: + product_functions.sh[12]return 0 -15: + interp_atmos_master.sh[56]export err=0 -15: + interp_atmos_master.sh[56]err=0 -15: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -15: ++ interp_atmos_master.sh[62]wc -l -15: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_16_1p00 -match 'LAND|ICEC' -15: + interp_atmos_master.sh[62]var_count=0 -15: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -15: + interp_atmos_master.sh[73]exit 0 - 9: + bash[8]'[' -z '' ']' - 9: + bash[9]case "$-" in - 9: + bash[12]__lmod_vx=x - 9: + bash[16]'[' -n x ']' - 9: + bash[16]set +x - 9: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 9: Shell debugging restarted - 9: + bash[224]unset __lmod_vx - 9: + interp_atmos_master.sh[7]input_file=tmpfile_f024_10 - 9: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_10 - 9: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 - 9: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 9: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 9: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 9: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 9: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 9: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 9: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 9: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 9: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 9: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 9: + interp_atmos_master.sh[31]IFS=: - 9: + interp_atmos_master.sh[31]read -ra grids - 9: + interp_atmos_master.sh[33]output_grids= - 9: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 9: + interp_atmos_master.sh[35]gridopt=grid0p25 - 9: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_10_0p25' - 9: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 9: + interp_atmos_master.sh[35]gridopt=grid0p50 - 9: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_10_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_10_0p50' - 9: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 9: + interp_atmos_master.sh[35]gridopt=grid1p00 - 9: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_10_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_10_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_10_1p00' - 9: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_10 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_10_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_10_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_10_1p00 - 9: 1:0:d=2021032312:RWMR:250 mb:24 hour fcst: - 9: 2:179:d=2021032312:SNMR:250 mb:24 hour fcst: - 9: 3:8984:d=2021032312:GRLE:250 mb:24 hour fcst: - 9: 4:9328:d=2021032312:O3MR:250 mb:24 hour fcst: - 9: 5:34764:d=2021032312:HGT:300 mb:24 hour fcst: - 9: 6:57074:d=2021032312:TMP:300 mb:24 hour fcst: - 9: 7:75642:d=2021032312:RH:300 mb:24 hour fcst: - 9: 8:94641:d=2021032312:TCDC:300 mb:24 hour fcst: - 9: 9:107373:d=2021032312:SPFH:300 mb:24 hour fcst: - 9: 10:134077:d=2021032312:VVEL:300 mb:24 hour fcst: - 9: 11:157992:d=2021032312:DZDT:300 mb:24 hour fcst: - 9: 12:186140:d=2021032312:UGRD:300 mb:24 hour fcst: - 9: 13:200645:d=2021032312:VGRD:300 mb:24 hour fcst: - 9: 14:215967:d=2021032312:ABSV:300 mb:24 hour fcst: - 9: 15:238276:d=2021032312:CLMR:300 mb:24 hour fcst: - 9: 16:238455:d=2021032312:ICMR:300 mb:24 hour fcst: - 9: 17:254274:d=2021032312:RWMR:300 mb:24 hour fcst: - 9: 18:254453:d=2021032312:SNMR:300 mb:24 hour fcst: - 9: 19:265854:d=2021032312:GRLE:300 mb:24 hour fcst: - 9: 20:266265:d=2021032312:O3MR:300 mb:24 hour fcst: - 9: 21:293264:d=2021032312:HGT:350 mb:24 hour fcst: - 9: 22:315401:d=2021032312:TMP:350 mb:24 hour fcst: - 9: 23:333615:d=2021032312:RH:350 mb:24 hour fcst: - 9: 24:352728:d=2021032312:TCDC:350 mb:24 hour fcst: - 9: 25:365820:d=2021032312:SPFH:350 mb:24 hour fcst: - 9: 26:394003:d=2021032312:VVEL:350 mb:24 hour fcst: - 9: 27:418530:d=2021032312:DZDT:350 mb:24 hour fcst: - 9: 28:446850:d=2021032312:UGRD:350 mb:24 hour fcst: - 9: 29:461440:d=2021032312:VGRD:350 mb:24 hour fcst: - 9: 30:476690:d=2021032312:ABSV:350 mb:24 hour fcst: - 9: 31:498959:d=2021032312:CLMR:350 mb:24 hour fcst: - 9: + interp_atmos_master.sh[47]export err=0 - 9: + interp_atmos_master.sh[47]err=0 - 9: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 9: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 9: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_10_0p25 - 9: + product_functions.sh[5]local filename=pgb2file_f024_10_0p25 - 9: + product_functions.sh[6]wgrib2 pgb2file_f024_10_0p25 -not_if :RH: -grib pgb2file_f024_10_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_10_0p25.new - 9: 1:0:d=2021032312:RWMR:250 mb:24 hour fcst: - 9: 2:179:d=2021032312:SNMR:250 mb:24 hour fcst: - 9: 3:171722:d=2021032312:GRLE:250 mb:24 hour fcst: - 9: 4:174774:d=2021032312:O3MR:250 mb:24 hour fcst: - 9: 5:841700:d=2021032312:HGT:300 mb:24 hour fcst: - 9: 6:1396363:d=2021032312:TMP:300 mb:24 hour fcst: - 9: 7:1807854:d=2021032312:RH:300 mb:24 hour fcst: - 9: 8:2233551:d=2021032312:TCDC:300 mb:24 hour fcst: - 9: 9:2472754:d=2021032312:SPFH:300 mb:24 hour fcst: - 9: 10:3179157:d=2021032312:VVEL:300 mb:24 hour fcst: - 9: 11:3812986:d=2021032312:DZDT:300 mb:24 hour fcst: - 9: 12:4607492:d=2021032312:UGRD:300 mb:24 hour fcst: - 9: 13:4924504:d=2021032312:VGRD:300 mb:24 hour fcst: - 9: 14:5261405:d=2021032312:ABSV:300 mb:24 hour fcst: - 9: 15:5812981:d=2021032312:CLMR:300 mb:24 hour fcst: - 9: 16:5813160:d=2021032312:ICMR:300 mb:24 hour fcst: - 9: 17:6160612:d=2021032312:RWMR:300 mb:24 hour fcst: - 9: 18:6160791:d=2021032312:SNMR:300 mb:24 hour fcst: - 9: 19:6394096:d=2021032312:GRLE:300 mb:24 hour fcst: - 9: 20:6398644:d=2021032312:O3MR:300 mb:24 hour fcst: - 9: 21:7119399:d=2021032312:HGT:350 mb:24 hour fcst: - 9: 22:7666626:d=2021032312:TMP:350 mb:24 hour fcst: - 9: 23:8073657:d=2021032312:RH:350 mb:24 hour fcst: - 9: 24:8504422:d=2021032312:TCDC:350 mb:24 hour fcst: - 9: 25:8744301:d=2021032312:SPFH:350 mb:24 hour fcst: - 9: 26:9491079:d=2021032312:VVEL:350 mb:24 hour fcst: - 9: 27:10144644:d=2021032312:DZDT:350 mb:24 hour fcst: - 9: 28:10940912:d=2021032312:UGRD:350 mb:24 hour fcst: - 9: 29:11252622:d=2021032312:VGRD:350 mb:24 hour fcst: - 9: 30:11583441:d=2021032312:ABSV:350 mb:24 hour fcst: - 9: 31:12131738:d=2021032312:CLMR:350 mb:24 hour fcst: - 9: + product_functions.sh[10]rc=0 - 9: + product_functions.sh[11](( rc == 0 )) - 9: + product_functions.sh[11]mv pgb2file_f024_10_0p25.new pgb2file_f024_10_0p25 - 9: + product_functions.sh[12]return 0 - 9: + interp_atmos_master.sh[56]export err=0 - 9: + interp_atmos_master.sh[56]err=0 - 9: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 9: ++ interp_atmos_master.sh[62]wc -l - 9: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_10_0p25 -match 'LAND|ICEC' - 9: + interp_atmos_master.sh[62]var_count=0 - 9: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 9: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 9: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_10_0p50 - 9: + product_functions.sh[5]local filename=pgb2file_f024_10_0p50 - 9: + product_functions.sh[6]wgrib2 pgb2file_f024_10_0p50 -not_if :RH: -grib pgb2file_f024_10_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_10_0p50.new - 9: 1:0:d=2021032312:RWMR:250 mb:24 hour fcst: - 9: 2:179:d=2021032312:SNMR:250 mb:24 hour fcst: - 9: 3:61660:d=2021032312:GRLE:250 mb:24 hour fcst: - 9: 4:63005:d=2021032312:O3MR:250 mb:24 hour fcst: - 9: 5:286749:d=2021032312:HGT:300 mb:24 hour fcst: - 9: 6:473452:d=2021032312:TMP:300 mb:24 hour fcst: - 9: 7:617279:d=2021032312:RH:300 mb:24 hour fcst: - 9: 8:765734:d=2021032312:TCDC:300 mb:24 hour fcst: - 9: 9:858225:d=2021032312:SPFH:300 mb:24 hour fcst: - 9: 10:1102444:d=2021032312:VVEL:300 mb:24 hour fcst: - 9: 11:1316999:d=2021032312:DZDT:300 mb:24 hour fcst: - 9: 12:1573719:d=2021032312:UGRD:300 mb:24 hour fcst: - 9: 13:1678590:d=2021032312:VGRD:300 mb:24 hour fcst: - 9: 14:1790136:d=2021032312:ABSV:300 mb:24 hour fcst: - 9: 15:1979008:d=2021032312:CLMR:300 mb:24 hour fcst: - 9: 16:1979187:d=2021032312:ICMR:300 mb:24 hour fcst: - 9: 17:2105292:d=2021032312:RWMR:300 mb:24 hour fcst: - 9: 18:2105471:d=2021032312:SNMR:300 mb:24 hour fcst: - 9: 19:2189645:d=2021032312:GRLE:300 mb:24 hour fcst: - 9: 20:2191520:d=2021032312:O3MR:300 mb:24 hour fcst: - 9: 21:2433238:d=2021032312:HGT:350 mb:24 hour fcst: - 9: 22:2618043:d=2021032312:TMP:350 mb:24 hour fcst: - 9: 23:2759476:d=2021032312:RH:350 mb:24 hour fcst: - 9: 24:2909969:d=2021032312:TCDC:350 mb:24 hour fcst: - 9: 25:3003183:d=2021032312:SPFH:350 mb:24 hour fcst: - 9: 26:3266258:d=2021032312:VVEL:350 mb:24 hour fcst: - 9: 27:3488752:d=2021032312:DZDT:350 mb:24 hour fcst: - 9: 28:3747818:d=2021032312:UGRD:350 mb:24 hour fcst: - 9: 29:3851095:d=2021032312:VGRD:350 mb:24 hour fcst: - 9: 30:3961596:d=2021032312:ABSV:350 mb:24 hour fcst: - 9: 31:4149726:d=2021032312:CLMR:350 mb:24 hour fcst: - 9: + product_functions.sh[10]rc=0 - 9: + product_functions.sh[11](( rc == 0 )) - 9: + product_functions.sh[11]mv pgb2file_f024_10_0p50.new pgb2file_f024_10_0p50 - 9: + product_functions.sh[12]return 0 - 9: + interp_atmos_master.sh[56]export err=0 - 9: + interp_atmos_master.sh[56]err=0 - 9: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 9: ++ interp_atmos_master.sh[62]wc -l - 9: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_10_0p50 -match 'LAND|ICEC' - 9: + interp_atmos_master.sh[62]var_count=0 - 9: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 9: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 9: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_10_1p00 - 9: + product_functions.sh[5]local filename=pgb2file_f024_10_1p00 - 9: + product_functions.sh[6]wgrib2 pgb2file_f024_10_1p00 -not_if :RH: -grib pgb2file_f024_10_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_10_1p00.new - 9: 1:0:d=2021032312:RWMR:250 mb:24 hour fcst: - 9: 2:179:d=2021032312:SNMR:250 mb:24 hour fcst: - 9: 3:21985:d=2021032312:GRLE:250 mb:24 hour fcst: - 9: 4:22601:d=2021032312:O3MR:250 mb:24 hour fcst: - 9: 5:92389:d=2021032312:HGT:300 mb:24 hour fcst: - 9: 6:152049:d=2021032312:TMP:300 mb:24 hour fcst: - 9: 7:199433:d=2021032312:RH:300 mb:24 hour fcst: - 9: 8:248266:d=2021032312:TCDC:300 mb:24 hour fcst: - 9: 9:280965:d=2021032312:SPFH:300 mb:24 hour fcst: - 9: 10:354615:d=2021032312:VVEL:300 mb:24 hour fcst: - 9: 11:420646:d=2021032312:DZDT:300 mb:24 hour fcst: - 9: 12:497570:d=2021032312:UGRD:300 mb:24 hour fcst: - 9: 13:533286:d=2021032312:VGRD:300 mb:24 hour fcst: - 9: 14:571520:d=2021032312:ABSV:300 mb:24 hour fcst: - 9: 15:631408:d=2021032312:CLMR:300 mb:24 hour fcst: - 9: 16:631587:d=2021032312:ICMR:300 mb:24 hour fcst: - 9: 17:673007:d=2021032312:RWMR:300 mb:24 hour fcst: - 9: 18:673186:d=2021032312:SNMR:300 mb:24 hour fcst: - 9: 19:702151:d=2021032312:GRLE:300 mb:24 hour fcst: - 9: 20:702977:d=2021032312:O3MR:300 mb:24 hour fcst: - 9: 21:777416:d=2021032312:HGT:350 mb:24 hour fcst: - 9: 22:836564:d=2021032312:TMP:350 mb:24 hour fcst: - 9: 23:882900:d=2021032312:RH:350 mb:24 hour fcst: - 9: 24:932280:d=2021032312:TCDC:350 mb:24 hour fcst: - 9: 25:965744:d=2021032312:SPFH:350 mb:24 hour fcst: - 9: 26:1044825:d=2021032312:VVEL:350 mb:24 hour fcst: - 9: 27:1112538:d=2021032312:DZDT:350 mb:24 hour fcst: - 9: 28:1189966:d=2021032312:UGRD:350 mb:24 hour fcst: - 9: 29:1225496:d=2021032312:VGRD:350 mb:24 hour fcst: - 9: 30:1263415:d=2021032312:ABSV:350 mb:24 hour fcst: - 9: 31:1323091:d=2021032312:CLMR:350 mb:24 hour fcst: - 9: + product_functions.sh[10]rc=0 - 9: + product_functions.sh[11](( rc == 0 )) - 9: + product_functions.sh[11]mv pgb2file_f024_10_1p00.new pgb2file_f024_10_1p00 - 9: + product_functions.sh[12]return 0 - 9: + interp_atmos_master.sh[56]export err=0 - 9: + interp_atmos_master.sh[56]err=0 - 9: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 9: ++ interp_atmos_master.sh[62]wc -l - 9: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_10_1p00 -match 'LAND|ICEC' - 9: + interp_atmos_master.sh[62]var_count=0 - 9: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 9: + interp_atmos_master.sh[73]exit 0 - 2: + bash[8]'[' -z '' ']' - 2: + bash[9]case "$-" in - 2: + bash[12]__lmod_vx=x - 2: + bash[16]'[' -n x ']' - 2: + bash[16]set +x - 2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 2: Shell debugging restarted - 2: + bash[224]unset __lmod_vx - 2: + interp_atmos_master.sh[7]input_file=tmpfile_f024_3 - 2: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_3 - 2: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 - 2: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 2: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 2: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 2: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 2: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 2: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 2: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 2: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 2: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 2: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 2: + interp_atmos_master.sh[31]IFS=: - 2: + interp_atmos_master.sh[31]read -ra grids - 2: + interp_atmos_master.sh[33]output_grids= - 2: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 2: + interp_atmos_master.sh[35]gridopt=grid0p25 - 2: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_3_0p25' - 2: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 2: + interp_atmos_master.sh[35]gridopt=grid0p50 - 2: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_3_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_3_0p50' - 2: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 2: + interp_atmos_master.sh[35]gridopt=grid1p00 - 2: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_3_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_3_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_3_1p00' - 2: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_3 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_3_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_3_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_3_1p00 - 2: 1:0:d=2021032312:O3MR:0.1 mb:24 hour fcst: - 2: 2:23624:d=2021032312:HGT:0.2 mb:24 hour fcst: - 2: 3:44775:d=2021032312:TMP:0.2 mb:24 hour fcst: - 2: 4:61089:d=2021032312:RH:0.2 mb:24 hour fcst: - 2: 5:72021:d=2021032312:SPFH:0.2 mb:24 hour fcst: - 2: 6:91691:d=2021032312:VVEL:0.2 mb:24 hour fcst: - 2: 7:120643:d=2021032312:DZDT:0.2 mb:24 hour fcst: - 2: 8:148180:d=2021032312:UGRD:0.2 mb:24 hour fcst: - 2: 9:167600:d=2021032312:VGRD:0.2 mb:24 hour fcst: - 2: 10:186514:d=2021032312:ABSV:0.2 mb:24 hour fcst: - 2: 11:204178:d=2021032312:O3MR:0.2 mb:24 hour fcst: - 2: 12:229960:d=2021032312:HGT:0.4 mb:24 hour fcst: - 2: 13:250722:d=2021032312:TMP:0.4 mb:24 hour fcst: - 2: 14:266737:d=2021032312:RH:0.4 mb:24 hour fcst: - 2: 15:275846:d=2021032312:SPFH:0.4 mb:24 hour fcst: - 2: 16:295810:d=2021032312:VVEL:0.4 mb:24 hour fcst: - 2: 17:320730:d=2021032312:DZDT:0.4 mb:24 hour fcst: - 2: 18:351258:d=2021032312:UGRD:0.4 mb:24 hour fcst: - 2: 19:370154:d=2021032312:VGRD:0.4 mb:24 hour fcst: - 2: 20:388425:d=2021032312:ABSV:0.4 mb:24 hour fcst: - 2: 21:405223:d=2021032312:O3MR:0.4 mb:24 hour fcst: - 2: 22:429843:d=2021032312:HGT:0.7 mb:24 hour fcst: - 2: 23:447975:d=2021032312:TMP:0.7 mb:24 hour fcst: - 2: 24:463821:d=2021032312:RH:0.7 mb:24 hour fcst: - 2: 25:472789:d=2021032312:SPFH:0.7 mb:24 hour fcst: - 2: 26:492035:d=2021032312:VVEL:0.7 mb:24 hour fcst: - 2: 27:517945:d=2021032312:DZDT:0.7 mb:24 hour fcst: - 2: 28:547085:d=2021032312:UGRD:0.7 mb:24 hour fcst: - 2: 29:565530:d=2021032312:VGRD:0.7 mb:24 hour fcst: - 2: 30:583389:d=2021032312:ABSV:0.7 mb:24 hour fcst: - 2: 31:599983:d=2021032312:O3MR:0.7 mb:24 hour fcst: - 2: + interp_atmos_master.sh[47]export err=0 - 2: + interp_atmos_master.sh[47]err=0 - 2: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 2: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 2: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_3_0p25 - 2: + product_functions.sh[5]local filename=pgb2file_f024_3_0p25 - 2: + product_functions.sh[6]wgrib2 pgb2file_f024_3_0p25 -not_if :RH: -grib pgb2file_f024_3_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_3_0p25.new - 2: 1:0:d=2021032312:O3MR:0.1 mb:24 hour fcst: - 2: 2:623720:d=2021032312:HGT:0.2 mb:24 hour fcst: - 2: 3:1143972:d=2021032312:TMP:0.2 mb:24 hour fcst: - 2: 4:1503275:d=2021032312:RH:0.2 mb:24 hour fcst: - 2: 5:1758936:d=2021032312:SPFH:0.2 mb:24 hour fcst: - 2: 6:2237560:d=2021032312:VVEL:0.2 mb:24 hour fcst: - 2: 7:3056429:d=2021032312:DZDT:0.2 mb:24 hour fcst: - 2: 8:3838683:d=2021032312:UGRD:0.2 mb:24 hour fcst: - 2: 9:4299855:d=2021032312:VGRD:0.2 mb:24 hour fcst: - 2: 10:4738352:d=2021032312:ABSV:0.2 mb:24 hour fcst: - 2: 11:5125717:d=2021032312:O3MR:0.2 mb:24 hour fcst: - 2: 12:5832731:d=2021032312:HGT:0.4 mb:24 hour fcst: - 2: 13:6342126:d=2021032312:TMP:0.4 mb:24 hour fcst: - 2: 14:6693845:d=2021032312:RH:0.4 mb:24 hour fcst: - 2: 15:6928083:d=2021032312:SPFH:0.4 mb:24 hour fcst: - 2: 16:7446272:d=2021032312:VVEL:0.4 mb:24 hour fcst: - 2: 17:8122885:d=2021032312:DZDT:0.4 mb:24 hour fcst: - 2: 18:9035368:d=2021032312:UGRD:0.4 mb:24 hour fcst: - 2: 19:9481056:d=2021032312:VGRD:0.4 mb:24 hour fcst: - 2: 20:9902489:d=2021032312:ABSV:0.4 mb:24 hour fcst: - 2: 21:10261357:d=2021032312:O3MR:0.4 mb:24 hour fcst: - 2: 22:10926173:d=2021032312:HGT:0.7 mb:24 hour fcst: - 2: 23:11341533:d=2021032312:TMP:0.7 mb:24 hour fcst: - 2: 24:11689462:d=2021032312:RH:0.7 mb:24 hour fcst: - 2: 25:11934421:d=2021032312:SPFH:0.7 mb:24 hour fcst: - 2: 26:12456663:d=2021032312:VVEL:0.7 mb:24 hour fcst: - 2: 27:13164370:d=2021032312:DZDT:0.7 mb:24 hour fcst: - 2: 28:14024663:d=2021032312:UGRD:0.7 mb:24 hour fcst: - 2: 29:14462313:d=2021032312:VGRD:0.7 mb:24 hour fcst: - 2: 30:14876290:d=2021032312:ABSV:0.7 mb:24 hour fcst: - 2: 31:15226050:d=2021032312:O3MR:0.7 mb:24 hour fcst: - 2: + product_functions.sh[10]rc=0 - 2: + product_functions.sh[11](( rc == 0 )) - 2: + product_functions.sh[11]mv pgb2file_f024_3_0p25.new pgb2file_f024_3_0p25 - 2: + product_functions.sh[12]return 0 - 2: + interp_atmos_master.sh[56]export err=0 - 2: + interp_atmos_master.sh[56]err=0 - 2: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 2: ++ interp_atmos_master.sh[62]wc -l - 2: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_3_0p25 -match 'LAND|ICEC' - 2: + interp_atmos_master.sh[62]var_count=0 - 2: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 2: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 2: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_3_0p50 - 2: + product_functions.sh[5]local filename=pgb2file_f024_3_0p50 - 2: + product_functions.sh[6]wgrib2 pgb2file_f024_3_0p50 -not_if :RH: -grib pgb2file_f024_3_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_3_0p50.new - 2: 1:0:d=2021032312:O3MR:0.1 mb:24 hour fcst: - 2: 2:211281:d=2021032312:HGT:0.2 mb:24 hour fcst: - 2: 3:388378:d=2021032312:TMP:0.2 mb:24 hour fcst: - 2: 4:510386:d=2021032312:RH:0.2 mb:24 hour fcst: - 2: 5:589844:d=2021032312:SPFH:0.2 mb:24 hour fcst: - 2: 6:748068:d=2021032312:VVEL:0.2 mb:24 hour fcst: - 2: 7:1030423:d=2021032312:DZDT:0.2 mb:24 hour fcst: - 2: 8:1285930:d=2021032312:UGRD:0.2 mb:24 hour fcst: - 2: 9:1443584:d=2021032312:VGRD:0.2 mb:24 hour fcst: - 2: 10:1595860:d=2021032312:ABSV:0.2 mb:24 hour fcst: - 2: 11:1728806:d=2021032312:O3MR:0.2 mb:24 hour fcst: - 2: 12:1967442:d=2021032312:HGT:0.4 mb:24 hour fcst: - 2: 13:2140990:d=2021032312:TMP:0.4 mb:24 hour fcst: - 2: 14:2261358:d=2021032312:RH:0.4 mb:24 hour fcst: - 2: 15:2329372:d=2021032312:SPFH:0.4 mb:24 hour fcst: - 2: 16:2503023:d=2021032312:VVEL:0.4 mb:24 hour fcst: - 2: 17:2732391:d=2021032312:DZDT:0.4 mb:24 hour fcst: - 2: 18:3028572:d=2021032312:UGRD:0.4 mb:24 hour fcst: - 2: 19:3181494:d=2021032312:VGRD:0.4 mb:24 hour fcst: - 2: 20:3325652:d=2021032312:ABSV:0.4 mb:24 hour fcst: - 2: 21:3448810:d=2021032312:O3MR:0.4 mb:24 hour fcst: - 2: 22:3672185:d=2021032312:HGT:0.7 mb:24 hour fcst: - 2: 23:3814675:d=2021032312:TMP:0.7 mb:24 hour fcst: - 2: 24:3933525:d=2021032312:RH:0.7 mb:24 hour fcst: - 2: 25:4002505:d=2021032312:SPFH:0.7 mb:24 hour fcst: - 2: 26:4174076:d=2021032312:VVEL:0.7 mb:24 hour fcst: - 2: 27:4416439:d=2021032312:DZDT:0.7 mb:24 hour fcst: - 2: 28:4695677:d=2021032312:UGRD:0.7 mb:24 hour fcst: - 2: 29:4845371:d=2021032312:VGRD:0.7 mb:24 hour fcst: - 2: 30:4986519:d=2021032312:ABSV:0.7 mb:24 hour fcst: - 2: 31:5107134:d=2021032312:O3MR:0.7 mb:24 hour fcst: - 2: + product_functions.sh[10]rc=0 - 2: + product_functions.sh[11](( rc == 0 )) - 2: + product_functions.sh[11]mv pgb2file_f024_3_0p50.new pgb2file_f024_3_0p50 - 2: + product_functions.sh[12]return 0 - 2: + interp_atmos_master.sh[56]export err=0 - 2: + interp_atmos_master.sh[56]err=0 - 2: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 2: ++ interp_atmos_master.sh[62]wc -l - 2: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_3_0p50 -match 'LAND|ICEC' - 2: + interp_atmos_master.sh[62]var_count=0 - 2: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 2: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 2: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_3_1p00 - 2: + product_functions.sh[5]local filename=pgb2file_f024_3_1p00 - 2: + product_functions.sh[6]wgrib2 pgb2file_f024_3_1p00 -not_if :RH: -grib pgb2file_f024_3_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_3_1p00.new - 2: 1:0:d=2021032312:O3MR:0.1 mb:24 hour fcst: - 2: 2:64766:d=2021032312:HGT:0.2 mb:24 hour fcst: - 2: 3:121035:d=2021032312:TMP:0.2 mb:24 hour fcst: - 2: 4:161769:d=2021032312:RH:0.2 mb:24 hour fcst: - 2: 5:187872:d=2021032312:SPFH:0.2 mb:24 hour fcst: - 2: 6:239293:d=2021032312:VVEL:0.2 mb:24 hour fcst: - 2: 7:322698:d=2021032312:DZDT:0.2 mb:24 hour fcst: - 2: 8:398992:d=2021032312:UGRD:0.2 mb:24 hour fcst: - 2: 9:449950:d=2021032312:VGRD:0.2 mb:24 hour fcst: - 2: 10:500038:d=2021032312:ABSV:0.2 mb:24 hour fcst: - 2: 11:544373:d=2021032312:O3MR:0.2 mb:24 hour fcst: - 2: 12:616695:d=2021032312:HGT:0.4 mb:24 hour fcst: - 2: 13:672006:d=2021032312:TMP:0.4 mb:24 hour fcst: - 2: 14:712181:d=2021032312:RH:0.4 mb:24 hour fcst: - 2: 15:733939:d=2021032312:SPFH:0.4 mb:24 hour fcst: - 2: 16:787986:d=2021032312:VVEL:0.4 mb:24 hour fcst: - 2: 17:857389:d=2021032312:DZDT:0.4 mb:24 hour fcst: - 2: 18:944406:d=2021032312:UGRD:0.4 mb:24 hour fcst: - 2: 19:994123:d=2021032312:VGRD:0.4 mb:24 hour fcst: - 2: 20:1041530:d=2021032312:ABSV:0.4 mb:24 hour fcst: - 2: 21:1083200:d=2021032312:O3MR:0.4 mb:24 hour fcst: - 2: 22:1151384:d=2021032312:HGT:0.7 mb:24 hour fcst: - 2: 23:1197865:d=2021032312:TMP:0.7 mb:24 hour fcst: - 2: 24:1237508:d=2021032312:RH:0.7 mb:24 hour fcst: - 2: 25:1259235:d=2021032312:SPFH:0.7 mb:24 hour fcst: - 2: 26:1311154:d=2021032312:VVEL:0.7 mb:24 hour fcst: - 2: 27:1384439:d=2021032312:DZDT:0.7 mb:24 hour fcst: - 2: 28:1466771:d=2021032312:UGRD:0.7 mb:24 hour fcst: - 2: 29:1515351:d=2021032312:VGRD:0.7 mb:24 hour fcst: - 2: 30:1561798:d=2021032312:ABSV:0.7 mb:24 hour fcst: - 2: 31:1602304:d=2021032312:O3MR:0.7 mb:24 hour fcst: - 2: + product_functions.sh[10]rc=0 - 2: + product_functions.sh[11](( rc == 0 )) - 2: + product_functions.sh[11]mv pgb2file_f024_3_1p00.new pgb2file_f024_3_1p00 - 2: + product_functions.sh[12]return 0 - 2: + interp_atmos_master.sh[56]export err=0 - 2: + interp_atmos_master.sh[56]err=0 - 2: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 2: ++ interp_atmos_master.sh[62]wc -l - 2: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_3_1p00 -match 'LAND|ICEC' - 2: + interp_atmos_master.sh[62]var_count=0 - 2: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 2: + interp_atmos_master.sh[73]exit 0 - 3: + bash[8]'[' -z '' ']' - 3: + bash[9]case "$-" in - 3: + bash[12]__lmod_vx=x - 3: + bash[16]'[' -n x ']' - 3: + bash[16]set +x - 3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 3: Shell debugging restarted - 3: + bash[224]unset __lmod_vx - 3: + interp_atmos_master.sh[7]input_file=tmpfile_f024_4 - 3: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_4 - 3: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 - 3: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 3: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 3: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 3: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 3: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 3: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 3: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 3: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 3: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 3: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 3: + interp_atmos_master.sh[31]IFS=: - 3: + interp_atmos_master.sh[31]read -ra grids - 3: + interp_atmos_master.sh[33]output_grids= - 3: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 3: + interp_atmos_master.sh[35]gridopt=grid0p25 - 3: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_4_0p25' - 3: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 3: + interp_atmos_master.sh[35]gridopt=grid0p50 - 3: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_4_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_4_0p50' - 3: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 3: + interp_atmos_master.sh[35]gridopt=grid1p00 - 3: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_4_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_4_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_4_1p00' - 3: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_4 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_4_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_4_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_4_1p00 - 3: 1:0:d=2021032312:HGT:1 mb:24 hour fcst: - 3: 2:18247:d=2021032312:TMP:1 mb:24 hour fcst: - 3: 3:34112:d=2021032312:RH:1 mb:24 hour fcst: - 3: 4:43938:d=2021032312:SPFH:1 mb:24 hour fcst: - 3: 5:62338:d=2021032312:VVEL:1 mb:24 hour fcst: - 3: 6:89192:d=2021032312:DZDT:1 mb:24 hour fcst: - 3: 7:115285:d=2021032312:UGRD:1 mb:24 hour fcst: - 3: 8:133886:d=2021032312:VGRD:1 mb:24 hour fcst: - 3: 9:151587:d=2021032312:ABSV:1 mb:24 hour fcst: - 3: 10:167941:d=2021032312:O3MR:1 mb:24 hour fcst: - 3: 11:189009:d=2021032312:HGT:2 mb:24 hour fcst: - 3: 12:209321:d=2021032312:TMP:2 mb:24 hour fcst: - 3: 13:225925:d=2021032312:RH:2 mb:24 hour fcst: - 3: 14:234506:d=2021032312:SPFH:2 mb:24 hour fcst: - 3: 15:253501:d=2021032312:VVEL:2 mb:24 hour fcst: - 3: 16:282188:d=2021032312:DZDT:2 mb:24 hour fcst: - 3: 17:308674:d=2021032312:UGRD:2 mb:24 hour fcst: - 3: 18:320259:d=2021032312:VGRD:2 mb:24 hour fcst: - 3: 19:338605:d=2021032312:ABSV:2 mb:24 hour fcst: - 3: 20:355688:d=2021032312:O3MR:2 mb:24 hour fcst: - 3: 21:378506:d=2021032312:HGT:3 mb:24 hour fcst: - 3: 22:396342:d=2021032312:TMP:3 mb:24 hour fcst: - 3: 23:412815:d=2021032312:RH:3 mb:24 hour fcst: - 3: 24:418845:d=2021032312:SPFH:3 mb:24 hour fcst: - 3: 25:440103:d=2021032312:VVEL:3 mb:24 hour fcst: - 3: 26:470159:d=2021032312:DZDT:3 mb:24 hour fcst: - 3: 27:496575:d=2021032312:UGRD:3 mb:24 hour fcst: - 3: 28:507958:d=2021032312:VGRD:3 mb:24 hour fcst: - 3: 29:526095:d=2021032312:ABSV:3 mb:24 hour fcst: - 3: 30:543109:d=2021032312:O3MR:3 mb:24 hour fcst: - 3: 31:566482:d=2021032312:HGT:5 mb:24 hour fcst: - 3: + interp_atmos_master.sh[47]export err=0 - 3: + interp_atmos_master.sh[47]err=0 - 3: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 3: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 3: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_4_0p25 - 3: + product_functions.sh[5]local filename=pgb2file_f024_4_0p25 - 3: + product_functions.sh[6]wgrib2 pgb2file_f024_4_0p25 -not_if :RH: -grib pgb2file_f024_4_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_4_0p25.new - 3: 1:0:d=2021032312:HGT:1 mb:24 hour fcst: - 3: 2:409041:d=2021032312:TMP:1 mb:24 hour fcst: - 3: 3:757856:d=2021032312:RH:1 mb:24 hour fcst: - 3: 4:1014158:d=2021032312:SPFH:1 mb:24 hour fcst: - 3: 5:1505366:d=2021032312:VVEL:1 mb:24 hour fcst: - 3: 6:2256586:d=2021032312:DZDT:1 mb:24 hour fcst: - 3: 7:2985525:d=2021032312:UGRD:1 mb:24 hour fcst: - 3: 8:3414946:d=2021032312:VGRD:1 mb:24 hour fcst: - 3: 9:3819166:d=2021032312:ABSV:1 mb:24 hour fcst: - 3: 10:4166285:d=2021032312:O3MR:1 mb:24 hour fcst: - 3: 11:4694211:d=2021032312:HGT:2 mb:24 hour fcst: - 3: 12:5182177:d=2021032312:TMP:2 mb:24 hour fcst: - 3: 13:5554555:d=2021032312:RH:2 mb:24 hour fcst: - 3: 14:5802949:d=2021032312:SPFH:2 mb:24 hour fcst: - 3: 15:6283840:d=2021032312:VVEL:2 mb:24 hour fcst: - 3: 16:7078946:d=2021032312:DZDT:2 mb:24 hour fcst: - 3: 17:7829441:d=2021032312:UGRD:2 mb:24 hour fcst: - 3: 18:8288473:d=2021032312:VGRD:2 mb:24 hour fcst: - 3: 19:8712478:d=2021032312:ABSV:2 mb:24 hour fcst: - 3: 20:9082112:d=2021032312:O3MR:2 mb:24 hour fcst: - 3: 21:9668996:d=2021032312:HGT:3 mb:24 hour fcst: - 3: 22:10070721:d=2021032312:TMP:3 mb:24 hour fcst: - 3: 23:10434986:d=2021032312:RH:3 mb:24 hour fcst: - 3: 24:10605434:d=2021032312:SPFH:3 mb:24 hour fcst: - 3: 25:11148300:d=2021032312:VVEL:3 mb:24 hour fcst: - 3: 26:11990001:d=2021032312:DZDT:3 mb:24 hour fcst: - 3: 27:12728606:d=2021032312:UGRD:3 mb:24 hour fcst: - 3: 28:13173516:d=2021032312:VGRD:3 mb:24 hour fcst: - 3: 29:13588205:d=2021032312:ABSV:3 mb:24 hour fcst: - 3: 30:13951908:d=2021032312:O3MR:3 mb:24 hour fcst: - 3: 31:14556189:d=2021032312:HGT:5 mb:24 hour fcst: - 3: + product_functions.sh[10]rc=0 - 3: + product_functions.sh[11](( rc == 0 )) - 3: + product_functions.sh[11]mv pgb2file_f024_4_0p25.new pgb2file_f024_4_0p25 - 3: + product_functions.sh[12]return 0 - 3: + interp_atmos_master.sh[56]export err=0 - 3: + interp_atmos_master.sh[56]err=0 - 3: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 3: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_4_0p25 -match 'LAND|ICEC' - 3: ++ interp_atmos_master.sh[62]wc -l - 3: + interp_atmos_master.sh[62]var_count=0 - 3: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 3: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 3: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_4_0p50 - 3: + product_functions.sh[5]local filename=pgb2file_f024_4_0p50 - 3: + product_functions.sh[6]wgrib2 pgb2file_f024_4_0p50 -not_if :RH: -grib pgb2file_f024_4_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_4_0p50.new - 3: 1:0:d=2021032312:HGT:1 mb:24 hour fcst: - 3: 2:140441:d=2021032312:TMP:1 mb:24 hour fcst: - 3: 3:258729:d=2021032312:RH:1 mb:24 hour fcst: - 3: 4:330812:d=2021032312:SPFH:1 mb:24 hour fcst: - 3: 5:491383:d=2021032312:VVEL:1 mb:24 hour fcst: - 3: 6:746850:d=2021032312:DZDT:1 mb:24 hour fcst: - 3: 7:986190:d=2021032312:UGRD:1 mb:24 hour fcst: - 3: 8:1133183:d=2021032312:VGRD:1 mb:24 hour fcst: - 3: 9:1271443:d=2021032312:ABSV:1 mb:24 hour fcst: - 3: 10:1389756:d=2021032312:O3MR:1 mb:24 hour fcst: - 3: 11:1570291:d=2021032312:HGT:2 mb:24 hour fcst: - 3: 12:1736145:d=2021032312:TMP:2 mb:24 hour fcst: - 3: 13:1862559:d=2021032312:RH:2 mb:24 hour fcst: - 3: 14:1930215:d=2021032312:SPFH:2 mb:24 hour fcst: - 3: 15:2089744:d=2021032312:VVEL:2 mb:24 hour fcst: - 3: 16:2367321:d=2021032312:DZDT:2 mb:24 hour fcst: - 3: 17:2612571:d=2021032312:UGRD:2 mb:24 hour fcst: - 3: 18:2771268:d=2021032312:VGRD:2 mb:24 hour fcst: - 3: 19:2916784:d=2021032312:ABSV:2 mb:24 hour fcst: - 3: 20:3044408:d=2021032312:O3MR:2 mb:24 hour fcst: - 3: 21:3242857:d=2021032312:HGT:3 mb:24 hour fcst: - 3: 22:3380251:d=2021032312:TMP:3 mb:24 hour fcst: - 3: 23:3504736:d=2021032312:RH:3 mb:24 hour fcst: - 3: 24:3561329:d=2021032312:SPFH:3 mb:24 hour fcst: - 3: 25:3743140:d=2021032312:VVEL:3 mb:24 hour fcst: - 3: 26:4038766:d=2021032312:DZDT:3 mb:24 hour fcst: - 3: 27:4280832:d=2021032312:UGRD:3 mb:24 hour fcst: - 3: 28:4435411:d=2021032312:VGRD:3 mb:24 hour fcst: - 3: 29:4578646:d=2021032312:ABSV:3 mb:24 hour fcst: - 3: 30:4704624:d=2021032312:O3MR:3 mb:24 hour fcst: - 3: 31:4908930:d=2021032312:HGT:5 mb:24 hour fcst: - 3: + product_functions.sh[10]rc=0 - 3: + product_functions.sh[11](( rc == 0 )) - 3: + product_functions.sh[11]mv pgb2file_f024_4_0p50.new pgb2file_f024_4_0p50 - 3: + product_functions.sh[12]return 0 - 3: + interp_atmos_master.sh[56]export err=0 - 3: + interp_atmos_master.sh[56]err=0 - 3: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 3: ++ interp_atmos_master.sh[62]wc -l - 3: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_4_0p50 -match 'LAND|ICEC' - 3: + interp_atmos_master.sh[62]var_count=0 - 3: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 3: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 3: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_4_1p00 - 3: + product_functions.sh[5]local filename=pgb2file_f024_4_1p00 - 3: + product_functions.sh[6]wgrib2 pgb2file_f024_4_1p00 -not_if :RH: -grib pgb2file_f024_4_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_4_1p00.new - 3: 1:0:d=2021032312:HGT:1 mb:24 hour fcst: - 3: 2:46183:d=2021032312:TMP:1 mb:24 hour fcst: - 3: 3:86076:d=2021032312:RH:1 mb:24 hour fcst: - 3: 4:109273:d=2021032312:SPFH:1 mb:24 hour fcst: - 3: 5:158630:d=2021032312:VVEL:1 mb:24 hour fcst: - 3: 6:235002:d=2021032312:DZDT:1 mb:24 hour fcst: - 3: 7:306913:d=2021032312:UGRD:1 mb:24 hour fcst: - 3: 8:354764:d=2021032312:VGRD:1 mb:24 hour fcst: - 3: 9:400605:d=2021032312:ABSV:1 mb:24 hour fcst: - 3: 10:440788:d=2021032312:O3MR:1 mb:24 hour fcst: - 3: 11:497984:d=2021032312:HGT:2 mb:24 hour fcst: - 3: 12:550911:d=2021032312:TMP:2 mb:24 hour fcst: - 3: 13:593129:d=2021032312:RH:2 mb:24 hour fcst: - 3: 14:613202:d=2021032312:SPFH:2 mb:24 hour fcst: - 3: 15:663189:d=2021032312:VVEL:2 mb:24 hour fcst: - 3: 16:745585:d=2021032312:DZDT:2 mb:24 hour fcst: - 3: 17:819372:d=2021032312:UGRD:2 mb:24 hour fcst: - 3: 18:870567:d=2021032312:VGRD:2 mb:24 hour fcst: - 3: 19:918127:d=2021032312:ABSV:2 mb:24 hour fcst: - 3: 20:960804:d=2021032312:O3MR:2 mb:24 hour fcst: - 3: 21:1022553:d=2021032312:HGT:3 mb:24 hour fcst: - 3: 22:1067532:d=2021032312:TMP:3 mb:24 hour fcst: - 3: 23:1108886:d=2021032312:RH:3 mb:24 hour fcst: - 3: 24:1125522:d=2021032312:SPFH:3 mb:24 hour fcst: - 3: 25:1182360:d=2021032312:VVEL:3 mb:24 hour fcst: - 3: 26:1269299:d=2021032312:DZDT:3 mb:24 hour fcst: - 3: 27:1342183:d=2021032312:UGRD:3 mb:24 hour fcst: - 3: 28:1392573:d=2021032312:VGRD:3 mb:24 hour fcst: - 3: 29:1439580:d=2021032312:ABSV:3 mb:24 hour fcst: - 3: 30:1482019:d=2021032312:O3MR:3 mb:24 hour fcst: - 3: 31:1545708:d=2021032312:HGT:5 mb:24 hour fcst: - 3: + product_functions.sh[10]rc=0 - 3: + product_functions.sh[11](( rc == 0 )) - 3: + product_functions.sh[11]mv pgb2file_f024_4_1p00.new pgb2file_f024_4_1p00 - 3: + product_functions.sh[12]return 0 - 3: + interp_atmos_master.sh[56]export err=0 - 3: + interp_atmos_master.sh[56]err=0 - 3: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 3: ++ interp_atmos_master.sh[62]wc -l - 3: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_4_1p00 -match 'LAND|ICEC' - 3: + interp_atmos_master.sh[62]var_count=0 - 3: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 3: + interp_atmos_master.sh[73]exit 0 -23: + bash[8]'[' -z '' ']' -23: + bash[9]case "$-" in -23: + bash[12]__lmod_vx=x -23: + bash[16]'[' -n x ']' -23: + bash[16]set +x -23: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -23: Shell debugging restarted -23: + bash[224]unset __lmod_vx -23: + interp_atmos_master.sh[7]input_file=tmpfile_f024_24 -23: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_24 -23: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -23: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -23: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -23: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -23: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -23: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -23: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -23: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -23: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -23: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -23: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -23: + interp_atmos_master.sh[31]IFS=: -23: + interp_atmos_master.sh[31]read -ra grids -23: + interp_atmos_master.sh[33]output_grids= -23: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -23: + interp_atmos_master.sh[35]gridopt=grid0p25 -23: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_24_0p25' -23: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -23: + interp_atmos_master.sh[35]gridopt=grid0p50 -23: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_24_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_24_0p50' -23: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -23: + interp_atmos_master.sh[35]gridopt=grid1p00 -23: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_24_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_24_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_24_1p00' -23: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_24 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_24_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_24_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_24_1p00 -23: 1:0:d=2021032312:CAPE:180-0 mb above ground:24 hour fcst: -23: 2:14000:d=2021032312:CIN:180-0 mb above ground:24 hour fcst: -23: 3:28112:d=2021032312:HPBL:surface:24 hour fcst: -23: 4:59266:d=2021032312:RH:0.33-1 sigma layer:24 hour fcst: -23: 5:77642:d=2021032312:RH:0.44-1 sigma layer:24 hour fcst: -23: 6:95644:d=2021032312:RH:0.72-0.94 sigma layer:24 hour fcst: -23: 7:114121:d=2021032312:RH:0.44-0.72 sigma layer:24 hour fcst: -23: 8:133429:d=2021032312:TMP:0.995 sigma level:24 hour fcst: -23: 9:154267:d=2021032312:POT:0.995 sigma level:24 hour fcst: -23: 10:175088:d=2021032312:RH:0.995 sigma level:24 hour fcst: -23: 11:192885:d=2021032312:UGRD:0.995 sigma level:24 hour fcst: -23: 12:214044:d=2021032312:VGRD:0.995 sigma level:24 hour fcst: -23: 13:235561:d=2021032312:VVEL:0.995 sigma level:24 hour fcst: -23: 14:256444:d=2021032312:CAPE:90-0 mb above ground:24 hour fcst: -23: 15:269255:d=2021032312:CIN:90-0 mb above ground:24 hour fcst: -23: 16:283621:d=2021032312:CAPE:255-0 mb above ground:24 hour fcst: -23: 17:296739:d=2021032312:CIN:255-0 mb above ground:24 hour fcst: -23: 18:308098:d=2021032312:PLPL:255-0 mb above ground:24 hour fcst: -23: 19:331819:d=2021032312:LAND:surface:24 hour fcst: -23: 20:333760:d=2021032312:ICEC:surface:24 hour fcst: -23: 21:336925:d=2021032312:ALBDO:surface:18-24 hour ave fcst: -23: 22:351687:d=2021032312:ICETMP:surface:24 hour fcst: -23: 23:358417:d=2021032312:UGRD:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 24:369140:d=2021032312:VGRD:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 25:379853:d=2021032312:TMP:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 26:394228:d=2021032312:HGT:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 27:411601:d=2021032312:PRES:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 28:430035:d=2021032312:VWSH:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 29:441992:d=2021032312:UGRD:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 30:452867:d=2021032312:VGRD:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 31:467677:d=2021032312:TMP:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 32:482093:d=2021032312:HGT:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 33:499595:d=2021032312:PRES:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 34:516568:d=2021032312:VWSH:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: + interp_atmos_master.sh[47]export err=0 -23: + interp_atmos_master.sh[47]err=0 -23: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -23: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -23: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_24_0p25 -23: + product_functions.sh[5]local filename=pgb2file_f024_24_0p25 -23: + product_functions.sh[6]wgrib2 pgb2file_f024_24_0p25 -not_if :RH: -grib pgb2file_f024_24_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_24_0p25.new -23: 1:0:d=2021032312:CAPE:180-0 mb above ground:24 hour fcst: -23: 2:298073:d=2021032312:CIN:180-0 mb above ground:24 hour fcst: -23: 3:580801:d=2021032312:HPBL:surface:24 hour fcst: -23: 4:1445390:d=2021032312:RH:0.33-1 sigma layer:24 hour fcst: -23: 5:1856748:d=2021032312:RH:0.44-1 sigma layer:24 hour fcst: -23: 6:2258424:d=2021032312:RH:0.72-0.94 sigma layer:24 hour fcst: -23: 7:2675771:d=2021032312:RH:0.44-0.72 sigma layer:24 hour fcst: -23: 8:3115582:d=2021032312:TMP:0.995 sigma level:24 hour fcst: -23: 9:3601106:d=2021032312:POT:0.995 sigma level:24 hour fcst: -23: 10:4090440:d=2021032312:RH:0.995 sigma level:24 hour fcst: -23: 11:4476759:d=2021032312:UGRD:0.995 sigma level:24 hour fcst: -23: 12:4989591:d=2021032312:VGRD:0.995 sigma level:24 hour fcst: -23: 13:5517982:d=2021032312:VVEL:0.995 sigma level:24 hour fcst: -23: 14:6013998:d=2021032312:CAPE:90-0 mb above ground:24 hour fcst: -23: 15:6286174:d=2021032312:CIN:90-0 mb above ground:24 hour fcst: -23: 16:6582038:d=2021032312:CAPE:255-0 mb above ground:24 hour fcst: -23: 17:6880262:d=2021032312:CIN:255-0 mb above ground:24 hour fcst: -23: 18:7112507:d=2021032312:PLPL:255-0 mb above ground:24 hour fcst: -23: 19:7670700:d=2021032312:LAND:surface:24 hour fcst: -23: 20:7690620:d=2021032312:ICEC:surface:24 hour fcst: -23: 21:7745820:d=2021032312:ALBDO:surface:18-24 hour ave fcst: -23: 22:8047006:d=2021032312:ICETMP:surface:24 hour fcst: -23: 23:8269101:d=2021032312:UGRD:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 24:8584400:d=2021032312:VGRD:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 25:8896006:d=2021032312:TMP:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 26:9321942:d=2021032312:HGT:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 27:9873686:d=2021032312:PRES:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 28:10457541:d=2021032312:VWSH:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 29:10810515:d=2021032312:UGRD:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 30:11269581:d=2021032312:VGRD:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 31:11723381:d=2021032312:TMP:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 32:12152379:d=2021032312:HGT:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 33:12706072:d=2021032312:PRES:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 34:13232187:d=2021032312:VWSH:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: + product_functions.sh[10]rc=0 -23: + product_functions.sh[11](( rc == 0 )) -23: + product_functions.sh[11]mv pgb2file_f024_24_0p25.new pgb2file_f024_24_0p25 -23: + product_functions.sh[12]return 0 -23: + interp_atmos_master.sh[56]export err=0 -23: + interp_atmos_master.sh[56]err=0 -23: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -23: ++ interp_atmos_master.sh[62]wc -l -23: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_24_0p25 -match 'LAND|ICEC' -23: + interp_atmos_master.sh[62]var_count=2 -23: + interp_atmos_master.sh[63][[ 2 -eq 2 ]] -23: + interp_atmos_master.sh[64]mod_icec pgb2file_f024_24_0p25 -23: + product_functions.sh[17]local filename=pgb2file_f024_24_0p25 -23: + product_functions.sh[18]wgrib2 pgb2file_f024_24_0p25 -if LAND -rpn sto_1 -fi -if ICEC -rpn 'rcl_1:0:==:*' -fi -set_grib_type same -set_scaling same same -grib_out pgb2file_f024_24_0p25.new -23: 1:0:d=2021032312:CAPE:180-0 mb above ground:24 hour fcst: -23: 2:298073:d=2021032312:CIN:180-0 mb above ground:24 hour fcst: -23: 3:580801:d=2021032312:HPBL:surface:24 hour fcst: -23: 4:1445390:d=2021032312:RH:0.33-1 sigma layer:24 hour fcst: -23: 5:1856748:d=2021032312:RH:0.44-1 sigma layer:24 hour fcst: -23: 6:2258424:d=2021032312:RH:0.72-0.94 sigma layer:24 hour fcst: -23: 7:2675771:d=2021032312:RH:0.44-0.72 sigma layer:24 hour fcst: -23: 8:3115582:d=2021032312:TMP:0.995 sigma level:24 hour fcst: -23: 9:3601106:d=2021032312:POT:0.995 sigma level:24 hour fcst: -23: 10:4090440:d=2021032312:RH:0.995 sigma level:24 hour fcst: -23: 11:4476759:d=2021032312:UGRD:0.995 sigma level:24 hour fcst: -23: 12:4989591:d=2021032312:VGRD:0.995 sigma level:24 hour fcst: -23: 13:5517982:d=2021032312:VVEL:0.995 sigma level:24 hour fcst: -23: 14:6013998:d=2021032312:CAPE:90-0 mb above ground:24 hour fcst: -23: 15:6286174:d=2021032312:CIN:90-0 mb above ground:24 hour fcst: -23: 16:6582038:d=2021032312:CAPE:255-0 mb above ground:24 hour fcst: -23: 17:6880262:d=2021032312:CIN:255-0 mb above ground:24 hour fcst: -23: 18:7112507:d=2021032312:PLPL:255-0 mb above ground:24 hour fcst: -23: 19:7670700:d=2021032312:LAND:surface:24 hour fcst: -23: 20:7690620:d=2021032312:ICEC:surface:24 hour fcst: -23: 21:7745820:d=2021032312:ALBDO:surface:18-24 hour ave fcst: -23: 22:8047006:d=2021032312:ICETMP:surface:24 hour fcst: -23: 23:8269101:d=2021032312:UGRD:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 24:8584400:d=2021032312:VGRD:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 25:8896006:d=2021032312:TMP:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 26:9321942:d=2021032312:HGT:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 27:9873686:d=2021032312:PRES:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 28:10457541:d=2021032312:VWSH:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 29:10810515:d=2021032312:UGRD:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 30:11269581:d=2021032312:VGRD:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 31:11723381:d=2021032312:TMP:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 32:12152379:d=2021032312:HGT:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 33:12706072:d=2021032312:PRES:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 34:13232187:d=2021032312:VWSH:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: + product_functions.sh[24]rc=0 -23: + product_functions.sh[25](( rc == 0 )) -23: + product_functions.sh[25]mv pgb2file_f024_24_0p25.new pgb2file_f024_24_0p25 -23: + product_functions.sh[26]return 0 -23: + interp_atmos_master.sh[65]export err=0 -23: + interp_atmos_master.sh[65]err=0 -23: + interp_atmos_master.sh[66][[ 0 -ne 0 ]] -23: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -23: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_24_0p50 -23: + product_functions.sh[5]local filename=pgb2file_f024_24_0p50 -23: + product_functions.sh[6]wgrib2 pgb2file_f024_24_0p50 -not_if :RH: -grib pgb2file_f024_24_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_24_0p50.new -23: 1:0:d=2021032312:CAPE:180-0 mb above ground:24 hour fcst: -23: 2:104759:d=2021032312:CIN:180-0 mb above ground:24 hour fcst: -23: 3:205703:d=2021032312:HPBL:surface:24 hour fcst: -23: 4:503256:d=2021032312:RH:0.33-1 sigma layer:24 hour fcst: -23: 5:645781:d=2021032312:RH:0.44-1 sigma layer:24 hour fcst: -23: 6:784811:d=2021032312:RH:0.72-0.94 sigma layer:24 hour fcst: -23: 7:929853:d=2021032312:RH:0.44-0.72 sigma layer:24 hour fcst: -23: 8:1082323:d=2021032312:TMP:0.995 sigma level:24 hour fcst: -23: 9:1250846:d=2021032312:POT:0.995 sigma level:24 hour fcst: -23: 10:1418749:d=2021032312:RH:0.995 sigma level:24 hour fcst: -23: 11:1556806:d=2021032312:UGRD:0.995 sigma level:24 hour fcst: -23: 12:1733304:d=2021032312:VGRD:0.995 sigma level:24 hour fcst: -23: 13:1914512:d=2021032312:VVEL:0.995 sigma level:24 hour fcst: -23: 14:2085263:d=2021032312:CAPE:90-0 mb above ground:24 hour fcst: -23: 15:2179952:d=2021032312:CIN:90-0 mb above ground:24 hour fcst: -23: 16:2285431:d=2021032312:CAPE:255-0 mb above ground:24 hour fcst: -23: 17:2389877:d=2021032312:CIN:255-0 mb above ground:24 hour fcst: -23: 18:2471438:d=2021032312:PLPL:255-0 mb above ground:24 hour fcst: -23: 19:2676794:d=2021032312:LAND:surface:24 hour fcst: -23: 20:2685619:d=2021032312:ICEC:surface:24 hour fcst: -23: 21:2706595:d=2021032312:ALBDO:surface:18-24 hour ave fcst: -23: 22:2814874:d=2021032312:ICETMP:surface:24 hour fcst: -23: 23:2882201:d=2021032312:UGRD:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 24:2977183:d=2021032312:VGRD:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 25:3071820:d=2021032312:TMP:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 26:3206649:d=2021032312:HGT:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 27:3379069:d=2021032312:PRES:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 28:3563308:d=2021032312:VWSH:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 29:3671229:d=2021032312:UGRD:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 30:3815153:d=2021032312:VGRD:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 31:3956848:d=2021032312:TMP:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 32:4091417:d=2021032312:HGT:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 33:4264787:d=2021032312:PRES:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 34:4431287:d=2021032312:VWSH:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: + product_functions.sh[10]rc=0 -23: + product_functions.sh[11](( rc == 0 )) -23: + product_functions.sh[11]mv pgb2file_f024_24_0p50.new pgb2file_f024_24_0p50 -23: + product_functions.sh[12]return 0 -23: + interp_atmos_master.sh[56]export err=0 -23: + interp_atmos_master.sh[56]err=0 -23: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -23: ++ interp_atmos_master.sh[62]wc -l -23: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_24_0p50 -match 'LAND|ICEC' -23: + interp_atmos_master.sh[62]var_count=2 -23: + interp_atmos_master.sh[63][[ 2 -eq 2 ]] -23: + interp_atmos_master.sh[64]mod_icec pgb2file_f024_24_0p50 -23: + product_functions.sh[17]local filename=pgb2file_f024_24_0p50 -23: + product_functions.sh[18]wgrib2 pgb2file_f024_24_0p50 -if LAND -rpn sto_1 -fi -if ICEC -rpn 'rcl_1:0:==:*' -fi -set_grib_type same -set_scaling same same -grib_out pgb2file_f024_24_0p50.new -23: 1:0:d=2021032312:CAPE:180-0 mb above ground:24 hour fcst: -23: 2:104759:d=2021032312:CIN:180-0 mb above ground:24 hour fcst: -23: 3:205703:d=2021032312:HPBL:surface:24 hour fcst: -23: 4:503256:d=2021032312:RH:0.33-1 sigma layer:24 hour fcst: -23: 5:645781:d=2021032312:RH:0.44-1 sigma layer:24 hour fcst: -23: 6:784811:d=2021032312:RH:0.72-0.94 sigma layer:24 hour fcst: -23: 7:929853:d=2021032312:RH:0.44-0.72 sigma layer:24 hour fcst: -23: 8:1082323:d=2021032312:TMP:0.995 sigma level:24 hour fcst: -23: 9:1250846:d=2021032312:POT:0.995 sigma level:24 hour fcst: -23: 10:1418749:d=2021032312:RH:0.995 sigma level:24 hour fcst: -23: 11:1556806:d=2021032312:UGRD:0.995 sigma level:24 hour fcst: -23: 12:1733304:d=2021032312:VGRD:0.995 sigma level:24 hour fcst: -23: 13:1914512:d=2021032312:VVEL:0.995 sigma level:24 hour fcst: -23: 14:2085263:d=2021032312:CAPE:90-0 mb above ground:24 hour fcst: -23: 15:2179952:d=2021032312:CIN:90-0 mb above ground:24 hour fcst: -23: 16:2285431:d=2021032312:CAPE:255-0 mb above ground:24 hour fcst: -23: 17:2389877:d=2021032312:CIN:255-0 mb above ground:24 hour fcst: -23: 18:2471438:d=2021032312:PLPL:255-0 mb above ground:24 hour fcst: -23: 19:2676794:d=2021032312:LAND:surface:24 hour fcst: -23: 20:2685619:d=2021032312:ICEC:surface:24 hour fcst: -23: 21:2706595:d=2021032312:ALBDO:surface:18-24 hour ave fcst: -23: 22:2814874:d=2021032312:ICETMP:surface:24 hour fcst: -23: 23:2882201:d=2021032312:UGRD:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 24:2977183:d=2021032312:VGRD:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 25:3071820:d=2021032312:TMP:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 26:3206649:d=2021032312:HGT:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 27:3379069:d=2021032312:PRES:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 28:3563308:d=2021032312:VWSH:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 29:3671229:d=2021032312:UGRD:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 30:3815153:d=2021032312:VGRD:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 31:3956848:d=2021032312:TMP:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 32:4091417:d=2021032312:HGT:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 33:4264787:d=2021032312:PRES:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 34:4431287:d=2021032312:VWSH:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: + product_functions.sh[24]rc=0 -23: + product_functions.sh[25](( rc == 0 )) -23: + product_functions.sh[25]mv pgb2file_f024_24_0p50.new pgb2file_f024_24_0p50 -23: + product_functions.sh[26]return 0 -23: + interp_atmos_master.sh[65]export err=0 -23: + interp_atmos_master.sh[65]err=0 -23: + interp_atmos_master.sh[66][[ 0 -ne 0 ]] -23: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -23: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_24_1p00 -23: + product_functions.sh[5]local filename=pgb2file_f024_24_1p00 -23: + product_functions.sh[6]wgrib2 pgb2file_f024_24_1p00 -not_if :RH: -grib pgb2file_f024_24_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_24_1p00.new -23: 1:0:d=2021032312:CAPE:180-0 mb above ground:24 hour fcst: -23: 2:35204:d=2021032312:CIN:180-0 mb above ground:24 hour fcst: -23: 3:70694:d=2021032312:HPBL:surface:24 hour fcst: -23: 4:159042:d=2021032312:RH:0.33-1 sigma layer:24 hour fcst: -23: 5:206424:d=2021032312:RH:0.44-1 sigma layer:24 hour fcst: -23: 6:252797:d=2021032312:RH:0.72-0.94 sigma layer:24 hour fcst: -23: 7:301061:d=2021032312:RH:0.44-0.72 sigma layer:24 hour fcst: -23: 8:351107:d=2021032312:TMP:0.995 sigma level:24 hour fcst: -23: 9:405885:d=2021032312:POT:0.995 sigma level:24 hour fcst: -23: 10:460597:d=2021032312:RH:0.995 sigma level:24 hour fcst: -23: 11:506523:d=2021032312:UGRD:0.995 sigma level:24 hour fcst: -23: 12:562744:d=2021032312:VGRD:0.995 sigma level:24 hour fcst: -23: 13:620407:d=2021032312:VVEL:0.995 sigma level:24 hour fcst: -23: 14:675329:d=2021032312:CAPE:90-0 mb above ground:24 hour fcst: -23: 15:707419:d=2021032312:CIN:90-0 mb above ground:24 hour fcst: -23: 16:744003:d=2021032312:CAPE:255-0 mb above ground:24 hour fcst: -23: 17:778244:d=2021032312:CIN:255-0 mb above ground:24 hour fcst: -23: 18:806611:d=2021032312:PLPL:255-0 mb above ground:24 hour fcst: -23: 19:871968:d=2021032312:LAND:surface:24 hour fcst: -23: 20:875747:d=2021032312:ICEC:surface:24 hour fcst: -23: 21:883717:d=2021032312:ALBDO:surface:18-24 hour ave fcst: -23: 22:920367:d=2021032312:ICETMP:surface:24 hour fcst: -23: 23:940282:d=2021032312:UGRD:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 24:969377:d=2021032312:VGRD:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 25:998499:d=2021032312:TMP:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 26:1039203:d=2021032312:HGT:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 27:1089788:d=2021032312:PRES:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 28:1143420:d=2021032312:VWSH:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 29:1176091:d=2021032312:UGRD:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 30:1218747:d=2021032312:VGRD:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 31:1261116:d=2021032312:TMP:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 32:1301885:d=2021032312:HGT:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 33:1352568:d=2021032312:PRES:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 34:1401878:d=2021032312:VWSH:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: + product_functions.sh[10]rc=0 -23: + product_functions.sh[11](( rc == 0 )) -23: + product_functions.sh[11]mv pgb2file_f024_24_1p00.new pgb2file_f024_24_1p00 -23: + product_functions.sh[12]return 0 -23: + interp_atmos_master.sh[56]export err=0 -23: + interp_atmos_master.sh[56]err=0 -23: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -23: ++ interp_atmos_master.sh[62]wc -l -23: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_24_1p00 -match 'LAND|ICEC' -23: + interp_atmos_master.sh[62]var_count=2 -23: + interp_atmos_master.sh[63][[ 2 -eq 2 ]] -23: + interp_atmos_master.sh[64]mod_icec pgb2file_f024_24_1p00 -23: + product_functions.sh[17]local filename=pgb2file_f024_24_1p00 -23: + product_functions.sh[18]wgrib2 pgb2file_f024_24_1p00 -if LAND -rpn sto_1 -fi -if ICEC -rpn 'rcl_1:0:==:*' -fi -set_grib_type same -set_scaling same same -grib_out pgb2file_f024_24_1p00.new -23: 1:0:d=2021032312:CAPE:180-0 mb above ground:24 hour fcst: -23: 2:35204:d=2021032312:CIN:180-0 mb above ground:24 hour fcst: -23: 3:70694:d=2021032312:HPBL:surface:24 hour fcst: -23: 4:159042:d=2021032312:RH:0.33-1 sigma layer:24 hour fcst: -23: 5:206424:d=2021032312:RH:0.44-1 sigma layer:24 hour fcst: -23: 6:252797:d=2021032312:RH:0.72-0.94 sigma layer:24 hour fcst: -23: 7:301061:d=2021032312:RH:0.44-0.72 sigma layer:24 hour fcst: -23: 8:351107:d=2021032312:TMP:0.995 sigma level:24 hour fcst: -23: 9:405885:d=2021032312:POT:0.995 sigma level:24 hour fcst: -23: 10:460597:d=2021032312:RH:0.995 sigma level:24 hour fcst: -23: 11:506523:d=2021032312:UGRD:0.995 sigma level:24 hour fcst: -23: 12:562744:d=2021032312:VGRD:0.995 sigma level:24 hour fcst: -23: 13:620407:d=2021032312:VVEL:0.995 sigma level:24 hour fcst: -23: 14:675329:d=2021032312:CAPE:90-0 mb above ground:24 hour fcst: -23: 15:707419:d=2021032312:CIN:90-0 mb above ground:24 hour fcst: -23: 16:744003:d=2021032312:CAPE:255-0 mb above ground:24 hour fcst: -23: 17:778244:d=2021032312:CIN:255-0 mb above ground:24 hour fcst: -23: 18:806611:d=2021032312:PLPL:255-0 mb above ground:24 hour fcst: -23: 19:871968:d=2021032312:LAND:surface:24 hour fcst: -23: 20:875747:d=2021032312:ICEC:surface:24 hour fcst: -23: 21:883717:d=2021032312:ALBDO:surface:18-24 hour ave fcst: -23: 22:920367:d=2021032312:ICETMP:surface:24 hour fcst: -23: 23:940282:d=2021032312:UGRD:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 24:969377:d=2021032312:VGRD:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 25:998499:d=2021032312:TMP:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 26:1039203:d=2021032312:HGT:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 27:1089788:d=2021032312:PRES:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 28:1143420:d=2021032312:VWSH:PV=2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 29:1176091:d=2021032312:UGRD:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 30:1218747:d=2021032312:VGRD:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 31:1261116:d=2021032312:TMP:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 32:1301885:d=2021032312:HGT:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 33:1352568:d=2021032312:PRES:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 34:1401878:d=2021032312:VWSH:PV=-2e-06 (Km^2/kg/s) surface:24 hour fcst: -23: + product_functions.sh[24]rc=0 -23: + product_functions.sh[25](( rc == 0 )) -23: + product_functions.sh[25]mv pgb2file_f024_24_1p00.new pgb2file_f024_24_1p00 -23: + product_functions.sh[26]return 0 -23: + interp_atmos_master.sh[65]export err=0 -23: + interp_atmos_master.sh[65]err=0 -23: + interp_atmos_master.sh[66][[ 0 -ne 0 ]] -23: + interp_atmos_master.sh[73]exit 0 - 4: + bash[8]'[' -z '' ']' - 4: + bash[9]case "$-" in - 4: + bash[12]__lmod_vx=x - 4: + bash[16]'[' -n x ']' - 4: + bash[16]set +x - 4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 4: Shell debugging restarted - 4: + bash[224]unset __lmod_vx - 4: + interp_atmos_master.sh[7]input_file=tmpfile_f024_5 - 4: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_5 - 4: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 - 4: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 4: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 4: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 4: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 4: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 4: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 4: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 4: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 4: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 4: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 4: + interp_atmos_master.sh[31]IFS=: - 4: + interp_atmos_master.sh[31]read -ra grids - 4: + interp_atmos_master.sh[33]output_grids= - 4: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 4: + interp_atmos_master.sh[35]gridopt=grid0p25 - 4: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_5_0p25' - 4: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 4: + interp_atmos_master.sh[35]gridopt=grid0p50 - 4: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_5_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_5_0p50' - 4: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 4: + interp_atmos_master.sh[35]gridopt=grid1p00 - 4: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_5_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_5_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_5_1p00' - 4: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_5 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_5_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_5_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_5_1p00 - 4: 1:0:d=2021032312:TMP:5 mb:24 hour fcst: - 4: 2:16629:d=2021032312:RH:5 mb:24 hour fcst: - 4: 3:25671:d=2021032312:SPFH:5 mb:24 hour fcst: - 4: 4:46419:d=2021032312:VVEL:5 mb:24 hour fcst: - 4: 5:75603:d=2021032312:DZDT:5 mb:24 hour fcst: - 4: 6:101668:d=2021032312:UGRD:5 mb:24 hour fcst: - 4: 7:113260:d=2021032312:VGRD:5 mb:24 hour fcst: - 4: 8:131830:d=2021032312:ABSV:5 mb:24 hour fcst: - 4: 9:149451:d=2021032312:O3MR:5 mb:24 hour fcst: - 4: 10:172111:d=2021032312:HGT:7 mb:24 hour fcst: - 4: 11:189870:d=2021032312:TMP:7 mb:24 hour fcst: - 4: 12:206059:d=2021032312:RH:7 mb:24 hour fcst: - 4: 13:212178:d=2021032312:SPFH:7 mb:24 hour fcst: - 4: 14:231601:d=2021032312:VVEL:7 mb:24 hour fcst: - 4: 15:256275:d=2021032312:DZDT:7 mb:24 hour fcst: - 4: 16:281784:d=2021032312:UGRD:7 mb:24 hour fcst: - 4: 17:293198:d=2021032312:VGRD:7 mb:24 hour fcst: - 4: 18:311890:d=2021032312:ABSV:7 mb:24 hour fcst: - 4: 19:329699:d=2021032312:O3MR:7 mb:24 hour fcst: - 4: 20:353473:d=2021032312:HGT:10 mb:24 hour fcst: - 4: 21:373321:d=2021032312:TMP:10 mb:24 hour fcst: - 4: 22:389477:d=2021032312:RH:10 mb:24 hour fcst: - 4: 23:397119:d=2021032312:SPFH:10 mb:24 hour fcst: - 4: 24:415365:d=2021032312:VVEL:10 mb:24 hour fcst: - 4: 25:440887:d=2021032312:DZDT:10 mb:24 hour fcst: - 4: 26:466297:d=2021032312:UGRD:10 mb:24 hour fcst: - 4: 27:477638:d=2021032312:VGRD:10 mb:24 hour fcst: - 4: 28:496301:d=2021032312:ABSV:10 mb:24 hour fcst: - 4: 29:514135:d=2021032312:O3MR:10 mb:24 hour fcst: - 4: 30:532913:d=2021032312:HGT:15 mb:24 hour fcst: - 4: 31:552424:d=2021032312:TMP:15 mb:24 hour fcst: - 4: + interp_atmos_master.sh[47]export err=0 - 4: + interp_atmos_master.sh[47]err=0 - 4: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 4: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 4: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_5_0p25 - 4: + product_functions.sh[5]local filename=pgb2file_f024_5_0p25 - 4: + product_functions.sh[6]wgrib2 pgb2file_f024_5_0p25 -not_if :RH: -grib pgb2file_f024_5_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_5_0p25.new - 4: 1:0:d=2021032312:TMP:5 mb:24 hour fcst: - 4: 2:365675:d=2021032312:RH:5 mb:24 hour fcst: - 4: 3:617005:d=2021032312:SPFH:5 mb:24 hour fcst: - 4: 4:1149045:d=2021032312:VVEL:5 mb:24 hour fcst: - 4: 5:1972602:d=2021032312:DZDT:5 mb:24 hour fcst: - 4: 6:2700006:d=2021032312:UGRD:5 mb:24 hour fcst: - 4: 7:3155905:d=2021032312:VGRD:5 mb:24 hour fcst: - 4: 8:3583670:d=2021032312:ABSV:5 mb:24 hour fcst: - 4: 9:3970721:d=2021032312:O3MR:5 mb:24 hour fcst: - 4: 10:4537598:d=2021032312:HGT:7 mb:24 hour fcst: - 4: 11:4934049:d=2021032312:TMP:7 mb:24 hour fcst: - 4: 12:5286205:d=2021032312:RH:7 mb:24 hour fcst: - 4: 13:5460954:d=2021032312:SPFH:7 mb:24 hour fcst: - 4: 14:5942778:d=2021032312:VVEL:7 mb:24 hour fcst: - 4: 15:6605699:d=2021032312:DZDT:7 mb:24 hour fcst: - 4: 16:7313793:d=2021032312:UGRD:7 mb:24 hour fcst: - 4: 17:7760964:d=2021032312:VGRD:7 mb:24 hour fcst: - 4: 18:8192964:d=2021032312:ABSV:7 mb:24 hour fcst: - 4: 19:8583487:d=2021032312:O3MR:7 mb:24 hour fcst: - 4: 20:9189064:d=2021032312:HGT:10 mb:24 hour fcst: - 4: 21:9660801:d=2021032312:TMP:10 mb:24 hour fcst: - 4: 22:10016805:d=2021032312:RH:10 mb:24 hour fcst: - 4: 23:10240822:d=2021032312:SPFH:10 mb:24 hour fcst: - 4: 24:10685072:d=2021032312:VVEL:10 mb:24 hour fcst: - 4: 25:11381521:d=2021032312:DZDT:10 mb:24 hour fcst: - 4: 26:12081110:d=2021032312:UGRD:10 mb:24 hour fcst: - 4: 27:12521887:d=2021032312:VGRD:10 mb:24 hour fcst: - 4: 28:12952285:d=2021032312:ABSV:10 mb:24 hour fcst: - 4: 29:13346816:d=2021032312:O3MR:10 mb:24 hour fcst: - 4: 30:13776234:d=2021032312:HGT:15 mb:24 hour fcst: - 4: 31:14238211:d=2021032312:TMP:15 mb:24 hour fcst: - 4: + product_functions.sh[10]rc=0 - 4: + product_functions.sh[11](( rc == 0 )) - 4: + product_functions.sh[11]mv pgb2file_f024_5_0p25.new pgb2file_f024_5_0p25 - 4: + product_functions.sh[12]return 0 - 4: + interp_atmos_master.sh[56]export err=0 - 4: + interp_atmos_master.sh[56]err=0 - 4: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 4: ++ interp_atmos_master.sh[62]wc -l - 4: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_5_0p25 -match 'LAND|ICEC' - 4: + interp_atmos_master.sh[62]var_count=0 - 4: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 4: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 4: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_5_0p50 - 4: + product_functions.sh[5]local filename=pgb2file_f024_5_0p50 - 4: + product_functions.sh[6]wgrib2 pgb2file_f024_5_0p50 -not_if :RH: -grib pgb2file_f024_5_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_5_0p50.new - 4: 1:0:d=2021032312:TMP:5 mb:24 hour fcst: - 4: 2:125712:d=2021032312:RH:5 mb:24 hour fcst: - 4: 3:194506:d=2021032312:SPFH:5 mb:24 hour fcst: - 4: 4:371002:d=2021032312:VVEL:5 mb:24 hour fcst: - 4: 5:656336:d=2021032312:DZDT:5 mb:24 hour fcst: - 4: 6:894951:d=2021032312:UGRD:5 mb:24 hour fcst: - 4: 7:1053082:d=2021032312:VGRD:5 mb:24 hour fcst: - 4: 8:1201437:d=2021032312:ABSV:5 mb:24 hour fcst: - 4: 9:1334253:d=2021032312:O3MR:5 mb:24 hour fcst: - 4: 10:1527561:d=2021032312:HGT:7 mb:24 hour fcst: - 4: 11:1663248:d=2021032312:TMP:7 mb:24 hour fcst: - 4: 12:1784973:d=2021032312:RH:7 mb:24 hour fcst: - 4: 13:1841322:d=2021032312:SPFH:7 mb:24 hour fcst: - 4: 14:2000568:d=2021032312:VVEL:7 mb:24 hour fcst: - 4: 15:2225404:d=2021032312:DZDT:7 mb:24 hour fcst: - 4: 16:2457786:d=2021032312:UGRD:7 mb:24 hour fcst: - 4: 17:2612983:d=2021032312:VGRD:7 mb:24 hour fcst: - 4: 18:2763158:d=2021032312:ABSV:7 mb:24 hour fcst: - 4: 19:2899048:d=2021032312:O3MR:7 mb:24 hour fcst: - 4: 20:3105618:d=2021032312:HGT:10 mb:24 hour fcst: - 4: 21:3266423:d=2021032312:TMP:10 mb:24 hour fcst: - 4: 22:3388985:d=2021032312:RH:10 mb:24 hour fcst: - 4: 23:3453716:d=2021032312:SPFH:10 mb:24 hour fcst: - 4: 24:3600733:d=2021032312:VVEL:10 mb:24 hour fcst: - 4: 25:3839259:d=2021032312:DZDT:10 mb:24 hour fcst: - 4: 26:4067807:d=2021032312:UGRD:10 mb:24 hour fcst: - 4: 27:4221725:d=2021032312:VGRD:10 mb:24 hour fcst: - 4: 28:4371556:d=2021032312:ABSV:10 mb:24 hour fcst: - 4: 29:4508439:d=2021032312:O3MR:10 mb:24 hour fcst: - 4: 30:4655591:d=2021032312:HGT:15 mb:24 hour fcst: - 4: 31:4812632:d=2021032312:TMP:15 mb:24 hour fcst: - 4: + product_functions.sh[10]rc=0 - 4: + product_functions.sh[11](( rc == 0 )) - 4: + product_functions.sh[11]mv pgb2file_f024_5_0p50.new pgb2file_f024_5_0p50 - 4: + product_functions.sh[12]return 0 - 4: + interp_atmos_master.sh[56]export err=0 - 4: + interp_atmos_master.sh[56]err=0 - 4: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 4: ++ interp_atmos_master.sh[62]wc -l - 4: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_5_0p50 -match 'LAND|ICEC' - 4: + interp_atmos_master.sh[62]var_count=0 - 4: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 4: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 4: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_5_1p00 - 4: + product_functions.sh[5]local filename=pgb2file_f024_5_1p00 - 4: + product_functions.sh[6]wgrib2 pgb2file_f024_5_1p00 -not_if :RH: -grib pgb2file_f024_5_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_5_1p00.new - 4: 1:0:d=2021032312:TMP:5 mb:24 hour fcst: - 4: 2:41786:d=2021032312:RH:5 mb:24 hour fcst: - 4: 3:62545:d=2021032312:SPFH:5 mb:24 hour fcst: - 4: 4:118226:d=2021032312:VVEL:5 mb:24 hour fcst: - 4: 5:202440:d=2021032312:DZDT:5 mb:24 hour fcst: - 4: 6:274560:d=2021032312:UGRD:5 mb:24 hour fcst: - 4: 7:325640:d=2021032312:VGRD:5 mb:24 hour fcst: - 4: 8:374548:d=2021032312:ABSV:5 mb:24 hour fcst: - 4: 9:418975:d=2021032312:O3MR:5 mb:24 hour fcst: - 4: 10:479958:d=2021032312:HGT:7 mb:24 hour fcst: - 4: 11:524399:d=2021032312:TMP:7 mb:24 hour fcst: - 4: 12:565093:d=2021032312:RH:7 mb:24 hour fcst: - 4: 13:581712:d=2021032312:SPFH:7 mb:24 hour fcst: - 4: 14:632268:d=2021032312:VVEL:7 mb:24 hour fcst: - 4: 15:700623:d=2021032312:DZDT:7 mb:24 hour fcst: - 4: 16:770570:d=2021032312:UGRD:7 mb:24 hour fcst: - 4: 17:821284:d=2021032312:VGRD:7 mb:24 hour fcst: - 4: 18:870298:d=2021032312:ABSV:7 mb:24 hour fcst: - 4: 19:915485:d=2021032312:O3MR:7 mb:24 hour fcst: - 4: 20:979884:d=2021032312:HGT:10 mb:24 hour fcst: - 4: 21:1031404:d=2021032312:TMP:10 mb:24 hour fcst: - 4: 22:1072146:d=2021032312:RH:10 mb:24 hour fcst: - 4: 23:1090768:d=2021032312:SPFH:10 mb:24 hour fcst: - 4: 24:1137938:d=2021032312:VVEL:10 mb:24 hour fcst: - 4: 25:1209895:d=2021032312:DZDT:10 mb:24 hour fcst: - 4: 26:1279175:d=2021032312:UGRD:10 mb:24 hour fcst: - 4: 27:1329461:d=2021032312:VGRD:10 mb:24 hour fcst: - 4: 28:1378361:d=2021032312:ABSV:10 mb:24 hour fcst: - 4: 29:1423532:d=2021032312:O3MR:10 mb:24 hour fcst: - 4: 30:1471816:d=2021032312:HGT:15 mb:24 hour fcst: - 4: 31:1522399:d=2021032312:TMP:15 mb:24 hour fcst: - 4: + product_functions.sh[10]rc=0 - 4: + product_functions.sh[11](( rc == 0 )) - 4: + product_functions.sh[11]mv pgb2file_f024_5_1p00.new pgb2file_f024_5_1p00 - 4: + product_functions.sh[12]return 0 - 4: + interp_atmos_master.sh[56]export err=0 - 4: + interp_atmos_master.sh[56]err=0 - 4: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 4: ++ interp_atmos_master.sh[62]wc -l - 4: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_5_1p00 -match 'LAND|ICEC' - 4: + interp_atmos_master.sh[62]var_count=0 - 4: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 4: + interp_atmos_master.sh[73]exit 0 -22: + bash[8]'[' -z '' ']' -22: + bash[9]case "$-" in -22: + bash[12]__lmod_vx=x -22: + bash[16]'[' -n x ']' -22: + bash[16]set +x -22: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -22: Shell debugging restarted -22: + bash[224]unset __lmod_vx -22: + interp_atmos_master.sh[7]input_file=tmpfile_f024_23 -22: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_23 -22: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -22: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -22: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -22: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -22: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -22: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -22: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -22: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -22: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -22: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -22: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -22: + interp_atmos_master.sh[31]IFS=: -22: + interp_atmos_master.sh[31]read -ra grids -22: + interp_atmos_master.sh[33]output_grids= -22: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -22: + interp_atmos_master.sh[35]gridopt=grid0p25 -22: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_23_0p25' -22: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -22: + interp_atmos_master.sh[35]gridopt=grid0p50 -22: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_23_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_23_0p50' -22: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -22: + interp_atmos_master.sh[35]gridopt=grid1p00 -22: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_23_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_23_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_23_1p00' -22: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_23 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_23_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_23_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_23_1p00 -22: 1:0:d=2021032312:UGRD:40 m above ground:24 hour fcst: -22: 2:20986:d=2021032312:VGRD:40 m above ground:24 hour fcst: -22: 3:42415:d=2021032312:UGRD:50 m above ground:24 hour fcst: -22: 4:63509:d=2021032312:VGRD:50 m above ground:24 hour fcst: -22: 5:85036:d=2021032312:TMP:80 m above ground:24 hour fcst: -22: 6:105815:d=2021032312:SPFH:80 m above ground:24 hour fcst: -22: 7:134462:d=2021032312:PRES:80 m above ground:24 hour fcst: -22: 8:160117:d=2021032312:UGRD:80 m above ground:24 hour fcst: -22: 9:181374:d=2021032312:VGRD:80 m above ground:24 hour fcst: -22: 10:203080:d=2021032312:TMP:100 m above ground:24 hour fcst: -22: 11:223848:d=2021032312:UGRD:100 m above ground:24 hour fcst: -22: 12:245175:d=2021032312:VGRD:100 m above ground:24 hour fcst: -22: 13:266937:d=2021032312:TMP:1829 m above mean sea level:24 hour fcst: -22: 14:288591:d=2021032312:UGRD:1829 m above mean sea level:24 hour fcst: -22: 15:310808:d=2021032312:VGRD:1829 m above mean sea level:24 hour fcst: -22: 16:333651:d=2021032312:TMP:2743 m above mean sea level:24 hour fcst: -22: 17:355121:d=2021032312:UGRD:2743 m above mean sea level:24 hour fcst: -22: 18:377795:d=2021032312:VGRD:2743 m above mean sea level:24 hour fcst: -22: 19:401116:d=2021032312:TMP:3658 m above mean sea level:24 hour fcst: -22: 20:422413:d=2021032312:UGRD:3658 m above mean sea level:24 hour fcst: -22: 21:445606:d=2021032312:VGRD:3658 m above mean sea level:24 hour fcst: -22: 22:469384:d=2021032312:HGT:0C isotherm:24 hour fcst: -22: 23:491109:d=2021032312:RH:0C isotherm:24 hour fcst: -22: 24:510388:d=2021032312:HGT:highest tropospheric freezing level:24 hour fcst: -22: 25:531748:d=2021032312:RH:highest tropospheric freezing level:24 hour fcst: -22: 26:550987:d=2021032312:TMP:30-0 mb above ground:24 hour fcst: -22: 27:571802:d=2021032312:RH:30-0 mb above ground:24 hour fcst: -22: 28:590018:d=2021032312:SPFH:30-0 mb above ground:24 hour fcst: -22: 29:618783:d=2021032312:UGRD:30-0 mb above ground:24 hour fcst: -22: 30:639903:d=2021032312:VGRD:30-0 mb above ground:24 hour fcst: -22: 31:661693:d=2021032312:4LFTX:surface:24 hour fcst: -22: + interp_atmos_master.sh[47]export err=0 -22: + interp_atmos_master.sh[47]err=0 -22: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -22: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -22: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_23_0p25 -22: + product_functions.sh[5]local filename=pgb2file_f024_23_0p25 -22: + product_functions.sh[6]wgrib2 pgb2file_f024_23_0p25 -not_if :RH: -grib pgb2file_f024_23_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_23_0p25.new -22: 1:0:d=2021032312:UGRD:40 m above ground:24 hour fcst: -22: 2:506690:d=2021032312:VGRD:40 m above ground:24 hour fcst: -22: 3:1030808:d=2021032312:UGRD:50 m above ground:24 hour fcst: -22: 4:1539925:d=2021032312:VGRD:50 m above ground:24 hour fcst: -22: 5:2067084:d=2021032312:TMP:80 m above ground:24 hour fcst: -22: 6:2550580:d=2021032312:SPFH:80 m above ground:24 hour fcst: -22: 7:3334186:d=2021032312:PRES:80 m above ground:24 hour fcst: -22: 8:3976467:d=2021032312:UGRD:80 m above ground:24 hour fcst: -22: 9:4490709:d=2021032312:VGRD:80 m above ground:24 hour fcst: -22: 10:5023901:d=2021032312:TMP:100 m above ground:24 hour fcst: -22: 11:5506638:d=2021032312:UGRD:100 m above ground:24 hour fcst: -22: 12:6023126:d=2021032312:VGRD:100 m above ground:24 hour fcst: -22: 13:6559012:d=2021032312:TMP:1829 m above mean sea level:24 hour fcst: -22: 14:7158402:d=2021032312:UGRD:1829 m above mean sea level:24 hour fcst: -22: 15:7784981:d=2021032312:VGRD:1829 m above mean sea level:24 hour fcst: -22: 16:8443396:d=2021032312:TMP:2743 m above mean sea level:24 hour fcst: -22: 17:9027026:d=2021032312:UGRD:2743 m above mean sea level:24 hour fcst: -22: 18:9655355:d=2021032312:VGRD:2743 m above mean sea level:24 hour fcst: -22: 19:10320665:d=2021032312:TMP:3658 m above mean sea level:24 hour fcst: -22: 20:10896157:d=2021032312:UGRD:3658 m above mean sea level:24 hour fcst: -22: 21:11540117:d=2021032312:VGRD:3658 m above mean sea level:24 hour fcst: -22: 22:12213695:d=2021032312:HGT:0C isotherm:24 hour fcst: -22: 23:12807175:d=2021032312:RH:0C isotherm:24 hour fcst: -22: 24:13241724:d=2021032312:HGT:highest tropospheric freezing level:24 hour fcst: -22: 25:13824214:d=2021032312:RH:highest tropospheric freezing level:24 hour fcst: -22: 26:14264624:d=2021032312:TMP:30-0 mb above ground:24 hour fcst: -22: 27:14748320:d=2021032312:RH:30-0 mb above ground:24 hour fcst: -22: 28:15144242:d=2021032312:SPFH:30-0 mb above ground:24 hour fcst: -22: 29:15929985:d=2021032312:UGRD:30-0 mb above ground:24 hour fcst: -22: 30:16445130:d=2021032312:VGRD:30-0 mb above ground:24 hour fcst: -22: 31:16980431:d=2021032312:4LFTX:surface:24 hour fcst: -22: + product_functions.sh[10]rc=0 -22: + product_functions.sh[11](( rc == 0 )) -22: + product_functions.sh[11]mv pgb2file_f024_23_0p25.new pgb2file_f024_23_0p25 -22: + product_functions.sh[12]return 0 -22: + interp_atmos_master.sh[56]export err=0 -22: + interp_atmos_master.sh[56]err=0 -22: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -22: ++ interp_atmos_master.sh[62]wc -l -22: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_23_0p25 -match 'LAND|ICEC' -22: + interp_atmos_master.sh[62]var_count=0 -22: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -22: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -22: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_23_0p50 -22: + product_functions.sh[5]local filename=pgb2file_f024_23_0p50 -22: + product_functions.sh[6]wgrib2 pgb2file_f024_23_0p50 -not_if :RH: -grib pgb2file_f024_23_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_23_0p50.new -22: 1:0:d=2021032312:UGRD:40 m above ground:24 hour fcst: -22: 2:172829:d=2021032312:VGRD:40 m above ground:24 hour fcst: -22: 3:352643:d=2021032312:UGRD:50 m above ground:24 hour fcst: -22: 4:528019:d=2021032312:VGRD:50 m above ground:24 hour fcst: -22: 5:708962:d=2021032312:TMP:80 m above ground:24 hour fcst: -22: 6:876749:d=2021032312:SPFH:80 m above ground:24 hour fcst: -22: 7:1146547:d=2021032312:PRES:80 m above ground:24 hour fcst: -22: 8:1367288:d=2021032312:UGRD:80 m above ground:24 hour fcst: -22: 9:1544382:d=2021032312:VGRD:80 m above ground:24 hour fcst: -22: 10:1727400:d=2021032312:TMP:100 m above ground:24 hour fcst: -22: 11:1895165:d=2021032312:UGRD:100 m above ground:24 hour fcst: -22: 12:2073117:d=2021032312:VGRD:100 m above ground:24 hour fcst: -22: 13:2256948:d=2021032312:TMP:1829 m above mean sea level:24 hour fcst: -22: 14:2449348:d=2021032312:UGRD:1829 m above mean sea level:24 hour fcst: -22: 15:2649752:d=2021032312:VGRD:1829 m above mean sea level:24 hour fcst: -22: 16:2859864:d=2021032312:TMP:2743 m above mean sea level:24 hour fcst: -22: 17:3046754:d=2021032312:UGRD:2743 m above mean sea level:24 hour fcst: -22: 18:3248331:d=2021032312:VGRD:2743 m above mean sea level:24 hour fcst: -22: 19:3460895:d=2021032312:TMP:3658 m above mean sea level:24 hour fcst: -22: 20:3644848:d=2021032312:UGRD:3658 m above mean sea level:24 hour fcst: -22: 21:3851436:d=2021032312:VGRD:3658 m above mean sea level:24 hour fcst: -22: 22:4066985:d=2021032312:HGT:0C isotherm:24 hour fcst: -22: 23:4267167:d=2021032312:RH:0C isotherm:24 hour fcst: -22: 24:4421539:d=2021032312:HGT:highest tropospheric freezing level:24 hour fcst: -22: 25:4621059:d=2021032312:RH:highest tropospheric freezing level:24 hour fcst: -22: 26:4773643:d=2021032312:TMP:30-0 mb above ground:24 hour fcst: -22: 27:4941821:d=2021032312:RH:30-0 mb above ground:24 hour fcst: -22: 28:5082964:d=2021032312:SPFH:30-0 mb above ground:24 hour fcst: -22: 29:5353506:d=2021032312:UGRD:30-0 mb above ground:24 hour fcst: -22: 30:5530987:d=2021032312:VGRD:30-0 mb above ground:24 hour fcst: -22: 31:5714768:d=2021032312:4LFTX:surface:24 hour fcst: -22: + product_functions.sh[10]rc=0 -22: + product_functions.sh[11](( rc == 0 )) -22: + product_functions.sh[11]mv pgb2file_f024_23_0p50.new pgb2file_f024_23_0p50 -22: + product_functions.sh[12]return 0 -22: + interp_atmos_master.sh[56]export err=0 -22: + interp_atmos_master.sh[56]err=0 -22: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -22: ++ interp_atmos_master.sh[62]wc -l -22: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_23_0p50 -match 'LAND|ICEC' -22: + interp_atmos_master.sh[62]var_count=0 -22: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -22: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -22: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_23_1p00 -22: + product_functions.sh[5]local filename=pgb2file_f024_23_1p00 -22: + product_functions.sh[6]wgrib2 pgb2file_f024_23_1p00 -not_if :RH: -grib pgb2file_f024_23_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_23_1p00.new -22: 1:0:d=2021032312:UGRD:40 m above ground:24 hour fcst: -22: 2:55700:d=2021032312:VGRD:40 m above ground:24 hour fcst: -22: 3:112860:d=2021032312:UGRD:50 m above ground:24 hour fcst: -22: 4:169124:d=2021032312:VGRD:50 m above ground:24 hour fcst: -22: 5:226656:d=2021032312:TMP:80 m above ground:24 hour fcst: -22: 6:281217:d=2021032312:SPFH:80 m above ground:24 hour fcst: -22: 7:361759:d=2021032312:PRES:80 m above ground:24 hour fcst: -22: 8:431084:d=2021032312:UGRD:80 m above ground:24 hour fcst: -22: 9:487842:d=2021032312:VGRD:80 m above ground:24 hour fcst: -22: 10:545997:d=2021032312:TMP:100 m above ground:24 hour fcst: -22: 11:600561:d=2021032312:UGRD:100 m above ground:24 hour fcst: -22: 12:657571:d=2021032312:VGRD:100 m above ground:24 hour fcst: -22: 13:715994:d=2021032312:TMP:1829 m above mean sea level:24 hour fcst: -22: 14:775408:d=2021032312:UGRD:1829 m above mean sea level:24 hour fcst: -22: 15:836902:d=2021032312:VGRD:1829 m above mean sea level:24 hour fcst: -22: 16:900935:d=2021032312:TMP:2743 m above mean sea level:24 hour fcst: -22: 17:959081:d=2021032312:UGRD:2743 m above mean sea level:24 hour fcst: -22: 18:1021333:d=2021032312:VGRD:2743 m above mean sea level:24 hour fcst: -22: 19:1086329:d=2021032312:TMP:3658 m above mean sea level:24 hour fcst: -22: 20:1143877:d=2021032312:UGRD:3658 m above mean sea level:24 hour fcst: -22: 21:1207554:d=2021032312:VGRD:3658 m above mean sea level:24 hour fcst: -22: 22:1273596:d=2021032312:HGT:0C isotherm:24 hour fcst: -22: 23:1334074:d=2021032312:RH:0C isotherm:24 hour fcst: -22: 24:1384575:d=2021032312:HGT:highest tropospheric freezing level:24 hour fcst: -22: 25:1444736:d=2021032312:RH:highest tropospheric freezing level:24 hour fcst: -22: 26:1494855:d=2021032312:TMP:30-0 mb above ground:24 hour fcst: -22: 27:1549643:d=2021032312:RH:30-0 mb above ground:24 hour fcst: -22: 28:1596511:d=2021032312:SPFH:30-0 mb above ground:24 hour fcst: -22: 29:1677300:d=2021032312:UGRD:30-0 mb above ground:24 hour fcst: -22: 30:1734177:d=2021032312:VGRD:30-0 mb above ground:24 hour fcst: -22: 31:1792634:d=2021032312:4LFTX:surface:24 hour fcst: -22: + product_functions.sh[10]rc=0 -22: + product_functions.sh[11](( rc == 0 )) -22: + product_functions.sh[11]mv pgb2file_f024_23_1p00.new pgb2file_f024_23_1p00 -22: + product_functions.sh[12]return 0 -22: + interp_atmos_master.sh[56]export err=0 -22: + interp_atmos_master.sh[56]err=0 -22: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -22: ++ interp_atmos_master.sh[62]wc -l -22: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_23_1p00 -match 'LAND|ICEC' -22: + interp_atmos_master.sh[62]var_count=0 -22: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -22: + interp_atmos_master.sh[73]exit 0 - 1: + bash[8]'[' -z '' ']' - 1: + bash[9]case "$-" in - 1: + bash[12]__lmod_vx=x - 1: + bash[16]'[' -n x ']' - 1: + bash[16]set +x - 1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 1: Shell debugging restarted - 1: + bash[224]unset __lmod_vx - 1: + interp_atmos_master.sh[7]input_file=tmpfile_f024_2 - 1: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_2 - 1: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 - 1: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 1: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 1: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 1: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 1: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 1: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 1: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 1: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 1: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 1: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 1: + interp_atmos_master.sh[31]IFS=: - 1: + interp_atmos_master.sh[31]read -ra grids - 1: + interp_atmos_master.sh[33]output_grids= - 1: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 1: + interp_atmos_master.sh[35]gridopt=grid0p25 - 1: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_2_0p25' - 1: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 1: + interp_atmos_master.sh[35]gridopt=grid0p50 - 1: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_2_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_2_0p50' - 1: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 1: + interp_atmos_master.sh[35]gridopt=grid1p00 - 1: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_2_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_2_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_2_1p00' - 1: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_2 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_2_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_2_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_2_1p00 - 1: 1:0:d=2021032312:ABSV:0.02 mb:24 hour fcst: - 1: 2:16961:d=2021032312:O3MR:0.02 mb:24 hour fcst: - 1: 3:32537:d=2021032312:HGT:0.04 mb:24 hour fcst: - 1: 4:56848:d=2021032312:TMP:0.04 mb:24 hour fcst: - 1: 5:74121:d=2021032312:RH:0.04 mb:24 hour fcst: - 1: 6:80108:d=2021032312:SPFH:0.04 mb:24 hour fcst: - 1: 7:105744:d=2021032312:VVEL:0.04 mb:24 hour fcst: - 1: 8:133206:d=2021032312:DZDT:0.04 mb:24 hour fcst: - 1: 9:158051:d=2021032312:UGRD:0.04 mb:24 hour fcst: - 1: 10:177984:d=2021032312:VGRD:0.04 mb:24 hour fcst: - 1: 11:197697:d=2021032312:ABSV:0.04 mb:24 hour fcst: - 1: 12:216080:d=2021032312:O3MR:0.04 mb:24 hour fcst: - 1: 13:239986:d=2021032312:HGT:0.07 mb:24 hour fcst: - 1: 14:264083:d=2021032312:TMP:0.07 mb:24 hour fcst: - 1: 15:280995:d=2021032312:RH:0.07 mb:24 hour fcst: - 1: 16:288824:d=2021032312:SPFH:0.07 mb:24 hour fcst: - 1: 17:312582:d=2021032312:VVEL:0.07 mb:24 hour fcst: - 1: 18:341505:d=2021032312:DZDT:0.07 mb:24 hour fcst: - 1: 19:366935:d=2021032312:UGRD:0.07 mb:24 hour fcst: - 1: 20:387291:d=2021032312:VGRD:0.07 mb:24 hour fcst: - 1: 21:407600:d=2021032312:ABSV:0.07 mb:24 hour fcst: - 1: 22:426510:d=2021032312:O3MR:0.07 mb:24 hour fcst: - 1: 23:446323:d=2021032312:HGT:0.1 mb:24 hour fcst: - 1: 24:467895:d=2021032312:TMP:0.1 mb:24 hour fcst: - 1: 25:484563:d=2021032312:RH:0.1 mb:24 hour fcst: - 1: 26:492550:d=2021032312:SPFH:0.1 mb:24 hour fcst: - 1: 27:513896:d=2021032312:VVEL:0.1 mb:24 hour fcst: - 1: 28:543743:d=2021032312:DZDT:0.1 mb:24 hour fcst: - 1: 29:569615:d=2021032312:UGRD:0.1 mb:24 hour fcst: - 1: 30:589608:d=2021032312:VGRD:0.1 mb:24 hour fcst: - 1: 31:609655:d=2021032312:ABSV:0.1 mb:24 hour fcst: - 1: + interp_atmos_master.sh[47]export err=0 - 1: + interp_atmos_master.sh[47]err=0 - 1: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 1: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 1: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_2_0p25 - 1: + product_functions.sh[5]local filename=pgb2file_f024_2_0p25 - 1: + product_functions.sh[6]wgrib2 pgb2file_f024_2_0p25 -not_if :RH: -grib pgb2file_f024_2_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_2_0p25.new - 1: 1:0:d=2021032312:ABSV:0.02 mb:24 hour fcst: - 1: 2:367168:d=2021032312:O3MR:0.02 mb:24 hour fcst: - 1: 3:694346:d=2021032312:HGT:0.04 mb:24 hour fcst: - 1: 4:1350618:d=2021032312:TMP:0.04 mb:24 hour fcst: - 1: 5:1745656:d=2021032312:RH:0.04 mb:24 hour fcst: - 1: 6:1892542:d=2021032312:SPFH:0.04 mb:24 hour fcst: - 1: 7:2614672:d=2021032312:VVEL:0.04 mb:24 hour fcst: - 1: 8:3379909:d=2021032312:DZDT:0.04 mb:24 hour fcst: - 1: 9:4043663:d=2021032312:UGRD:0.04 mb:24 hour fcst: - 1: 10:4533625:d=2021032312:VGRD:0.04 mb:24 hour fcst: - 1: 11:5011911:d=2021032312:ABSV:0.04 mb:24 hour fcst: - 1: 12:5427129:d=2021032312:O3MR:0.04 mb:24 hour fcst: - 1: 13:6062692:d=2021032312:HGT:0.07 mb:24 hour fcst: - 1: 14:6707160:d=2021032312:TMP:0.07 mb:24 hour fcst: - 1: 15:7090825:d=2021032312:RH:0.07 mb:24 hour fcst: - 1: 16:7284530:d=2021032312:SPFH:0.07 mb:24 hour fcst: - 1: 17:7881745:d=2021032312:VVEL:0.07 mb:24 hour fcst: - 1: 18:8699702:d=2021032312:DZDT:0.07 mb:24 hour fcst: - 1: 19:9399636:d=2021032312:UGRD:0.07 mb:24 hour fcst: - 1: 20:9898501:d=2021032312:VGRD:0.07 mb:24 hour fcst: - 1: 21:10401624:d=2021032312:ABSV:0.07 mb:24 hour fcst: - 1: 22:10834155:d=2021032312:O3MR:0.07 mb:24 hour fcst: - 1: 23:11313077:d=2021032312:HGT:0.1 mb:24 hour fcst: - 1: 24:11852893:d=2021032312:TMP:0.1 mb:24 hour fcst: - 1: 25:12226523:d=2021032312:RH:0.1 mb:24 hour fcst: - 1: 26:12430821:d=2021032312:SPFH:0.1 mb:24 hour fcst: - 1: 27:12941518:d=2021032312:VVEL:0.1 mb:24 hour fcst: - 1: 28:13775030:d=2021032312:DZDT:0.1 mb:24 hour fcst: - 1: 29:14489416:d=2021032312:UGRD:0.1 mb:24 hour fcst: - 1: 30:14976601:d=2021032312:VGRD:0.1 mb:24 hour fcst: - 1: 31:15460328:d=2021032312:ABSV:0.1 mb:24 hour fcst: - 1: + product_functions.sh[10]rc=0 - 1: + product_functions.sh[11](( rc == 0 )) - 1: + product_functions.sh[11]mv pgb2file_f024_2_0p25.new pgb2file_f024_2_0p25 - 1: + product_functions.sh[12]return 0 - 1: + interp_atmos_master.sh[56]export err=0 - 1: + interp_atmos_master.sh[56]err=0 - 1: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 1: ++ interp_atmos_master.sh[62]wc -l - 1: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_2_0p25 -match 'LAND|ICEC' - 1: + interp_atmos_master.sh[62]var_count=0 - 1: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 1: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 1: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_2_0p50 - 1: + product_functions.sh[5]local filename=pgb2file_f024_2_0p50 - 1: + product_functions.sh[6]wgrib2 pgb2file_f024_2_0p50 -not_if :RH: -grib pgb2file_f024_2_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_2_0p50.new - 1: 1:0:d=2021032312:ABSV:0.02 mb:24 hour fcst: - 1: 2:127104:d=2021032312:O3MR:0.02 mb:24 hour fcst: - 1: 3:237068:d=2021032312:HGT:0.04 mb:24 hour fcst: - 1: 4:455066:d=2021032312:TMP:0.04 mb:24 hour fcst: - 1: 5:590896:d=2021032312:RH:0.04 mb:24 hour fcst: - 1: 6:637269:d=2021032312:SPFH:0.04 mb:24 hour fcst: - 1: 7:876504:d=2021032312:VVEL:0.04 mb:24 hour fcst: - 1: 8:1140535:d=2021032312:DZDT:0.04 mb:24 hour fcst: - 1: 9:1361745:d=2021032312:UGRD:0.04 mb:24 hour fcst: - 1: 10:1528524:d=2021032312:VGRD:0.04 mb:24 hour fcst: - 1: 11:1691345:d=2021032312:ABSV:0.04 mb:24 hour fcst: - 1: 12:1834057:d=2021032312:O3MR:0.04 mb:24 hour fcst: - 1: 13:2046517:d=2021032312:HGT:0.07 mb:24 hour fcst: - 1: 14:2262841:d=2021032312:TMP:0.07 mb:24 hour fcst: - 1: 15:2393630:d=2021032312:RH:0.07 mb:24 hour fcst: - 1: 16:2452336:d=2021032312:SPFH:0.07 mb:24 hour fcst: - 1: 17:2656634:d=2021032312:VVEL:0.07 mb:24 hour fcst: - 1: 18:2938663:d=2021032312:DZDT:0.07 mb:24 hour fcst: - 1: 19:3169657:d=2021032312:UGRD:0.07 mb:24 hour fcst: - 1: 20:3340034:d=2021032312:VGRD:0.07 mb:24 hour fcst: - 1: 21:3509249:d=2021032312:ABSV:0.07 mb:24 hour fcst: - 1: 22:3659063:d=2021032312:O3MR:0.07 mb:24 hour fcst: - 1: 23:3821831:d=2021032312:HGT:0.1 mb:24 hour fcst: - 1: 24:4005449:d=2021032312:TMP:0.1 mb:24 hour fcst: - 1: 25:4133682:d=2021032312:RH:0.1 mb:24 hour fcst: - 1: 26:4194075:d=2021032312:SPFH:0.1 mb:24 hour fcst: - 1: 27:4368032:d=2021032312:VVEL:0.1 mb:24 hour fcst: - 1: 28:4659393:d=2021032312:DZDT:0.1 mb:24 hour fcst: - 1: 29:4894644:d=2021032312:UGRD:0.1 mb:24 hour fcst: - 1: 30:5060653:d=2021032312:VGRD:0.1 mb:24 hour fcst: - 1: 31:5225389:d=2021032312:ABSV:0.1 mb:24 hour fcst: - 1: + product_functions.sh[10]rc=0 - 1: + product_functions.sh[11](( rc == 0 )) - 1: + product_functions.sh[11]mv pgb2file_f024_2_0p50.new pgb2file_f024_2_0p50 - 1: + product_functions.sh[12]return 0 - 1: + interp_atmos_master.sh[56]export err=0 - 1: + interp_atmos_master.sh[56]err=0 - 1: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 1: ++ interp_atmos_master.sh[62]wc -l - 1: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_2_0p50 -match 'LAND|ICEC' - 1: + interp_atmos_master.sh[62]var_count=0 - 1: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 1: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 1: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_2_1p00 - 1: + product_functions.sh[5]local filename=pgb2file_f024_2_1p00 - 1: + product_functions.sh[6]wgrib2 pgb2file_f024_2_1p00 -not_if :RH: -grib pgb2file_f024_2_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_2_1p00.new - 1: 1:0:d=2021032312:ABSV:0.02 mb:24 hour fcst: - 1: 2:42744:d=2021032312:O3MR:0.02 mb:24 hour fcst: - 1: 3:80247:d=2021032312:HGT:0.04 mb:24 hour fcst: - 1: 4:147195:d=2021032312:TMP:0.04 mb:24 hour fcst: - 1: 5:191598:d=2021032312:RH:0.04 mb:24 hour fcst: - 1: 6:206372:d=2021032312:SPFH:0.04 mb:24 hour fcst: - 1: 7:278813:d=2021032312:VVEL:0.04 mb:24 hour fcst: - 1: 8:357527:d=2021032312:DZDT:0.04 mb:24 hour fcst: - 1: 9:425321:d=2021032312:UGRD:0.04 mb:24 hour fcst: - 1: 10:478649:d=2021032312:VGRD:0.04 mb:24 hour fcst: - 1: 11:531152:d=2021032312:ABSV:0.04 mb:24 hour fcst: - 1: 12:578284:d=2021032312:O3MR:0.04 mb:24 hour fcst: - 1: 13:643276:d=2021032312:HGT:0.07 mb:24 hour fcst: - 1: 14:709685:d=2021032312:TMP:0.07 mb:24 hour fcst: - 1: 15:752772:d=2021032312:RH:0.07 mb:24 hour fcst: - 1: 16:771896:d=2021032312:SPFH:0.07 mb:24 hour fcst: - 1: 17:836087:d=2021032312:VVEL:0.07 mb:24 hour fcst: - 1: 18:919240:d=2021032312:DZDT:0.07 mb:24 hour fcst: - 1: 19:989160:d=2021032312:UGRD:0.07 mb:24 hour fcst: - 1: 20:1043564:d=2021032312:VGRD:0.07 mb:24 hour fcst: - 1: 21:1097882:d=2021032312:ABSV:0.07 mb:24 hour fcst: - 1: 22:1147030:d=2021032312:O3MR:0.07 mb:24 hour fcst: - 1: 23:1198862:d=2021032312:HGT:0.1 mb:24 hour fcst: - 1: 24:1256663:d=2021032312:TMP:0.1 mb:24 hour fcst: - 1: 25:1298845:d=2021032312:RH:0.1 mb:24 hour fcst: - 1: 26:1318264:d=2021032312:SPFH:0.1 mb:24 hour fcst: - 1: 27:1374821:d=2021032312:VVEL:0.1 mb:24 hour fcst: - 1: 28:1460548:d=2021032312:DZDT:0.1 mb:24 hour fcst: - 1: 29:1531879:d=2021032312:UGRD:0.1 mb:24 hour fcst: - 1: 30:1585119:d=2021032312:VGRD:0.1 mb:24 hour fcst: - 1: 31:1638415:d=2021032312:ABSV:0.1 mb:24 hour fcst: - 1: + product_functions.sh[10]rc=0 - 1: + product_functions.sh[11](( rc == 0 )) - 1: + product_functions.sh[11]mv pgb2file_f024_2_1p00.new pgb2file_f024_2_1p00 - 1: + product_functions.sh[12]return 0 - 1: + interp_atmos_master.sh[56]export err=0 - 1: + interp_atmos_master.sh[56]err=0 - 1: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 1: ++ interp_atmos_master.sh[62]wc -l - 1: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_2_1p00 -match 'LAND|ICEC' - 1: + interp_atmos_master.sh[62]var_count=0 - 1: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 1: + interp_atmos_master.sh[73]exit 0 - 7: + bash[8]'[' -z '' ']' - 7: + bash[9]case "$-" in - 7: + bash[12]__lmod_vx=x - 7: + bash[16]'[' -n x ']' - 7: + bash[16]set +x - 7: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 7: Shell debugging restarted - 7: + bash[224]unset __lmod_vx - 7: + interp_atmos_master.sh[7]input_file=tmpfile_f024_8 - 7: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_8 - 7: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 - 7: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 7: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 7: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 7: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 7: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 7: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 7: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 7: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 7: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 7: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 7: + interp_atmos_master.sh[31]IFS=: - 7: + interp_atmos_master.sh[31]read -ra grids - 7: + interp_atmos_master.sh[33]output_grids= - 7: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 7: + interp_atmos_master.sh[35]gridopt=grid0p25 - 7: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_8_0p25' - 7: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 7: + interp_atmos_master.sh[35]gridopt=grid0p50 - 7: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_8_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_8_0p50' - 7: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 7: + interp_atmos_master.sh[35]gridopt=grid1p00 - 7: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_8_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_8_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_8_1p00' - 7: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_8 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_8_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_8_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_8_1p00 - 7: 1:0:d=2021032312:ABSV:70 mb:24 hour fcst: - 7: 2:17809:d=2021032312:O3MR:70 mb:24 hour fcst: - 7: 3:43613:d=2021032312:HGT:100 mb:24 hour fcst: - 7: 4:63262:d=2021032312:TMP:100 mb:24 hour fcst: - 7: 5:80180:d=2021032312:RH:100 mb:24 hour fcst: - 7: 6:90329:d=2021032312:TCDC:100 mb:24 hour fcst: - 7: 7:90508:d=2021032312:SPFH:100 mb:24 hour fcst: - 7: 8:116342:d=2021032312:VVEL:100 mb:24 hour fcst: - 7: 9:141399:d=2021032312:DZDT:100 mb:24 hour fcst: - 7: 10:166011:d=2021032312:UGRD:100 mb:24 hour fcst: - 7: 11:185377:d=2021032312:VGRD:100 mb:24 hour fcst: - 7: 12:204809:d=2021032312:ABSV:100 mb:24 hour fcst: - 7: 13:222998:d=2021032312:CLMR:100 mb:24 hour fcst: - 7: 14:223177:d=2021032312:ICMR:100 mb:24 hour fcst: - 7: 15:223983:d=2021032312:RWMR:100 mb:24 hour fcst: - 7: 16:224162:d=2021032312:SNMR:100 mb:24 hour fcst: - 7: 17:226129:d=2021032312:GRLE:100 mb:24 hour fcst: - 7: 18:226308:d=2021032312:O3MR:100 mb:24 hour fcst: - 7: 19:250371:d=2021032312:HGT:150 mb:24 hour fcst: - 7: 20:270771:d=2021032312:TMP:150 mb:24 hour fcst: - 7: 21:287740:d=2021032312:RH:150 mb:24 hour fcst: - 7: 22:299199:d=2021032312:TCDC:150 mb:24 hour fcst: - 7: 23:303230:d=2021032312:SPFH:150 mb:24 hour fcst: - 7: 24:327607:d=2021032312:VVEL:150 mb:24 hour fcst: - 7: 25:354919:d=2021032312:DZDT:150 mb:24 hour fcst: - 7: 26:380412:d=2021032312:UGRD:150 mb:24 hour fcst: - 7: 27:400385:d=2021032312:VGRD:150 mb:24 hour fcst: - 7: 28:420411:d=2021032312:ABSV:150 mb:24 hour fcst: - 7: 29:439289:d=2021032312:CLMR:150 mb:24 hour fcst: - 7: 30:439468:d=2021032312:ICMR:150 mb:24 hour fcst: - 7: 31:444740:d=2021032312:RWMR:150 mb:24 hour fcst: - 7: + interp_atmos_master.sh[47]export err=0 - 7: + interp_atmos_master.sh[47]err=0 - 7: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 7: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 7: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_8_0p25 - 7: + product_functions.sh[5]local filename=pgb2file_f024_8_0p25 - 7: + product_functions.sh[6]wgrib2 pgb2file_f024_8_0p25 -not_if :RH: -grib pgb2file_f024_8_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_8_0p25.new - 7: 1:0:d=2021032312:ABSV:70 mb:24 hour fcst: - 7: 2:399741:d=2021032312:O3MR:70 mb:24 hour fcst: - 7: 3:1112157:d=2021032312:HGT:100 mb:24 hour fcst: - 7: 4:1580520:d=2021032312:TMP:100 mb:24 hour fcst: - 7: 5:1958210:d=2021032312:RH:100 mb:24 hour fcst: - 7: 6:2191653:d=2021032312:TCDC:100 mb:24 hour fcst: - 7: 7:2191832:d=2021032312:SPFH:100 mb:24 hour fcst: - 7: 8:2852166:d=2021032312:VVEL:100 mb:24 hour fcst: - 7: 9:3524013:d=2021032312:DZDT:100 mb:24 hour fcst: - 7: 10:4178860:d=2021032312:UGRD:100 mb:24 hour fcst: - 7: 11:4645728:d=2021032312:VGRD:100 mb:24 hour fcst: - 7: 12:5111004:d=2021032312:ABSV:100 mb:24 hour fcst: - 7: 13:5518342:d=2021032312:CLMR:100 mb:24 hour fcst: - 7: 14:5518521:d=2021032312:ICMR:100 mb:24 hour fcst: - 7: 15:5530235:d=2021032312:RWMR:100 mb:24 hour fcst: - 7: 16:5530414:d=2021032312:SNMR:100 mb:24 hour fcst: - 7: 17:5571264:d=2021032312:GRLE:100 mb:24 hour fcst: - 7: 18:5571443:d=2021032312:O3MR:100 mb:24 hour fcst: - 7: 19:6189843:d=2021032312:HGT:150 mb:24 hour fcst: - 7: 20:6685689:d=2021032312:TMP:150 mb:24 hour fcst: - 7: 21:7067463:d=2021032312:RH:150 mb:24 hour fcst: - 7: 22:7333087:d=2021032312:TCDC:150 mb:24 hour fcst: - 7: 23:7416088:d=2021032312:SPFH:150 mb:24 hour fcst: - 7: 24:8048467:d=2021032312:VVEL:150 mb:24 hour fcst: - 7: 25:8804350:d=2021032312:DZDT:150 mb:24 hour fcst: - 7: 26:9504816:d=2021032312:UGRD:150 mb:24 hour fcst: - 7: 27:9998249:d=2021032312:VGRD:150 mb:24 hour fcst: - 7: 28:10492327:d=2021032312:ABSV:150 mb:24 hour fcst: - 7: 29:10923095:d=2021032312:CLMR:150 mb:24 hour fcst: - 7: 30:10923274:d=2021032312:ICMR:150 mb:24 hour fcst: - 7: 31:11045779:d=2021032312:RWMR:150 mb:24 hour fcst: - 7: + product_functions.sh[10]rc=0 - 7: + product_functions.sh[11](( rc == 0 )) - 7: + product_functions.sh[11]mv pgb2file_f024_8_0p25.new pgb2file_f024_8_0p25 - 7: + product_functions.sh[12]return 0 - 7: + interp_atmos_master.sh[56]export err=0 - 7: + interp_atmos_master.sh[56]err=0 - 7: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 7: ++ interp_atmos_master.sh[62]wc -l - 7: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_8_0p25 -match 'LAND|ICEC' - 7: + interp_atmos_master.sh[62]var_count=0 - 7: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 7: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 7: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_8_0p50 - 7: + product_functions.sh[5]local filename=pgb2file_f024_8_0p50 - 7: + product_functions.sh[6]wgrib2 pgb2file_f024_8_0p50 -not_if :RH: -grib pgb2file_f024_8_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_8_0p50.new - 7: 1:0:d=2021032312:ABSV:70 mb:24 hour fcst: - 7: 2:136320:d=2021032312:O3MR:70 mb:24 hour fcst: - 7: 3:374664:d=2021032312:HGT:100 mb:24 hour fcst: - 7: 4:532641:d=2021032312:TMP:100 mb:24 hour fcst: - 7: 5:662413:d=2021032312:RH:100 mb:24 hour fcst: - 7: 6:739837:d=2021032312:TCDC:100 mb:24 hour fcst: - 7: 7:740016:d=2021032312:SPFH:100 mb:24 hour fcst: - 7: 8:971247:d=2021032312:VVEL:100 mb:24 hour fcst: - 7: 9:1199578:d=2021032312:DZDT:100 mb:24 hour fcst: - 7: 10:1416053:d=2021032312:UGRD:100 mb:24 hour fcst: - 7: 11:1574824:d=2021032312:VGRD:100 mb:24 hour fcst: - 7: 12:1734434:d=2021032312:ABSV:100 mb:24 hour fcst: - 7: 13:1876518:d=2021032312:CLMR:100 mb:24 hour fcst: - 7: 14:1876697:d=2021032312:ICMR:100 mb:24 hour fcst: - 7: 15:1881250:d=2021032312:RWMR:100 mb:24 hour fcst: - 7: 16:1881429:d=2021032312:SNMR:100 mb:24 hour fcst: - 7: 17:1896660:d=2021032312:GRLE:100 mb:24 hour fcst: - 7: 18:1896839:d=2021032312:O3MR:100 mb:24 hour fcst: - 7: 19:2107344:d=2021032312:HGT:150 mb:24 hour fcst: - 7: 20:2274277:d=2021032312:TMP:150 mb:24 hour fcst: - 7: 21:2403810:d=2021032312:RH:150 mb:24 hour fcst: - 7: 22:2491637:d=2021032312:TCDC:150 mb:24 hour fcst: - 7: 23:2521653:d=2021032312:SPFH:150 mb:24 hour fcst: - 7: 24:2738186:d=2021032312:VVEL:150 mb:24 hour fcst: - 7: 25:2994515:d=2021032312:DZDT:150 mb:24 hour fcst: - 7: 26:3223645:d=2021032312:UGRD:150 mb:24 hour fcst: - 7: 27:3390279:d=2021032312:VGRD:150 mb:24 hour fcst: - 7: 28:3557458:d=2021032312:ABSV:150 mb:24 hour fcst: - 7: 29:3705818:d=2021032312:CLMR:150 mb:24 hour fcst: - 7: 30:3705997:d=2021032312:ICMR:150 mb:24 hour fcst: - 7: 31:3748294:d=2021032312:RWMR:150 mb:24 hour fcst: - 7: + product_functions.sh[10]rc=0 - 7: + product_functions.sh[11](( rc == 0 )) - 7: + product_functions.sh[11]mv pgb2file_f024_8_0p50.new pgb2file_f024_8_0p50 - 7: + product_functions.sh[12]return 0 - 7: + interp_atmos_master.sh[56]export err=0 - 7: + interp_atmos_master.sh[56]err=0 - 7: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 7: ++ interp_atmos_master.sh[62]wc -l - 7: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_8_0p50 -match 'LAND|ICEC' - 7: + interp_atmos_master.sh[62]var_count=0 - 7: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 7: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 7: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_8_1p00 - 7: + product_functions.sh[5]local filename=pgb2file_f024_8_1p00 - 7: + product_functions.sh[6]wgrib2 pgb2file_f024_8_1p00 -not_if :RH: -grib pgb2file_f024_8_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_8_1p00.new - 7: 1:0:d=2021032312:ABSV:70 mb:24 hour fcst: - 7: 2:45184:d=2021032312:O3MR:70 mb:24 hour fcst: - 7: 3:117436:d=2021032312:HGT:100 mb:24 hour fcst: - 7: 4:168499:d=2021032312:TMP:100 mb:24 hour fcst: - 7: 5:211321:d=2021032312:RH:100 mb:24 hour fcst: - 7: 6:236668:d=2021032312:TCDC:100 mb:24 hour fcst: - 7: 7:236847:d=2021032312:SPFH:100 mb:24 hour fcst: - 7: 8:307110:d=2021032312:VVEL:100 mb:24 hour fcst: - 7: 9:376465:d=2021032312:DZDT:100 mb:24 hour fcst: - 7: 10:442589:d=2021032312:UGRD:100 mb:24 hour fcst: - 7: 11:493578:d=2021032312:VGRD:100 mb:24 hour fcst: - 7: 12:545004:d=2021032312:ABSV:100 mb:24 hour fcst: - 7: 13:591847:d=2021032312:CLMR:100 mb:24 hour fcst: - 7: 14:592026:d=2021032312:ICMR:100 mb:24 hour fcst: - 7: 15:593856:d=2021032312:RWMR:100 mb:24 hour fcst: - 7: 16:594035:d=2021032312:SNMR:100 mb:24 hour fcst: - 7: 17:599237:d=2021032312:GRLE:100 mb:24 hour fcst: - 7: 18:599416:d=2021032312:O3MR:100 mb:24 hour fcst: - 7: 19:664958:d=2021032312:HGT:150 mb:24 hour fcst: - 7: 20:718652:d=2021032312:TMP:150 mb:24 hour fcst: - 7: 21:761651:d=2021032312:RH:150 mb:24 hour fcst: - 7: 22:790214:d=2021032312:TCDC:150 mb:24 hour fcst: - 7: 23:800568:d=2021032312:SPFH:150 mb:24 hour fcst: - 7: 24:866412:d=2021032312:VVEL:150 mb:24 hour fcst: - 7: 25:943125:d=2021032312:DZDT:150 mb:24 hour fcst: - 7: 26:1012611:d=2021032312:UGRD:150 mb:24 hour fcst: - 7: 27:1065729:d=2021032312:VGRD:150 mb:24 hour fcst: - 7: 28:1119209:d=2021032312:ABSV:150 mb:24 hour fcst: - 7: 29:1168055:d=2021032312:CLMR:150 mb:24 hour fcst: - 7: 30:1168234:d=2021032312:ICMR:150 mb:24 hour fcst: - 7: 31:1182052:d=2021032312:RWMR:150 mb:24 hour fcst: - 7: + product_functions.sh[10]rc=0 - 7: + product_functions.sh[11](( rc == 0 )) - 7: + product_functions.sh[11]mv pgb2file_f024_8_1p00.new pgb2file_f024_8_1p00 - 7: + product_functions.sh[12]return 0 - 7: + interp_atmos_master.sh[56]export err=0 - 7: + interp_atmos_master.sh[56]err=0 - 7: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 7: ++ interp_atmos_master.sh[62]wc -l - 7: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_8_1p00 -match 'LAND|ICEC' - 7: + interp_atmos_master.sh[62]var_count=0 - 7: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 7: + interp_atmos_master.sh[73]exit 0 -21: + bash[8]'[' -z '' ']' -21: + bash[9]case "$-" in -21: + bash[12]__lmod_vx=x -21: + bash[16]'[' -n x ']' -21: + bash[16]set +x -21: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -21: Shell debugging restarted -21: + bash[224]unset __lmod_vx -21: + interp_atmos_master.sh[7]input_file=tmpfile_f024_22 -21: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_22 -21: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -21: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -21: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -21: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -21: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -21: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -21: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -21: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -21: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -21: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -21: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -21: + interp_atmos_master.sh[31]IFS=: -21: + interp_atmos_master.sh[31]read -ra grids -21: + interp_atmos_master.sh[33]output_grids= -21: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -21: + interp_atmos_master.sh[35]gridopt=grid0p25 -21: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_22_0p25' -21: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -21: + interp_atmos_master.sh[35]gridopt=grid0p50 -21: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_22_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_22_0p50' -21: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -21: + interp_atmos_master.sh[35]gridopt=grid1p00 -21: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_22_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_22_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_22_1p00' -21: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_22 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_22_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_22_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_22_1p00 -21: 1:0:d=2021032312:DLWRF:surface:18-24 hour ave fcst: -21: 2:21270:d=2021032312:USWRF:surface:18-24 hour ave fcst: -21: 3:41054:d=2021032312:ULWRF:surface:18-24 hour ave fcst: -21: 4:59808:d=2021032312:USWRF:top of atmosphere:18-24 hour ave fcst: -21: 5:82397:d=2021032312:ULWRF:top of atmosphere:18-24 hour ave fcst: -21: 6:103160:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Total Aerosol:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 7:130702:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Dust Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 8:148843:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Sea Salt Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 9:176262:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Sulphate Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 10:201923:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Particulate Organic Matter Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 11:227492:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Black Carbon Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 12:249771:d=2021032312:HLCY:3000-0 m above ground:24 hour fcst: -21: 13:265663:d=2021032312:HLCY:1000-0 m above ground:24 hour fcst: -21: 14:287943:d=2021032312:USTM:6000-0 m above ground:24 hour fcst: -21: 15:311169:d=2021032312:VSTM:6000-0 m above ground:24 hour fcst: -21: 16:333988:d=2021032312:PRES:tropopause:24 hour fcst: -21: 17:363780:d=2021032312:ICAHT:tropopause:24 hour fcst: -21: 18:394556:d=2021032312:HGT:tropopause:24 hour fcst: -21: 19:422751:d=2021032312:TMP:tropopause:24 hour fcst: -21: 20:444317:d=2021032312:UGRD:tropopause:24 hour fcst: -21: 21:460304:d=2021032312:VGRD:tropopause:24 hour fcst: -21: 22:476148:d=2021032312:VWSH:tropopause:24 hour fcst: -21: 23:494822:d=2021032312:PRES:max wind:24 hour fcst: -21: 24:529108:d=2021032312:ICAHT:max wind:24 hour fcst: -21: 25:564132:d=2021032312:HGT:max wind:24 hour fcst: -21: 26:599329:d=2021032312:UGRD:max wind:24 hour fcst: -21: 27:615815:d=2021032312:VGRD:max wind:24 hour fcst: -21: 28:633158:d=2021032312:TMP:max wind:24 hour fcst: -21: 29:659893:d=2021032312:UGRD:20 m above ground:24 hour fcst: -21: 30:680417:d=2021032312:VGRD:20 m above ground:24 hour fcst: -21: 31:701515:d=2021032312:UGRD:30 m above ground:24 hour fcst: -21: 32:722401:d=2021032312:VGRD:30 m above ground:24 hour fcst: -21: + interp_atmos_master.sh[47]export err=0 -21: + interp_atmos_master.sh[47]err=0 -21: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -21: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -21: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_22_0p25 -21: + product_functions.sh[5]local filename=pgb2file_f024_22_0p25 -21: + product_functions.sh[6]wgrib2 pgb2file_f024_22_0p25 -not_if :RH: -grib pgb2file_f024_22_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_22_0p25.new -21: 1:0:d=2021032312:DLWRF:surface:18-24 hour ave fcst: -21: 2:517034:d=2021032312:USWRF:surface:18-24 hour ave fcst: -21: 3:1006312:d=2021032312:ULWRF:surface:18-24 hour ave fcst: -21: 4:1432788:d=2021032312:USWRF:top of atmosphere:18-24 hour ave fcst: -21: 5:2013152:d=2021032312:ULWRF:top of atmosphere:18-24 hour ave fcst: -21: 6:2503202:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Total Aerosol:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 7:3248002:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Dust Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 8:3694813:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Sea Salt Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 9:4431805:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Sulphate Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 10:5109744:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Particulate Organic Matter Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 11:5768905:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Black Carbon Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 12:6321996:d=2021032312:HLCY:3000-0 m above ground:24 hour fcst: -21: 13:6659759:d=2021032312:HLCY:1000-0 m above ground:24 hour fcst: -21: 14:7197967:d=2021032312:USTM:6000-0 m above ground:24 hour fcst: -21: 15:7769525:d=2021032312:VSTM:6000-0 m above ground:24 hour fcst: -21: 16:8334792:d=2021032312:PRES:tropopause:24 hour fcst: -21: 17:9134986:d=2021032312:ICAHT:tropopause:24 hour fcst: -21: 18:9986304:d=2021032312:HGT:tropopause:24 hour fcst: -21: 19:10734708:d=2021032312:TMP:tropopause:24 hour fcst: -21: 20:11250447:d=2021032312:UGRD:tropopause:24 hour fcst: -21: 21:11592420:d=2021032312:VGRD:tropopause:24 hour fcst: -21: 22:11933008:d=2021032312:VWSH:tropopause:24 hour fcst: -21: 23:12346772:d=2021032312:PRES:max wind:24 hour fcst: -21: 24:13236319:d=2021032312:ICAHT:max wind:24 hour fcst: -21: 25:14139293:d=2021032312:HGT:max wind:24 hour fcst: -21: 26:15045177:d=2021032312:UGRD:max wind:24 hour fcst: -21: 27:15401662:d=2021032312:VGRD:max wind:24 hour fcst: -21: 28:15779892:d=2021032312:TMP:max wind:24 hour fcst: -21: 29:16444397:d=2021032312:UGRD:20 m above ground:24 hour fcst: -21: 30:16940220:d=2021032312:VGRD:20 m above ground:24 hour fcst: -21: 31:17453692:d=2021032312:UGRD:30 m above ground:24 hour fcst: -21: 32:17957179:d=2021032312:VGRD:30 m above ground:24 hour fcst: -21: + product_functions.sh[10]rc=0 -21: + product_functions.sh[11](( rc == 0 )) -21: + product_functions.sh[11]mv pgb2file_f024_22_0p25.new pgb2file_f024_22_0p25 -21: + product_functions.sh[12]return 0 -21: + interp_atmos_master.sh[56]export err=0 -21: + interp_atmos_master.sh[56]err=0 -21: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -21: ++ interp_atmos_master.sh[62]wc -l -21: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_22_0p25 -match 'LAND|ICEC' -21: + interp_atmos_master.sh[62]var_count=0 -21: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -21: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -21: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_22_0p50 -21: + product_functions.sh[5]local filename=pgb2file_f024_22_0p50 -21: + product_functions.sh[6]wgrib2 pgb2file_f024_22_0p50 -not_if :RH: -grib pgb2file_f024_22_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_22_0p50.new -21: 1:0:d=2021032312:DLWRF:surface:18-24 hour ave fcst: -21: 2:178193:d=2021032312:USWRF:surface:18-24 hour ave fcst: -21: 3:346173:d=2021032312:ULWRF:surface:18-24 hour ave fcst: -21: 4:490504:d=2021032312:USWRF:top of atmosphere:18-24 hour ave fcst: -21: 5:693153:d=2021032312:ULWRF:top of atmosphere:18-24 hour ave fcst: -21: 6:864195:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Total Aerosol:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 7:1114564:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Dust Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 8:1261519:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Sea Salt Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 9:1515950:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Sulphate Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 10:1743447:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Particulate Organic Matter Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 11:1967186:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Black Carbon Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 12:2154050:d=2021032312:HLCY:3000-0 m above ground:24 hour fcst: -21: 13:2268221:d=2021032312:HLCY:1000-0 m above ground:24 hour fcst: -21: 14:2452166:d=2021032312:USTM:6000-0 m above ground:24 hour fcst: -21: 15:2649099:d=2021032312:VSTM:6000-0 m above ground:24 hour fcst: -21: 16:2842070:d=2021032312:PRES:tropopause:24 hour fcst: -21: 17:3113536:d=2021032312:ICAHT:tropopause:24 hour fcst: -21: 18:3400600:d=2021032312:HGT:tropopause:24 hour fcst: -21: 19:3654040:d=2021032312:TMP:tropopause:24 hour fcst: -21: 20:3830215:d=2021032312:UGRD:tropopause:24 hour fcst: -21: 21:3946732:d=2021032312:VGRD:tropopause:24 hour fcst: -21: 22:4062343:d=2021032312:VWSH:tropopause:24 hour fcst: -21: 23:4206727:d=2021032312:PRES:max wind:24 hour fcst: -21: 24:4535364:d=2021032312:ICAHT:max wind:24 hour fcst: -21: 25:4874315:d=2021032312:HGT:max wind:24 hour fcst: -21: 26:5214039:d=2021032312:UGRD:max wind:24 hour fcst: -21: 27:5335261:d=2021032312:VGRD:max wind:24 hour fcst: -21: 28:5465247:d=2021032312:TMP:max wind:24 hour fcst: -21: 29:5699732:d=2021032312:UGRD:20 m above ground:24 hour fcst: -21: 30:5868858:d=2021032312:VGRD:20 m above ground:24 hour fcst: -21: 31:6044913:d=2021032312:UGRD:30 m above ground:24 hour fcst: -21: 32:6216656:d=2021032312:VGRD:30 m above ground:24 hour fcst: -21: + product_functions.sh[10]rc=0 -21: + product_functions.sh[11](( rc == 0 )) -21: + product_functions.sh[11]mv pgb2file_f024_22_0p50.new pgb2file_f024_22_0p50 -21: + product_functions.sh[12]return 0 -21: + interp_atmos_master.sh[56]export err=0 -21: + interp_atmos_master.sh[56]err=0 -21: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -21: ++ interp_atmos_master.sh[62]wc -l -21: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_22_0p50 -match 'LAND|ICEC' -21: + interp_atmos_master.sh[62]var_count=0 -21: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -21: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -21: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_22_1p00 -21: + product_functions.sh[5]local filename=pgb2file_f024_22_1p00 -21: + product_functions.sh[6]wgrib2 pgb2file_f024_22_1p00 -not_if :RH: -grib pgb2file_f024_22_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_22_1p00.new -21: 1:0:d=2021032312:DLWRF:surface:18-24 hour ave fcst: -21: 2:56856:d=2021032312:USWRF:surface:18-24 hour ave fcst: -21: 3:109293:d=2021032312:ULWRF:surface:18-24 hour ave fcst: -21: 4:157014:d=2021032312:USWRF:top of atmosphere:18-24 hour ave fcst: -21: 5:218861:d=2021032312:ULWRF:top of atmosphere:18-24 hour ave fcst: -21: 6:273843:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Total Aerosol:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 7:349892:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Dust Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 8:396687:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Sea Salt Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 9:473641:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Sulphate Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 10:543928:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Particulate Organic Matter Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 11:613518:d=2021032312:AOTK:entire atmosphere:24 hour fcst:aerosol=Black Carbon Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 12:672955:d=2021032312:HLCY:3000-0 m above ground:24 hour fcst: -21: 13:711666:d=2021032312:HLCY:1000-0 m above ground:24 hour fcst: -21: 14:770741:d=2021032312:USTM:6000-0 m above ground:24 hour fcst: -21: 15:833756:d=2021032312:VSTM:6000-0 m above ground:24 hour fcst: -21: 16:895868:d=2021032312:PRES:tropopause:24 hour fcst: -21: 17:979021:d=2021032312:ICAHT:tropopause:24 hour fcst: -21: 18:1066036:d=2021032312:HGT:tropopause:24 hour fcst: -21: 19:1144067:d=2021032312:TMP:tropopause:24 hour fcst: -21: 20:1200950:d=2021032312:UGRD:tropopause:24 hour fcst: -21: 21:1240476:d=2021032312:VGRD:tropopause:24 hour fcst: -21: 22:1279552:d=2021032312:VWSH:tropopause:24 hour fcst: -21: 23:1326738:d=2021032312:PRES:max wind:24 hour fcst: -21: 24:1426144:d=2021032312:ICAHT:max wind:24 hour fcst: -21: 25:1527634:d=2021032312:HGT:max wind:24 hour fcst: -21: 26:1629532:d=2021032312:UGRD:max wind:24 hour fcst: -21: 27:1670774:d=2021032312:VGRD:max wind:24 hour fcst: -21: 28:1714797:d=2021032312:TMP:max wind:24 hour fcst: -21: 29:1788641:d=2021032312:UGRD:20 m above ground:24 hour fcst: -21: 30:1843208:d=2021032312:VGRD:20 m above ground:24 hour fcst: -21: 31:1899403:d=2021032312:UGRD:30 m above ground:24 hour fcst: -21: 32:1954718:d=2021032312:VGRD:30 m above ground:24 hour fcst: -21: + product_functions.sh[10]rc=0 -21: + product_functions.sh[11](( rc == 0 )) -21: + product_functions.sh[11]mv pgb2file_f024_22_1p00.new pgb2file_f024_22_1p00 -21: + product_functions.sh[12]return 0 -21: + interp_atmos_master.sh[56]export err=0 -21: + interp_atmos_master.sh[56]err=0 -21: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -21: ++ interp_atmos_master.sh[62]wc -l -21: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_22_1p00 -match 'LAND|ICEC' -21: + interp_atmos_master.sh[62]var_count=0 -21: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -21: + interp_atmos_master.sh[73]exit 0 -12: + bash[8]'[' -z '' ']' -12: + bash[9]case "$-" in -12: + bash[12]__lmod_vx=x -12: + bash[16]'[' -n x ']' -12: + bash[16]set +x -12: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -12: Shell debugging restarted -12: + bash[224]unset __lmod_vx -12: + interp_atmos_master.sh[7]input_file=tmpfile_f024_13 -12: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_13 -12: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -12: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -12: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -12: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -12: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -12: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -12: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -12: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -12: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -12: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -12: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -12: + interp_atmos_master.sh[31]IFS=: -12: + interp_atmos_master.sh[31]read -ra grids -12: + interp_atmos_master.sh[33]output_grids= -12: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -12: + interp_atmos_master.sh[35]gridopt=grid0p25 -12: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_13_0p25' -12: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -12: + interp_atmos_master.sh[35]gridopt=grid0p50 -12: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_13_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_13_0p50' -12: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -12: + interp_atmos_master.sh[35]gridopt=grid1p00 -12: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_13_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_13_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_13_1p00' -12: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_13 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_13_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_13_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_13_1p00 -12: 1:0:d=2021032312:ABSV:550 mb:24 hour fcst: -12: 2:20954:d=2021032312:CLMR:550 mb:24 hour fcst: -12: 3:24149:d=2021032312:ICMR:550 mb:24 hour fcst: -12: 4:40721:d=2021032312:RWMR:550 mb:24 hour fcst: -12: 5:42240:d=2021032312:SNMR:550 mb:24 hour fcst: -12: 6:56635:d=2021032312:GRLE:550 mb:24 hour fcst: -12: 7:58476:d=2021032312:O3MR:550 mb:24 hour fcst: -12: 8:81398:d=2021032312:HGT:600 mb:24 hour fcst: -12: 9:105623:d=2021032312:TMP:600 mb:24 hour fcst: -12: 10:124644:d=2021032312:RH:600 mb:24 hour fcst: -12: 11:144721:d=2021032312:TCDC:600 mb:24 hour fcst: -12: 12:156837:d=2021032312:SPFH:600 mb:24 hour fcst: -12: 13:187282:d=2021032312:VVEL:600 mb:24 hour fcst: -12: 14:212695:d=2021032312:DZDT:600 mb:24 hour fcst: -12: 15:240254:d=2021032312:UGRD:600 mb:24 hour fcst: -12: 16:261299:d=2021032312:VGRD:600 mb:24 hour fcst: -12: 17:282921:d=2021032312:ABSV:600 mb:24 hour fcst: -12: 18:303681:d=2021032312:CLMR:600 mb:24 hour fcst: -12: 19:308456:d=2021032312:ICMR:600 mb:24 hour fcst: -12: 20:322389:d=2021032312:RWMR:600 mb:24 hour fcst: -12: 21:325930:d=2021032312:SNMR:600 mb:24 hour fcst: -12: 22:341751:d=2021032312:GRLE:600 mb:24 hour fcst: -12: 23:344149:d=2021032312:O3MR:600 mb:24 hour fcst: -12: 24:372360:d=2021032312:HGT:650 mb:24 hour fcst: -12: 25:396536:d=2021032312:TMP:650 mb:24 hour fcst: -12: 26:415874:d=2021032312:RH:650 mb:24 hour fcst: -12: 27:436124:d=2021032312:TCDC:650 mb:24 hour fcst: -12: 28:447841:d=2021032312:SPFH:650 mb:24 hour fcst: -12: 29:476516:d=2021032312:VVEL:650 mb:24 hour fcst: -12: 30:502068:d=2021032312:DZDT:650 mb:24 hour fcst: -12: 31:529464:d=2021032312:UGRD:650 mb:24 hour fcst: -12: 32:550341:d=2021032312:VGRD:650 mb:24 hour fcst: -12: + interp_atmos_master.sh[47]export err=0 -12: + interp_atmos_master.sh[47]err=0 -12: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -12: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -12: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_13_0p25 -12: + product_functions.sh[5]local filename=pgb2file_f024_13_0p25 -12: + product_functions.sh[6]wgrib2 pgb2file_f024_13_0p25 -not_if :RH: -grib pgb2file_f024_13_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_13_0p25.new -12: 1:0:d=2021032312:ABSV:550 mb:24 hour fcst: -12: 2:496044:d=2021032312:CLMR:550 mb:24 hour fcst: -12: 3:566337:d=2021032312:ICMR:550 mb:24 hour fcst: -12: 4:931475:d=2021032312:RWMR:550 mb:24 hour fcst: -12: 5:956618:d=2021032312:SNMR:550 mb:24 hour fcst: -12: 6:1268492:d=2021032312:GRLE:550 mb:24 hour fcst: -12: 7:1300666:d=2021032312:O3MR:550 mb:24 hour fcst: -12: 8:1874374:d=2021032312:HGT:600 mb:24 hour fcst: -12: 9:2506892:d=2021032312:TMP:600 mb:24 hour fcst: -12: 10:2933084:d=2021032312:RH:600 mb:24 hour fcst: -12: 11:3400010:d=2021032312:TCDC:600 mb:24 hour fcst: -12: 12:3624651:d=2021032312:SPFH:600 mb:24 hour fcst: -12: 13:4451414:d=2021032312:VVEL:600 mb:24 hour fcst: -12: 14:5125097:d=2021032312:DZDT:600 mb:24 hour fcst: -12: 15:5891453:d=2021032312:UGRD:600 mb:24 hour fcst: -12: 16:6399386:d=2021032312:VGRD:600 mb:24 hour fcst: -12: 17:6935306:d=2021032312:ABSV:600 mb:24 hour fcst: -12: 18:7418678:d=2021032312:CLMR:600 mb:24 hour fcst: -12: 19:7518136:d=2021032312:ICMR:600 mb:24 hour fcst: -12: 20:7827871:d=2021032312:RWMR:600 mb:24 hour fcst: -12: 21:7902453:d=2021032312:SNMR:600 mb:24 hour fcst: -12: 22:8246202:d=2021032312:GRLE:600 mb:24 hour fcst: -12: 23:8290802:d=2021032312:O3MR:600 mb:24 hour fcst: -12: 24:9072760:d=2021032312:HGT:650 mb:24 hour fcst: -12: 25:9702418:d=2021032312:TMP:650 mb:24 hour fcst: -12: 26:10138879:d=2021032312:RH:650 mb:24 hour fcst: -12: 27:10610518:d=2021032312:TCDC:650 mb:24 hour fcst: -12: 28:10834691:d=2021032312:SPFH:650 mb:24 hour fcst: -12: 29:11608418:d=2021032312:VVEL:650 mb:24 hour fcst: -12: 30:12287205:d=2021032312:DZDT:650 mb:24 hour fcst: -12: 31:13043819:d=2021032312:UGRD:650 mb:24 hour fcst: -12: 32:13548237:d=2021032312:VGRD:650 mb:24 hour fcst: -12: + product_functions.sh[10]rc=0 -12: + product_functions.sh[11](( rc == 0 )) -12: + product_functions.sh[11]mv pgb2file_f024_13_0p25.new pgb2file_f024_13_0p25 -12: + product_functions.sh[12]return 0 -12: + interp_atmos_master.sh[56]export err=0 -12: + interp_atmos_master.sh[56]err=0 -12: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -12: ++ interp_atmos_master.sh[62]wc -l -12: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_13_0p25 -match 'LAND|ICEC' -12: + interp_atmos_master.sh[62]var_count=0 -12: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -12: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -12: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_13_0p50 -12: + product_functions.sh[5]local filename=pgb2file_f024_13_0p50 -12: + product_functions.sh[6]wgrib2 pgb2file_f024_13_0p50 -not_if :RH: -grib pgb2file_f024_13_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_13_0p50.new -12: 1:0:d=2021032312:ABSV:550 mb:24 hour fcst: -12: 2:172071:d=2021032312:CLMR:550 mb:24 hour fcst: -12: 3:197834:d=2021032312:ICMR:550 mb:24 hour fcst: -12: 4:330770:d=2021032312:RWMR:550 mb:24 hour fcst: -12: 5:340569:d=2021032312:SNMR:550 mb:24 hour fcst: -12: 6:451834:d=2021032312:GRLE:550 mb:24 hour fcst: -12: 7:464099:d=2021032312:O3MR:550 mb:24 hour fcst: -12: 8:660482:d=2021032312:HGT:600 mb:24 hour fcst: -12: 9:873250:d=2021032312:TMP:600 mb:24 hour fcst: -12: 10:1022304:d=2021032312:RH:600 mb:24 hour fcst: -12: 11:1183403:d=2021032312:TCDC:600 mb:24 hour fcst: -12: 12:1268763:d=2021032312:SPFH:600 mb:24 hour fcst: -12: 13:1558197:d=2021032312:VVEL:600 mb:24 hour fcst: -12: 14:1788124:d=2021032312:DZDT:600 mb:24 hour fcst: -12: 15:2038799:d=2021032312:UGRD:600 mb:24 hour fcst: -12: 16:2211805:d=2021032312:VGRD:600 mb:24 hour fcst: -12: 17:2394651:d=2021032312:ABSV:600 mb:24 hour fcst: -12: 18:2565029:d=2021032312:CLMR:600 mb:24 hour fcst: -12: 19:2602292:d=2021032312:ICMR:600 mb:24 hour fcst: -12: 20:2712775:d=2021032312:RWMR:600 mb:24 hour fcst: -12: 21:2739340:d=2021032312:SNMR:600 mb:24 hour fcst: -12: 22:2863420:d=2021032312:GRLE:600 mb:24 hour fcst: -12: 23:2880390:d=2021032312:O3MR:600 mb:24 hour fcst: -12: 24:3141949:d=2021032312:HGT:650 mb:24 hour fcst: -12: 25:3354338:d=2021032312:TMP:650 mb:24 hour fcst: -12: 26:3505401:d=2021032312:RH:650 mb:24 hour fcst: -12: 27:3668180:d=2021032312:TCDC:650 mb:24 hour fcst: -12: 28:3753157:d=2021032312:SPFH:650 mb:24 hour fcst: -12: 29:4023355:d=2021032312:VVEL:650 mb:24 hour fcst: -12: 30:4255189:d=2021032312:DZDT:650 mb:24 hour fcst: -12: 31:4505049:d=2021032312:UGRD:650 mb:24 hour fcst: -12: 32:4678646:d=2021032312:VGRD:650 mb:24 hour fcst: -12: + product_functions.sh[10]rc=0 -12: + product_functions.sh[11](( rc == 0 )) -12: + product_functions.sh[11]mv pgb2file_f024_13_0p50.new pgb2file_f024_13_0p50 -12: + product_functions.sh[12]return 0 -12: + interp_atmos_master.sh[56]export err=0 -12: + interp_atmos_master.sh[56]err=0 -12: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -12: ++ interp_atmos_master.sh[62]wc -l -12: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_13_0p50 -match 'LAND|ICEC' -12: + interp_atmos_master.sh[62]var_count=0 -12: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -12: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -12: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_13_1p00 -12: + product_functions.sh[5]local filename=pgb2file_f024_13_1p00 -12: + product_functions.sh[6]wgrib2 pgb2file_f024_13_1p00 -not_if :RH: -grib pgb2file_f024_13_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_13_1p00.new -12: 1:0:d=2021032312:ABSV:550 mb:24 hour fcst: -12: 2:55381:d=2021032312:CLMR:550 mb:24 hour fcst: -12: 3:64600:d=2021032312:ICMR:550 mb:24 hour fcst: -12: 4:108849:d=2021032312:RWMR:550 mb:24 hour fcst: -12: 5:112711:d=2021032312:SNMR:550 mb:24 hour fcst: -12: 6:149847:d=2021032312:GRLE:550 mb:24 hour fcst: -12: 7:154456:d=2021032312:O3MR:550 mb:24 hour fcst: -12: 8:216072:d=2021032312:HGT:600 mb:24 hour fcst: -12: 9:282521:d=2021032312:TMP:600 mb:24 hour fcst: -12: 10:331630:d=2021032312:RH:600 mb:24 hour fcst: -12: 11:384189:d=2021032312:TCDC:600 mb:24 hour fcst: -12: 12:415008:d=2021032312:SPFH:600 mb:24 hour fcst: -12: 13:501301:d=2021032312:VVEL:600 mb:24 hour fcst: -12: 14:571797:d=2021032312:DZDT:600 mb:24 hour fcst: -12: 15:647407:d=2021032312:UGRD:600 mb:24 hour fcst: -12: 16:703418:d=2021032312:VGRD:600 mb:24 hour fcst: -12: 17:761669:d=2021032312:ABSV:600 mb:24 hour fcst: -12: 18:816470:d=2021032312:CLMR:600 mb:24 hour fcst: -12: 19:829639:d=2021032312:ICMR:600 mb:24 hour fcst: -12: 20:866215:d=2021032312:RWMR:600 mb:24 hour fcst: -12: 21:875389:d=2021032312:SNMR:600 mb:24 hour fcst: -12: 22:916885:d=2021032312:GRLE:600 mb:24 hour fcst: -12: 23:923201:d=2021032312:O3MR:600 mb:24 hour fcst: -12: 24:1001657:d=2021032312:HGT:650 mb:24 hour fcst: -12: 25:1067932:d=2021032312:TMP:650 mb:24 hour fcst: -12: 26:1117720:d=2021032312:RH:650 mb:24 hour fcst: -12: 27:1170747:d=2021032312:TCDC:650 mb:24 hour fcst: -12: 28:1201388:d=2021032312:SPFH:650 mb:24 hour fcst: -12: 29:1282271:d=2021032312:VVEL:650 mb:24 hour fcst: -12: 30:1353435:d=2021032312:DZDT:650 mb:24 hour fcst: -12: 31:1428548:d=2021032312:UGRD:650 mb:24 hour fcst: -12: 32:1484349:d=2021032312:VGRD:650 mb:24 hour fcst: -12: + product_functions.sh[10]rc=0 -12: + product_functions.sh[11](( rc == 0 )) -12: + product_functions.sh[11]mv pgb2file_f024_13_1p00.new pgb2file_f024_13_1p00 -12: + product_functions.sh[12]return 0 -12: + interp_atmos_master.sh[56]export err=0 -12: + interp_atmos_master.sh[56]err=0 -12: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -12: ++ interp_atmos_master.sh[62]wc -l -12: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_13_1p00 -match 'LAND|ICEC' -12: + interp_atmos_master.sh[62]var_count=0 -12: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -12: + interp_atmos_master.sh[73]exit 0 -19: + bash[8]'[' -z '' ']' -19: + bash[9]case "$-" in -19: + bash[12]__lmod_vx=x -19: + bash[16]'[' -n x ']' -19: + bash[16]set +x -19: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -19: Shell debugging restarted -19: + bash[224]unset __lmod_vx -19: + interp_atmos_master.sh[7]input_file=tmpfile_f024_20 -19: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_20 -19: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -19: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -19: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -19: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -19: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -19: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -19: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -19: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -19: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -19: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -19: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -19: + interp_atmos_master.sh[31]IFS=: -19: + interp_atmos_master.sh[31]read -ra grids -19: + interp_atmos_master.sh[33]output_grids= -19: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -19: + interp_atmos_master.sh[35]gridopt=grid0p25 -19: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_20_0p25' -19: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -19: + interp_atmos_master.sh[35]gridopt=grid0p50 -19: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_20_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_20_0p50' -19: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -19: + interp_atmos_master.sh[35]gridopt=grid1p00 -19: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_20_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_20_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_20_1p00' -19: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_20 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_20_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_20_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_20_1p00 -19: 1:0:d=2021032312:CPRAT:surface:18-24 hour ave fcst: -19: 2:24054:d=2021032312:PRATE:surface:18-24 hour ave fcst: -19: 3:49112:d=2021032312:APCP:surface:18-24 hour acc fcst: -19: 4:59330:d=2021032312:APCP:surface:0-1 day acc fcst: -19: 5:72299:d=2021032312:ACPCP:surface:18-24 hour acc fcst: -19: 6:80638:d=2021032312:ACPCP:surface:0-1 day acc fcst: -19: 7:91630:d=2021032312:FROZR:surface:0-1 day acc fcst: -19: 8:97931:d=2021032312:FRZR:surface:0-1 day acc fcst: -19: 9:101034:d=2021032312:TSNOWP:surface:0-1 day acc fcst: -19: 10:108580:d=2021032312:WATR:surface:18-24 hour acc fcst: -19: 11:120991:d=2021032312:FROZR:surface:18-24 hour acc fcst: -19: 12:124951:d=2021032312:FRZR:surface:18-24 hour acc fcst: -19: 13:126724:d=2021032312:TSNOWP:surface:18-24 hour acc fcst: -19: 14:134650:d=2021032312:CSNOW:surface:24 hour fcst: -19: 15:136430:d=2021032312:CICEP:surface:24 hour fcst: -19: 16:136609:d=2021032312:CFRZR:surface:24 hour fcst: -19: 17:136829:d=2021032312:CRAIN:surface:24 hour fcst: -19: 18:140364:d=2021032312:LHTFL:surface:18-24 hour ave fcst: -19: 19:162008:d=2021032312:SHTFL:surface:18-24 hour ave fcst: -19: 20:182669:d=2021032312:GFLUX:surface:18-24 hour ave fcst: -19: 21:194727:d=2021032312:UFLX:surface:18-24 hour ave fcst: -19: 22:210970:d=2021032312:VFLX:surface:18-24 hour ave fcst: -19: 23:227333:d=2021032312:SFCR:surface:24 hour fcst: -19: 24:245799:d=2021032312:FRICV:surface:24 hour fcst: -19: 25:267503:d=2021032312:U-GWD:surface:18-24 hour ave fcst: -19: 26:282262:d=2021032312:V-GWD:surface:18-24 hour ave fcst: -19: 27:297045:d=2021032312:VEG:surface:24 hour fcst: -19: 28:305069:d=2021032312:SOTYP:surface:24 hour fcst: -19: 29:316118:d=2021032312:WILT:surface:24 hour fcst: -19: 30:326028:d=2021032312:FLDCP:surface:24 hour fcst: -19: 31:336234:d=2021032312:SUNSD:surface:24 hour fcst: -19: + interp_atmos_master.sh[47]export err=0 -19: + interp_atmos_master.sh[47]err=0 -19: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -19: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -19: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_20_0p25 -19: + product_functions.sh[5]local filename=pgb2file_f024_20_0p25 -19: + product_functions.sh[6]wgrib2 pgb2file_f024_20_0p25 -not_if :RH: -grib pgb2file_f024_20_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_20_0p25.new -19: 1:0:d=2021032312:CPRAT:surface:18-24 hour ave fcst: -19: 2:620796:d=2021032312:PRATE:surface:18-24 hour ave fcst: -19: 3:1257652:d=2021032312:APCP:surface:18-24 hour acc fcst: -19: 4:1450326:d=2021032312:APCP:surface:0-1 day acc fcst: -19: 5:1709783:d=2021032312:ACPCP:surface:18-24 hour acc fcst: -19: 6:1865242:d=2021032312:ACPCP:surface:0-1 day acc fcst: -19: 7:2080777:d=2021032312:FROZR:surface:0-1 day acc fcst: -19: 8:2202630:d=2021032312:FRZR:surface:0-1 day acc fcst: -19: 9:2256918:d=2021032312:TSNOWP:surface:0-1 day acc fcst: -19: 10:2398199:d=2021032312:WATR:surface:18-24 hour acc fcst: -19: 11:2763707:d=2021032312:FROZR:surface:18-24 hour acc fcst: -19: 12:2837442:d=2021032312:FRZR:surface:18-24 hour acc fcst: -19: 13:2865304:d=2021032312:TSNOWP:surface:18-24 hour acc fcst: -19: 14:3018914:d=2021032312:CSNOW:surface:24 hour fcst: -19: 15:3039056:d=2021032312:CICEP:surface:24 hour fcst: -19: 16:3039235:d=2021032312:CFRZR:surface:24 hour fcst: -19: 17:3039514:d=2021032312:CRAIN:surface:24 hour fcst: -19: 18:3106346:d=2021032312:LHTFL:surface:18-24 hour ave fcst: -19: 19:3652434:d=2021032312:SHTFL:surface:18-24 hour ave fcst: -19: 20:4141015:d=2021032312:GFLUX:surface:18-24 hour ave fcst: -19: 21:4508408:d=2021032312:UFLX:surface:18-24 hour ave fcst: -19: 22:4864880:d=2021032312:VFLX:surface:18-24 hour ave fcst: -19: 23:5221747:d=2021032312:SFCR:surface:24 hour fcst: -19: 24:5614282:d=2021032312:FRICV:surface:24 hour fcst: -19: 25:6125319:d=2021032312:U-GWD:surface:18-24 hour ave fcst: -19: 26:6460305:d=2021032312:V-GWD:surface:18-24 hour ave fcst: -19: 27:6794553:d=2021032312:VEG:surface:24 hour fcst: -19: 28:7067732:d=2021032312:SOTYP:surface:24 hour fcst: -19: 29:7305599:d=2021032312:WILT:surface:24 hour fcst: -19: 30:7618080:d=2021032312:FLDCP:surface:24 hour fcst: -19: 31:7939841:d=2021032312:SUNSD:surface:24 hour fcst: -19: + product_functions.sh[10]rc=0 -19: + product_functions.sh[11](( rc == 0 )) -19: + product_functions.sh[11]mv pgb2file_f024_20_0p25.new pgb2file_f024_20_0p25 -19: + product_functions.sh[12]return 0 -19: + interp_atmos_master.sh[56]export err=0 -19: + interp_atmos_master.sh[56]err=0 -19: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -19: ++ interp_atmos_master.sh[62]wc -l -19: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_20_0p25 -match 'LAND|ICEC' -19: + interp_atmos_master.sh[62]var_count=0 -19: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -19: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -19: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_20_0p50 -19: + product_functions.sh[5]local filename=pgb2file_f024_20_0p50 -19: + product_functions.sh[6]wgrib2 pgb2file_f024_20_0p50 -not_if :RH: -grib pgb2file_f024_20_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_20_0p50.new -19: 1:0:d=2021032312:CPRAT:surface:18-24 hour ave fcst: -19: 2:210328:d=2021032312:PRATE:surface:18-24 hour ave fcst: -19: 3:422932:d=2021032312:APCP:surface:18-24 hour acc fcst: -19: 4:486605:d=2021032312:APCP:surface:0-1 day acc fcst: -19: 5:574047:d=2021032312:ACPCP:surface:18-24 hour acc fcst: -19: 6:625464:d=2021032312:ACPCP:surface:0-1 day acc fcst: -19: 7:697819:d=2021032312:FROZR:surface:0-1 day acc fcst: -19: 8:741464:d=2021032312:FRZR:surface:0-1 day acc fcst: -19: 9:761666:d=2021032312:TSNOWP:surface:0-1 day acc fcst: -19: 10:813309:d=2021032312:WATR:surface:18-24 hour acc fcst: -19: 11:928309:d=2021032312:FROZR:surface:18-24 hour acc fcst: -19: 12:955260:d=2021032312:FRZR:surface:18-24 hour acc fcst: -19: 13:965706:d=2021032312:TSNOWP:surface:18-24 hour acc fcst: -19: 14:1022653:d=2021032312:CSNOW:surface:24 hour fcst: -19: 15:1031798:d=2021032312:CICEP:surface:24 hour fcst: -19: 16:1031977:d=2021032312:CFRZR:surface:24 hour fcst: -19: 17:1032218:d=2021032312:CRAIN:surface:24 hour fcst: -19: 18:1056028:d=2021032312:LHTFL:surface:18-24 hour ave fcst: -19: 19:1239549:d=2021032312:SHTFL:surface:18-24 hour ave fcst: -19: 20:1407983:d=2021032312:GFLUX:surface:18-24 hour ave fcst: -19: 21:1521469:d=2021032312:UFLX:surface:18-24 hour ave fcst: -19: 22:1641165:d=2021032312:VFLX:surface:18-24 hour ave fcst: -19: 23:1760985:d=2021032312:SFCR:surface:24 hour fcst: -19: 24:1911704:d=2021032312:FRICV:surface:24 hour fcst: -19: 25:2089802:d=2021032312:U-GWD:surface:18-24 hour ave fcst: -19: 26:2211538:d=2021032312:V-GWD:surface:18-24 hour ave fcst: -19: 27:2332338:d=2021032312:VEG:surface:24 hour fcst: -19: 28:2414171:d=2021032312:SOTYP:surface:24 hour fcst: -19: 29:2502302:d=2021032312:WILT:surface:24 hour fcst: -19: 30:2599606:d=2021032312:FLDCP:surface:24 hour fcst: -19: 31:2700330:d=2021032312:SUNSD:surface:24 hour fcst: -19: + product_functions.sh[10]rc=0 -19: + product_functions.sh[11](( rc == 0 )) -19: + product_functions.sh[11]mv pgb2file_f024_20_0p50.new pgb2file_f024_20_0p50 -19: + product_functions.sh[12]return 0 -19: + interp_atmos_master.sh[56]export err=0 -19: + interp_atmos_master.sh[56]err=0 -19: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -19: ++ interp_atmos_master.sh[62]wc -l -19: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_20_0p50 -match 'LAND|ICEC' -19: + interp_atmos_master.sh[62]var_count=0 -19: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -19: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -19: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_20_1p00 -19: + product_functions.sh[5]local filename=pgb2file_f024_20_1p00 -19: + product_functions.sh[6]wgrib2 pgb2file_f024_20_1p00 -not_if :RH: -grib pgb2file_f024_20_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_20_1p00.new -19: 1:0:d=2021032312:CPRAT:surface:18-24 hour ave fcst: -19: 2:65604:d=2021032312:PRATE:surface:18-24 hour ave fcst: -19: 3:132390:d=2021032312:APCP:surface:18-24 hour acc fcst: -19: 4:154995:d=2021032312:APCP:surface:0-1 day acc fcst: -19: 5:185284:d=2021032312:ACPCP:surface:18-24 hour acc fcst: -19: 6:203181:d=2021032312:ACPCP:surface:0-1 day acc fcst: -19: 7:228063:d=2021032312:FROZR:surface:0-1 day acc fcst: -19: 8:243351:d=2021032312:FRZR:surface:0-1 day acc fcst: -19: 9:250877:d=2021032312:TSNOWP:surface:0-1 day acc fcst: -19: 10:269306:d=2021032312:WATR:surface:18-24 hour acc fcst: -19: 11:304210:d=2021032312:FROZR:surface:18-24 hour acc fcst: -19: 12:314110:d=2021032312:FRZR:surface:18-24 hour acc fcst: -19: 13:318259:d=2021032312:TSNOWP:surface:18-24 hour acc fcst: -19: 14:338405:d=2021032312:CSNOW:surface:24 hour fcst: -19: 15:342103:d=2021032312:CICEP:surface:24 hour fcst: -19: 16:342282:d=2021032312:CFRZR:surface:24 hour fcst: -19: 17:342507:d=2021032312:CRAIN:surface:24 hour fcst: -19: 18:350384:d=2021032312:LHTFL:surface:18-24 hour ave fcst: -19: 19:408495:d=2021032312:SHTFL:surface:18-24 hour ave fcst: -19: 20:462986:d=2021032312:GFLUX:surface:18-24 hour ave fcst: -19: 21:496833:d=2021032312:UFLX:surface:18-24 hour ave fcst: -19: 22:537284:d=2021032312:VFLX:surface:18-24 hour ave fcst: -19: 23:577607:d=2021032312:SFCR:surface:24 hour fcst: -19: 24:626055:d=2021032312:FRICV:surface:24 hour fcst: -19: 25:683516:d=2021032312:U-GWD:surface:18-24 hour ave fcst: -19: 26:723428:d=2021032312:V-GWD:surface:18-24 hour ave fcst: -19: 27:763106:d=2021032312:VEG:surface:24 hour fcst: -19: 28:786961:d=2021032312:SOTYP:surface:24 hour fcst: -19: 29:816359:d=2021032312:WILT:surface:24 hour fcst: -19: 30:845136:d=2021032312:FLDCP:surface:24 hour fcst: -19: 31:874786:d=2021032312:SUNSD:surface:24 hour fcst: -19: + product_functions.sh[10]rc=0 -19: + product_functions.sh[11](( rc == 0 )) -19: + product_functions.sh[11]mv pgb2file_f024_20_1p00.new pgb2file_f024_20_1p00 -19: + product_functions.sh[12]return 0 -19: + interp_atmos_master.sh[56]export err=0 -19: + interp_atmos_master.sh[56]err=0 -19: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -19: ++ interp_atmos_master.sh[62]wc -l -19: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_20_1p00 -match 'LAND|ICEC' -19: + interp_atmos_master.sh[62]var_count=0 -19: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -19: + interp_atmos_master.sh[73]exit 0 - 8: + bash[8]'[' -z '' ']' - 8: + bash[9]case "$-" in - 8: + bash[12]__lmod_vx=x - 8: + bash[16]'[' -n x ']' - 8: + bash[16]set +x - 8: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 8: Shell debugging restarted - 8: + bash[224]unset __lmod_vx - 8: + interp_atmos_master.sh[7]input_file=tmpfile_f024_9 - 8: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_9 - 8: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 - 8: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 8: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 8: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 8: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 8: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 8: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 8: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 8: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 8: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 8: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 8: + interp_atmos_master.sh[31]IFS=: - 8: + interp_atmos_master.sh[31]read -ra grids - 8: + interp_atmos_master.sh[33]output_grids= - 8: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 8: + interp_atmos_master.sh[35]gridopt=grid0p25 - 8: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_9_0p25' - 8: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 8: + interp_atmos_master.sh[35]gridopt=grid0p50 - 8: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_9_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_9_0p50' - 8: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 8: + interp_atmos_master.sh[35]gridopt=grid1p00 - 8: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_9_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_9_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_9_1p00' - 8: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_9 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_9_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_9_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_9_1p00 - 8: 1:0:d=2021032312:SNMR:150 mb:24 hour fcst: - 8: 2:3519:d=2021032312:GRLE:150 mb:24 hour fcst: - 8: 3:4071:d=2021032312:O3MR:150 mb:24 hour fcst: - 8: 4:25981:d=2021032312:HGT:200 mb:24 hour fcst: - 8: 5:47019:d=2021032312:TMP:200 mb:24 hour fcst: - 8: 6:65124:d=2021032312:RH:200 mb:24 hour fcst: - 8: 7:80210:d=2021032312:TCDC:200 mb:24 hour fcst: - 8: 8:86702:d=2021032312:SPFH:200 mb:24 hour fcst: - 8: 9:112471:d=2021032312:VVEL:200 mb:24 hour fcst: - 8: 10:139206:d=2021032312:DZDT:200 mb:24 hour fcst: - 8: 11:165555:d=2021032312:UGRD:200 mb:24 hour fcst: - 8: 12:178958:d=2021032312:VGRD:200 mb:24 hour fcst: - 8: 13:200235:d=2021032312:ABSV:200 mb:24 hour fcst: - 8: 14:220183:d=2021032312:CLMR:200 mb:24 hour fcst: - 8: 15:220362:d=2021032312:ICMR:200 mb:24 hour fcst: - 8: 16:228773:d=2021032312:RWMR:200 mb:24 hour fcst: - 8: 17:228952:d=2021032312:SNMR:200 mb:24 hour fcst: - 8: 18:234260:d=2021032312:GRLE:200 mb:24 hour fcst: - 8: 19:234679:d=2021032312:O3MR:200 mb:24 hour fcst: - 8: 20:255333:d=2021032312:HGT:250 mb:24 hour fcst: - 8: 21:277258:d=2021032312:TMP:250 mb:24 hour fcst: - 8: 22:296144:d=2021032312:RH:250 mb:24 hour fcst: - 8: 23:313935:d=2021032312:TCDC:250 mb:24 hour fcst: - 8: 24:324473:d=2021032312:SPFH:250 mb:24 hour fcst: - 8: 25:352776:d=2021032312:VVEL:250 mb:24 hour fcst: - 8: 26:381161:d=2021032312:DZDT:250 mb:24 hour fcst: - 8: 27:408308:d=2021032312:UGRD:250 mb:24 hour fcst: - 8: 28:422608:d=2021032312:VGRD:250 mb:24 hour fcst: - 8: 29:437425:d=2021032312:ABSV:250 mb:24 hour fcst: - 8: 30:458874:d=2021032312:CLMR:250 mb:24 hour fcst: - 8: 31:459053:d=2021032312:ICMR:250 mb:24 hour fcst: - 8: + interp_atmos_master.sh[47]export err=0 - 8: + interp_atmos_master.sh[47]err=0 - 8: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 8: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 8: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_9_0p25 - 8: + product_functions.sh[5]local filename=pgb2file_f024_9_0p25 - 8: + product_functions.sh[6]wgrib2 pgb2file_f024_9_0p25 -not_if :RH: -grib pgb2file_f024_9_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_9_0p25.new - 8: 1:0:d=2021032312:SNMR:150 mb:24 hour fcst: - 8: 2:57276:d=2021032312:GRLE:150 mb:24 hour fcst: - 8: 3:65482:d=2021032312:O3MR:150 mb:24 hour fcst: - 8: 4:596337:d=2021032312:HGT:200 mb:24 hour fcst: - 8: 5:1122867:d=2021032312:TMP:200 mb:24 hour fcst: - 8: 6:1534264:d=2021032312:RH:200 mb:24 hour fcst: - 8: 7:1882646:d=2021032312:TCDC:200 mb:24 hour fcst: - 8: 8:2018229:d=2021032312:SPFH:200 mb:24 hour fcst: - 8: 9:2692697:d=2021032312:VVEL:200 mb:24 hour fcst: - 8: 10:3414976:d=2021032312:DZDT:200 mb:24 hour fcst: - 8: 11:4144323:d=2021032312:UGRD:200 mb:24 hour fcst: - 8: 12:4671199:d=2021032312:VGRD:200 mb:24 hour fcst: - 8: 13:5207947:d=2021032312:ABSV:200 mb:24 hour fcst: - 8: 14:5680778:d=2021032312:CLMR:200 mb:24 hour fcst: - 8: 15:5680957:d=2021032312:ICMR:200 mb:24 hour fcst: - 8: 16:5859239:d=2021032312:RWMR:200 mb:24 hour fcst: - 8: 17:5859418:d=2021032312:SNMR:200 mb:24 hour fcst: - 8: 18:5960457:d=2021032312:GRLE:200 mb:24 hour fcst: - 8: 19:5964825:d=2021032312:O3MR:200 mb:24 hour fcst: - 8: 20:6448584:d=2021032312:HGT:250 mb:24 hour fcst: - 8: 21:6997671:d=2021032312:TMP:250 mb:24 hour fcst: - 8: 22:7420602:d=2021032312:RH:250 mb:24 hour fcst: - 8: 23:7829164:d=2021032312:TCDC:250 mb:24 hour fcst: - 8: 24:8036052:d=2021032312:SPFH:250 mb:24 hour fcst: - 8: 25:8792389:d=2021032312:VVEL:250 mb:24 hour fcst: - 8: 26:9575753:d=2021032312:DZDT:250 mb:24 hour fcst: - 8: 27:10339458:d=2021032312:UGRD:250 mb:24 hour fcst: - 8: 28:10653546:d=2021032312:VGRD:250 mb:24 hour fcst: - 8: 29:10982158:d=2021032312:ABSV:250 mb:24 hour fcst: - 8: 30:11507773:d=2021032312:CLMR:250 mb:24 hour fcst: - 8: 31:11507952:d=2021032312:ICMR:250 mb:24 hour fcst: - 8: + product_functions.sh[10]rc=0 - 8: + product_functions.sh[11](( rc == 0 )) - 8: + product_functions.sh[11]mv pgb2file_f024_9_0p25.new pgb2file_f024_9_0p25 - 8: + product_functions.sh[12]return 0 - 8: + interp_atmos_master.sh[56]export err=0 - 8: + interp_atmos_master.sh[56]err=0 - 8: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 8: ++ interp_atmos_master.sh[62]wc -l - 8: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_9_0p25 -match 'LAND|ICEC' - 8: + interp_atmos_master.sh[62]var_count=0 - 8: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 8: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 8: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_9_0p50 - 8: + product_functions.sh[5]local filename=pgb2file_f024_9_0p50 - 8: + product_functions.sh[6]wgrib2 pgb2file_f024_9_0p50 -not_if :RH: -grib pgb2file_f024_9_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_9_0p50.new - 8: 1:0:d=2021032312:SNMR:150 mb:24 hour fcst: - 8: 2:19596:d=2021032312:GRLE:150 mb:24 hour fcst: - 8: 3:22859:d=2021032312:O3MR:150 mb:24 hour fcst: - 8: 4:205237:d=2021032312:HGT:200 mb:24 hour fcst: - 8: 5:380906:d=2021032312:TMP:200 mb:24 hour fcst: - 8: 6:521568:d=2021032312:RH:200 mb:24 hour fcst: - 8: 7:638635:d=2021032312:TCDC:200 mb:24 hour fcst: - 8: 8:688062:d=2021032312:SPFH:200 mb:24 hour fcst: - 8: 9:919565:d=2021032312:VVEL:200 mb:24 hour fcst: - 8: 10:1167362:d=2021032312:DZDT:200 mb:24 hour fcst: - 8: 11:1406801:d=2021032312:UGRD:200 mb:24 hour fcst: - 8: 12:1587878:d=2021032312:VGRD:200 mb:24 hour fcst: - 8: 13:1768380:d=2021032312:ABSV:200 mb:24 hour fcst: - 8: 14:1930359:d=2021032312:CLMR:200 mb:24 hour fcst: - 8: 15:1930538:d=2021032312:ICMR:200 mb:24 hour fcst: - 8: 16:1994745:d=2021032312:RWMR:200 mb:24 hour fcst: - 8: 17:1994924:d=2021032312:SNMR:200 mb:24 hour fcst: - 8: 18:2030746:d=2021032312:GRLE:200 mb:24 hour fcst: - 8: 19:2032609:d=2021032312:O3MR:200 mb:24 hour fcst: - 8: 20:2198462:d=2021032312:HGT:250 mb:24 hour fcst: - 8: 21:2382576:d=2021032312:TMP:250 mb:24 hour fcst: - 8: 22:2527616:d=2021032312:RH:250 mb:24 hour fcst: - 8: 23:2666293:d=2021032312:TCDC:250 mb:24 hour fcst: - 8: 24:2744762:d=2021032312:SPFH:250 mb:24 hour fcst: - 8: 25:3006560:d=2021032312:VVEL:250 mb:24 hour fcst: - 8: 26:3273679:d=2021032312:DZDT:250 mb:24 hour fcst: - 8: 27:3522066:d=2021032312:UGRD:250 mb:24 hour fcst: - 8: 28:3625044:d=2021032312:VGRD:250 mb:24 hour fcst: - 8: 29:3733556:d=2021032312:ABSV:250 mb:24 hour fcst: - 8: 30:3912509:d=2021032312:CLMR:250 mb:24 hour fcst: - 8: 31:3912688:d=2021032312:ICMR:250 mb:24 hour fcst: - 8: + product_functions.sh[10]rc=0 - 8: + product_functions.sh[11](( rc == 0 )) - 8: + product_functions.sh[11]mv pgb2file_f024_9_0p50.new pgb2file_f024_9_0p50 - 8: + product_functions.sh[12]return 0 - 8: + interp_atmos_master.sh[56]export err=0 - 8: + interp_atmos_master.sh[56]err=0 - 8: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 8: ++ interp_atmos_master.sh[62]wc -l - 8: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_9_0p50 -match 'LAND|ICEC' - 8: + interp_atmos_master.sh[62]var_count=0 - 8: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 8: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 8: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_9_1p00 - 8: + product_functions.sh[5]local filename=pgb2file_f024_9_1p00 - 8: + product_functions.sh[6]wgrib2 pgb2file_f024_9_1p00 -not_if :RH: -grib pgb2file_f024_9_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_9_1p00.new - 8: 1:0:d=2021032312:SNMR:150 mb:24 hour fcst: - 8: 2:7483:d=2021032312:GRLE:150 mb:24 hour fcst: - 8: 3:8752:d=2021032312:O3MR:150 mb:24 hour fcst: - 8: 4:67048:d=2021032312:HGT:200 mb:24 hour fcst: - 8: 5:122988:d=2021032312:TMP:200 mb:24 hour fcst: - 8: 6:168881:d=2021032312:RH:200 mb:24 hour fcst: - 8: 7:207195:d=2021032312:TCDC:200 mb:24 hour fcst: - 8: 8:224244:d=2021032312:SPFH:200 mb:24 hour fcst: - 8: 9:294519:d=2021032312:VVEL:200 mb:24 hour fcst: - 8: 10:369347:d=2021032312:DZDT:200 mb:24 hour fcst: - 8: 11:441413:d=2021032312:UGRD:200 mb:24 hour fcst: - 8: 12:498845:d=2021032312:VGRD:200 mb:24 hour fcst: - 8: 13:555999:d=2021032312:ABSV:200 mb:24 hour fcst: - 8: 14:608322:d=2021032312:CLMR:200 mb:24 hour fcst: - 8: 15:608501:d=2021032312:ICMR:200 mb:24 hour fcst: - 8: 16:630018:d=2021032312:RWMR:200 mb:24 hour fcst: - 8: 17:630197:d=2021032312:SNMR:200 mb:24 hour fcst: - 8: 18:642951:d=2021032312:GRLE:200 mb:24 hour fcst: - 8: 19:643777:d=2021032312:O3MR:200 mb:24 hour fcst: - 8: 20:697884:d=2021032312:HGT:250 mb:24 hour fcst: - 8: 21:756605:d=2021032312:TMP:250 mb:24 hour fcst: - 8: 22:804776:d=2021032312:RH:250 mb:24 hour fcst: - 8: 23:850812:d=2021032312:TCDC:250 mb:24 hour fcst: - 8: 24:878493:d=2021032312:SPFH:250 mb:24 hour fcst: - 8: 25:956861:d=2021032312:VVEL:250 mb:24 hour fcst: - 8: 26:1036723:d=2021032312:DZDT:250 mb:24 hour fcst: - 8: 27:1111595:d=2021032312:UGRD:250 mb:24 hour fcst: - 8: 28:1146403:d=2021032312:VGRD:250 mb:24 hour fcst: - 8: 29:1182979:d=2021032312:ABSV:250 mb:24 hour fcst: - 8: 30:1240450:d=2021032312:CLMR:250 mb:24 hour fcst: - 8: 31:1240629:d=2021032312:ICMR:250 mb:24 hour fcst: - 8: + product_functions.sh[10]rc=0 - 8: + product_functions.sh[11](( rc == 0 )) - 8: + product_functions.sh[11]mv pgb2file_f024_9_1p00.new pgb2file_f024_9_1p00 - 8: + product_functions.sh[12]return 0 - 8: + interp_atmos_master.sh[56]export err=0 - 8: + interp_atmos_master.sh[56]err=0 - 8: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 8: ++ interp_atmos_master.sh[62]wc -l - 8: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_9_1p00 -match 'LAND|ICEC' - 8: + interp_atmos_master.sh[62]var_count=0 - 8: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 8: + interp_atmos_master.sh[73]exit 0 -16: + bash[8]'[' -z '' ']' -16: + bash[9]case "$-" in -16: + bash[12]__lmod_vx=x -16: + bash[16]'[' -n x ']' -16: + bash[16]set +x -16: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -16: Shell debugging restarted -16: + bash[224]unset __lmod_vx -16: + interp_atmos_master.sh[7]input_file=tmpfile_f024_17 -16: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_17 -16: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -16: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -16: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -16: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -16: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -16: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -16: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -16: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -16: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -16: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -16: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -16: + interp_atmos_master.sh[31]IFS=: -16: + interp_atmos_master.sh[31]read -ra grids -16: + interp_atmos_master.sh[33]output_grids= -16: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -16: + interp_atmos_master.sh[35]gridopt=grid0p25 -16: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_17_0p25' -16: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -16: + interp_atmos_master.sh[35]gridopt=grid0p50 -16: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_17_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_17_0p50' -16: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -16: + interp_atmos_master.sh[35]gridopt=grid1p00 -16: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_17_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_17_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_17_1p00' -16: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_17 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_17_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_17_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_17_1p00 -16: 1:0:d=2021032312:ABSV:925 mb:24 hour fcst: -16: 2:21292:d=2021032312:CLMR:925 mb:24 hour fcst: -16: 3:29995:d=2021032312:ICMR:925 mb:24 hour fcst: -16: 4:37623:d=2021032312:RWMR:925 mb:24 hour fcst: -16: 5:50129:d=2021032312:SNMR:925 mb:24 hour fcst: -16: 6:59609:d=2021032312:GRLE:925 mb:24 hour fcst: -16: 7:62793:d=2021032312:O3MR:925 mb:24 hour fcst: -16: 8:88824:d=2021032312:HGT:950 mb:24 hour fcst: -16: 9:114177:d=2021032312:TMP:950 mb:24 hour fcst: -16: 10:135296:d=2021032312:RH:950 mb:24 hour fcst: -16: 11:154824:d=2021032312:TCDC:950 mb:24 hour fcst: -16: 12:164989:d=2021032312:SPFH:950 mb:24 hour fcst: -16: 13:192191:d=2021032312:VVEL:950 mb:24 hour fcst: -16: 14:216014:d=2021032312:DZDT:950 mb:24 hour fcst: -16: 15:240693:d=2021032312:UGRD:950 mb:24 hour fcst: -16: 16:262131:d=2021032312:VGRD:950 mb:24 hour fcst: -16: 17:284031:d=2021032312:ABSV:950 mb:24 hour fcst: -16: 18:305258:d=2021032312:CLMR:950 mb:24 hour fcst: -16: 19:311683:d=2021032312:ICMR:950 mb:24 hour fcst: -16: 20:318858:d=2021032312:RWMR:950 mb:24 hour fcst: -16: 21:330174:d=2021032312:SNMR:950 mb:24 hour fcst: -16: 22:338926:d=2021032312:GRLE:950 mb:24 hour fcst: -16: 23:341811:d=2021032312:O3MR:950 mb:24 hour fcst: -16: 24:367805:d=2021032312:HINDEX:surface:24 hour fcst: -16: 25:372898:d=2021032312:HGT:975 mb:24 hour fcst: -16: 26:398410:d=2021032312:TMP:975 mb:24 hour fcst: -16: 27:419315:d=2021032312:RH:975 mb:24 hour fcst: -16: 28:438434:d=2021032312:TCDC:975 mb:24 hour fcst: -16: 29:446711:d=2021032312:SPFH:975 mb:24 hour fcst: -16: 30:473777:d=2021032312:VVEL:975 mb:24 hour fcst: -16: 31:496523:d=2021032312:DZDT:975 mb:24 hour fcst: -16: + interp_atmos_master.sh[47]export err=0 -16: + interp_atmos_master.sh[47]err=0 -16: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -16: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -16: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_17_0p25 -16: + product_functions.sh[5]local filename=pgb2file_f024_17_0p25 -16: + product_functions.sh[6]wgrib2 pgb2file_f024_17_0p25 -not_if :RH: -grib pgb2file_f024_17_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_17_0p25.new -16: 1:0:d=2021032312:ABSV:925 mb:24 hour fcst: -16: 2:497430:d=2021032312:CLMR:925 mb:24 hour fcst: -16: 3:691576:d=2021032312:ICMR:925 mb:24 hour fcst: -16: 4:866813:d=2021032312:RWMR:925 mb:24 hour fcst: -16: 5:1119334:d=2021032312:SNMR:925 mb:24 hour fcst: -16: 6:1336925:d=2021032312:GRLE:925 mb:24 hour fcst: -16: 7:1398731:d=2021032312:O3MR:925 mb:24 hour fcst: -16: 8:2104496:d=2021032312:HGT:950 mb:24 hour fcst: -16: 9:2777141:d=2021032312:TMP:950 mb:24 hour fcst: -16: 10:3273224:d=2021032312:RH:950 mb:24 hour fcst: -16: 11:3716271:d=2021032312:TCDC:950 mb:24 hour fcst: -16: 12:3916807:d=2021032312:SPFH:950 mb:24 hour fcst: -16: 13:4656146:d=2021032312:VVEL:950 mb:24 hour fcst: -16: 14:5268282:d=2021032312:DZDT:950 mb:24 hour fcst: -16: 15:5917892:d=2021032312:UGRD:950 mb:24 hour fcst: -16: 16:6437535:d=2021032312:VGRD:950 mb:24 hour fcst: -16: 17:6982412:d=2021032312:ABSV:950 mb:24 hour fcst: -16: 18:7485339:d=2021032312:CLMR:950 mb:24 hour fcst: -16: 19:7626563:d=2021032312:ICMR:950 mb:24 hour fcst: -16: 20:7792651:d=2021032312:RWMR:950 mb:24 hour fcst: -16: 21:8018116:d=2021032312:SNMR:950 mb:24 hour fcst: -16: 22:8219013:d=2021032312:GRLE:950 mb:24 hour fcst: -16: 23:8274524:d=2021032312:O3MR:950 mb:24 hour fcst: -16: 24:8977232:d=2021032312:HINDEX:surface:24 hour fcst: -16: 25:9156755:d=2021032312:HGT:975 mb:24 hour fcst: -16: 26:9836273:d=2021032312:TMP:975 mb:24 hour fcst: -16: 27:10325465:d=2021032312:RH:975 mb:24 hour fcst: -16: 28:10754304:d=2021032312:TCDC:975 mb:24 hour fcst: -16: 29:10913045:d=2021032312:SPFH:975 mb:24 hour fcst: -16: 30:11646927:d=2021032312:VVEL:975 mb:24 hour fcst: -16: 31:12219427:d=2021032312:DZDT:975 mb:24 hour fcst: -16: + product_functions.sh[10]rc=0 -16: + product_functions.sh[11](( rc == 0 )) -16: + product_functions.sh[11]mv pgb2file_f024_17_0p25.new pgb2file_f024_17_0p25 -16: + product_functions.sh[12]return 0 -16: + interp_atmos_master.sh[56]export err=0 -16: + interp_atmos_master.sh[56]err=0 -16: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -16: ++ interp_atmos_master.sh[62]wc -l -16: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_17_0p25 -match 'LAND|ICEC' -16: + interp_atmos_master.sh[62]var_count=0 -16: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -16: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -16: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_17_0p50 -16: + product_functions.sh[5]local filename=pgb2file_f024_17_0p50 -16: + product_functions.sh[6]wgrib2 pgb2file_f024_17_0p50 -not_if :RH: -grib pgb2file_f024_17_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_17_0p50.new -16: 1:0:d=2021032312:ABSV:925 mb:24 hour fcst: -16: 2:172994:d=2021032312:CLMR:925 mb:24 hour fcst: -16: 3:245634:d=2021032312:ICMR:925 mb:24 hour fcst: -16: 4:307359:d=2021032312:RWMR:925 mb:24 hour fcst: -16: 5:400798:d=2021032312:SNMR:925 mb:24 hour fcst: -16: 6:478523:d=2021032312:GRLE:925 mb:24 hour fcst: -16: 7:501258:d=2021032312:O3MR:925 mb:24 hour fcst: -16: 8:737979:d=2021032312:HGT:950 mb:24 hour fcst: -16: 9:962323:d=2021032312:TMP:950 mb:24 hour fcst: -16: 10:1134358:d=2021032312:RH:950 mb:24 hour fcst: -16: 11:1289970:d=2021032312:TCDC:950 mb:24 hour fcst: -16: 12:1364624:d=2021032312:SPFH:950 mb:24 hour fcst: -16: 13:1614628:d=2021032312:VVEL:950 mb:24 hour fcst: -16: 14:1823673:d=2021032312:DZDT:950 mb:24 hour fcst: -16: 15:2039963:d=2021032312:UGRD:950 mb:24 hour fcst: -16: 16:2219217:d=2021032312:VGRD:950 mb:24 hour fcst: -16: 17:2404081:d=2021032312:ABSV:950 mb:24 hour fcst: -16: 18:2578213:d=2021032312:CLMR:950 mb:24 hour fcst: -16: 19:2629709:d=2021032312:ICMR:950 mb:24 hour fcst: -16: 20:2688574:d=2021032312:RWMR:950 mb:24 hour fcst: -16: 21:2770021:d=2021032312:SNMR:950 mb:24 hour fcst: -16: 22:2839901:d=2021032312:GRLE:950 mb:24 hour fcst: -16: 23:2859847:d=2021032312:O3MR:950 mb:24 hour fcst: -16: 24:3096019:d=2021032312:HINDEX:surface:24 hour fcst: -16: 25:3146520:d=2021032312:HGT:975 mb:24 hour fcst: -16: 26:3373173:d=2021032312:TMP:975 mb:24 hour fcst: -16: 27:3542427:d=2021032312:RH:975 mb:24 hour fcst: -16: 28:3693415:d=2021032312:TCDC:975 mb:24 hour fcst: -16: 29:3752581:d=2021032312:SPFH:975 mb:24 hour fcst: -16: 30:4000680:d=2021032312:VVEL:975 mb:24 hour fcst: -16: 31:4196250:d=2021032312:DZDT:975 mb:24 hour fcst: -16: + product_functions.sh[10]rc=0 -16: + product_functions.sh[11](( rc == 0 )) -16: + product_functions.sh[11]mv pgb2file_f024_17_0p50.new pgb2file_f024_17_0p50 -16: + product_functions.sh[12]return 0 -16: + interp_atmos_master.sh[56]export err=0 -16: + interp_atmos_master.sh[56]err=0 -16: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -16: ++ interp_atmos_master.sh[62]wc -l -16: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_17_0p50 -match 'LAND|ICEC' -16: + interp_atmos_master.sh[62]var_count=0 -16: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -16: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -16: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_17_1p00 -16: + product_functions.sh[5]local filename=pgb2file_f024_17_1p00 -16: + product_functions.sh[6]wgrib2 pgb2file_f024_17_1p00 -not_if :RH: -grib pgb2file_f024_17_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_17_1p00.new -16: 1:0:d=2021032312:ABSV:925 mb:24 hour fcst: -16: 2:56317:d=2021032312:CLMR:925 mb:24 hour fcst: -16: 3:80991:d=2021032312:ICMR:925 mb:24 hour fcst: -16: 4:101288:d=2021032312:RWMR:925 mb:24 hour fcst: -16: 5:133710:d=2021032312:SNMR:925 mb:24 hour fcst: -16: 6:159007:d=2021032312:GRLE:925 mb:24 hour fcst: -16: 7:167184:d=2021032312:O3MR:925 mb:24 hour fcst: -16: 8:239562:d=2021032312:HGT:950 mb:24 hour fcst: -16: 9:309222:d=2021032312:TMP:950 mb:24 hour fcst: -16: 10:364927:d=2021032312:RH:950 mb:24 hour fcst: -16: 11:415864:d=2021032312:TCDC:950 mb:24 hour fcst: -16: 12:442782:d=2021032312:SPFH:950 mb:24 hour fcst: -16: 13:517909:d=2021032312:VVEL:950 mb:24 hour fcst: -16: 14:583130:d=2021032312:DZDT:950 mb:24 hour fcst: -16: 15:650103:d=2021032312:UGRD:950 mb:24 hour fcst: -16: 16:707570:d=2021032312:VGRD:950 mb:24 hour fcst: -16: 17:766695:d=2021032312:ABSV:950 mb:24 hour fcst: -16: 18:823030:d=2021032312:CLMR:950 mb:24 hour fcst: -16: 19:840672:d=2021032312:ICMR:950 mb:24 hour fcst: -16: 20:860099:d=2021032312:RWMR:950 mb:24 hour fcst: -16: 21:888799:d=2021032312:SNMR:950 mb:24 hour fcst: -16: 22:911921:d=2021032312:GRLE:950 mb:24 hour fcst: -16: 23:919159:d=2021032312:O3MR:950 mb:24 hour fcst: -16: 24:991400:d=2021032312:HINDEX:surface:24 hour fcst: -16: 25:1005996:d=2021032312:HGT:975 mb:24 hour fcst: -16: 26:1076623:d=2021032312:TMP:975 mb:24 hour fcst: -16: 27:1131547:d=2021032312:RH:975 mb:24 hour fcst: -16: 28:1181082:d=2021032312:TCDC:975 mb:24 hour fcst: -16: 29:1202985:d=2021032312:SPFH:975 mb:24 hour fcst: -16: 30:1277900:d=2021032312:VVEL:975 mb:24 hour fcst: -16: 31:1339311:d=2021032312:DZDT:975 mb:24 hour fcst: -16: + product_functions.sh[10]rc=0 -16: + product_functions.sh[11](( rc == 0 )) -16: + product_functions.sh[11]mv pgb2file_f024_17_1p00.new pgb2file_f024_17_1p00 -16: + product_functions.sh[12]return 0 -16: + interp_atmos_master.sh[56]export err=0 -16: + interp_atmos_master.sh[56]err=0 -16: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -16: ++ interp_atmos_master.sh[62]wc -l -16: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_17_1p00 -match 'LAND|ICEC' -16: + interp_atmos_master.sh[62]var_count=0 -16: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -16: + interp_atmos_master.sh[73]exit 0 -18: + bash[8]'[' -z '' ']' -18: + bash[9]case "$-" in -18: + bash[12]__lmod_vx=x -18: + bash[16]'[' -n x ']' -18: + bash[16]set +x -18: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -18: Shell debugging restarted -18: + bash[224]unset __lmod_vx -18: + interp_atmos_master.sh[7]input_file=tmpfile_f024_19 -18: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_19 -18: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -18: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -18: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -18: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -18: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -18: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -18: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -18: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -18: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -18: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -18: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -18: + interp_atmos_master.sh[31]IFS=: -18: + interp_atmos_master.sh[31]read -ra grids -18: + interp_atmos_master.sh[33]output_grids= -18: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -18: + interp_atmos_master.sh[35]gridopt=grid0p25 -18: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_19_0p25' -18: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -18: + interp_atmos_master.sh[35]gridopt=grid0p50 -18: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_19_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_19_0p50' -18: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -18: + interp_atmos_master.sh[35]gridopt=grid1p00 -18: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_19_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_19_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_19_1p00' -18: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_19 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_19_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_19_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_19_1p00 -18: 1:0:d=2021032312:TMP:surface:24 hour fcst: -18: 2:16630:d=2021032312:TSOIL:0-0.1 m below ground:24 hour fcst: -18: 3:28561:d=2021032312:SOILW:0-0.1 m below ground:24 hour fcst: -18: 4:37472:d=2021032312:SOILL:0-0.1 m below ground:24 hour fcst: -18: 5:46269:d=2021032312:TSOIL:0.1-0.4 m below ground:24 hour fcst: -18: 6:61729:d=2021032312:SOILW:0.1-0.4 m below ground:24 hour fcst: -18: 7:70746:d=2021032312:SOILL:0.1-0.4 m below ground:24 hour fcst: -18: 8:79640:d=2021032312:TSOIL:0.4-1 m below ground:24 hour fcst: -18: 9:94079:d=2021032312:SOILW:0.4-1 m below ground:24 hour fcst: -18: 10:103212:d=2021032312:SOILL:0.4-1 m below ground:24 hour fcst: -18: 11:112116:d=2021032312:TSOIL:1-2 m below ground:24 hour fcst: -18: 12:126378:d=2021032312:SOILW:1-2 m below ground:24 hour fcst: -18: 13:135438:d=2021032312:SOILL:1-2 m below ground:24 hour fcst: -18: 14:144514:d=2021032312:CNWAT:surface:24 hour fcst: -18: 15:150099:d=2021032312:WEASD:surface:24 hour fcst: -18: 16:162896:d=2021032312:SNOD:surface:24 hour fcst: -18: 17:177242:d=2021032312:PEVPR:surface:24 hour fcst: -18: 18:192892:d=2021032312:ICETK:surface:24 hour fcst: -18: 19:196306:d=2021032312:TMP:2 m above ground:24 hour fcst: -18: 20:217066:d=2021032312:SPFH:2 m above ground:24 hour fcst: -18: 21:245933:d=2021032312:DPT:2 m above ground:24 hour fcst: -18: 22:267172:d=2021032312:RH:2 m above ground:24 hour fcst: -18: 23:285709:d=2021032312:APTMP:2 m above ground:24 hour fcst: -18: 24:300087:d=2021032312:TMAX:2 m above ground:18-24 hour max fcst: -18: 25:320783:d=2021032312:TMIN:2 m above ground:18-24 hour min fcst: -18: 26:341440:d=2021032312:UGRD:10 m above ground:24 hour fcst: -18: 27:361955:d=2021032312:VGRD:10 m above ground:24 hour fcst: -18: 28:382996:d=2021032312:ICEG:10 m above mean sea level:24 hour fcst: -18: 29:383643:d=2021032312:CPOFP:surface:24 hour fcst: -18: 30:403750:d=2021032312:CPRAT:surface:24 hour fcst: -18: 31:425051:d=2021032312:PRATE:surface:24 hour fcst: -18: + interp_atmos_master.sh[47]export err=0 -18: + interp_atmos_master.sh[47]err=0 -18: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -18: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -18: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_19_0p25 -18: + product_functions.sh[5]local filename=pgb2file_f024_19_0p25 -18: + product_functions.sh[6]wgrib2 pgb2file_f024_19_0p25 -not_if :RH: -grib pgb2file_f024_19_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_19_0p25.new -18: 1:0:d=2021032312:TMP:surface:24 hour fcst: -18: 2:343612:d=2021032312:TSOIL:0-0.1 m below ground:24 hour fcst: -18: 3:678042:d=2021032312:SOILW:0-0.1 m below ground:24 hour fcst: -18: 4:964878:d=2021032312:SOILL:0-0.1 m below ground:24 hour fcst: -18: 5:1253520:d=2021032312:TSOIL:0.1-0.4 m below ground:24 hour fcst: -18: 6:1711815:d=2021032312:SOILW:0.1-0.4 m below ground:24 hour fcst: -18: 7:2003258:d=2021032312:SOILL:0.1-0.4 m below ground:24 hour fcst: -18: 8:2297242:d=2021032312:TSOIL:0.4-1 m below ground:24 hour fcst: -18: 9:2734263:d=2021032312:SOILW:0.4-1 m below ground:24 hour fcst: -18: 10:3030758:d=2021032312:SOILL:0.4-1 m below ground:24 hour fcst: -18: 11:3326304:d=2021032312:TSOIL:1-2 m below ground:24 hour fcst: -18: 12:3755465:d=2021032312:SOILW:1-2 m below ground:24 hour fcst: -18: 13:4046545:d=2021032312:SOILL:1-2 m below ground:24 hour fcst: -18: 14:4342793:d=2021032312:CNWAT:surface:24 hour fcst: -18: 15:4529348:d=2021032312:WEASD:surface:24 hour fcst: -18: 16:4897658:d=2021032312:SNOD:surface:24 hour fcst: -18: 17:5290912:d=2021032312:PEVPR:surface:24 hour fcst: -18: 18:5772754:d=2021032312:ICETK:surface:24 hour fcst: -18: 19:5825992:d=2021032312:TMP:2 m above ground:24 hour fcst: -18: 20:6308602:d=2021032312:SPFH:2 m above ground:24 hour fcst: -18: 21:7094330:d=2021032312:DPT:2 m above ground:24 hour fcst: -18: 22:7600083:d=2021032312:RH:2 m above ground:24 hour fcst: -18: 23:8009740:d=2021032312:APTMP:2 m above ground:24 hour fcst: -18: 24:8321016:d=2021032312:TMAX:2 m above ground:18-24 hour max fcst: -18: 25:8799684:d=2021032312:TMIN:2 m above ground:18-24 hour min fcst: -18: 26:9280417:d=2021032312:UGRD:10 m above ground:24 hour fcst: -18: 27:9770883:d=2021032312:VGRD:10 m above ground:24 hour fcst: -18: 28:10282197:d=2021032312:ICEG:10 m above mean sea level:24 hour fcst: -18: 29:10293396:d=2021032312:CPOFP:surface:24 hour fcst: -18: 30:10715126:d=2021032312:CPRAT:surface:24 hour fcst: -18: 31:11247383:d=2021032312:PRATE:surface:24 hour fcst: -18: + product_functions.sh[10]rc=0 -18: + product_functions.sh[11](( rc == 0 )) -18: + product_functions.sh[11]mv pgb2file_f024_19_0p25.new pgb2file_f024_19_0p25 -18: + product_functions.sh[12]return 0 -18: + interp_atmos_master.sh[56]export err=0 -18: + interp_atmos_master.sh[56]err=0 -18: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -18: ++ interp_atmos_master.sh[62]wc -l -18: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_19_0p25 -match 'LAND|ICEC' -18: + interp_atmos_master.sh[62]var_count=0 -18: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -18: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -18: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_19_0p50 -18: + product_functions.sh[5]local filename=pgb2file_f024_19_0p50 -18: + product_functions.sh[6]wgrib2 pgb2file_f024_19_0p50 -not_if :RH: -grib pgb2file_f024_19_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_19_0p50.new -18: 1:0:d=2021032312:TMP:surface:24 hour fcst: -18: 2:120223:d=2021032312:TSOIL:0-0.1 m below ground:24 hour fcst: -18: 3:225653:d=2021032312:SOILW:0-0.1 m below ground:24 hour fcst: -18: 4:313197:d=2021032312:SOILL:0-0.1 m below ground:24 hour fcst: -18: 5:400988:d=2021032312:TSOIL:0.1-0.4 m below ground:24 hour fcst: -18: 6:547293:d=2021032312:SOILW:0.1-0.4 m below ground:24 hour fcst: -18: 7:636118:d=2021032312:SOILL:0.1-0.4 m below ground:24 hour fcst: -18: 8:725111:d=2021032312:TSOIL:0.4-1 m below ground:24 hour fcst: -18: 9:863582:d=2021032312:SOILW:0.4-1 m below ground:24 hour fcst: -18: 10:954008:d=2021032312:SOILL:0.4-1 m below ground:24 hour fcst: -18: 11:1043303:d=2021032312:TSOIL:1-2 m below ground:24 hour fcst: -18: 12:1179634:d=2021032312:SOILW:1-2 m below ground:24 hour fcst: -18: 13:1268022:d=2021032312:SOILL:1-2 m below ground:24 hour fcst: -18: 14:1357490:d=2021032312:CNWAT:surface:24 hour fcst: -18: 15:1410879:d=2021032312:WEASD:surface:24 hour fcst: -18: 16:1532694:d=2021032312:SNOD:surface:24 hour fcst: -18: 17:1666529:d=2021032312:PEVPR:surface:24 hour fcst: -18: 18:1818234:d=2021032312:ICETK:surface:24 hour fcst: -18: 19:1838498:d=2021032312:TMP:2 m above ground:24 hour fcst: -18: 20:2006068:d=2021032312:SPFH:2 m above ground:24 hour fcst: -18: 21:2277386:d=2021032312:DPT:2 m above ground:24 hour fcst: -18: 22:2452170:d=2021032312:RH:2 m above ground:24 hour fcst: -18: 23:2595387:d=2021032312:APTMP:2 m above ground:24 hour fcst: -18: 24:2695338:d=2021032312:TMAX:2 m above ground:18-24 hour max fcst: -18: 25:2861830:d=2021032312:TMIN:2 m above ground:18-24 hour min fcst: -18: 26:3028385:d=2021032312:UGRD:10 m above ground:24 hour fcst: -18: 27:3198027:d=2021032312:VGRD:10 m above ground:24 hour fcst: -18: 28:3373293:d=2021032312:ICEG:10 m above mean sea level:24 hour fcst: -18: 29:3377412:d=2021032312:CPOFP:surface:24 hour fcst: -18: 30:3536158:d=2021032312:CPRAT:surface:24 hour fcst: -18: 31:3723190:d=2021032312:PRATE:surface:24 hour fcst: -18: + product_functions.sh[10]rc=0 -18: + product_functions.sh[11](( rc == 0 )) -18: + product_functions.sh[11]mv pgb2file_f024_19_0p50.new pgb2file_f024_19_0p50 -18: + product_functions.sh[12]return 0 -18: + interp_atmos_master.sh[56]export err=0 -18: + interp_atmos_master.sh[56]err=0 -18: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -18: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_19_0p50 -match 'LAND|ICEC' -18: ++ interp_atmos_master.sh[62]wc -l -18: + interp_atmos_master.sh[62]var_count=0 -18: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -18: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -18: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_19_1p00 -18: + product_functions.sh[5]local filename=pgb2file_f024_19_1p00 -18: + product_functions.sh[6]wgrib2 pgb2file_f024_19_1p00 -not_if :RH: -grib pgb2file_f024_19_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_19_1p00.new -18: 1:0:d=2021032312:TMP:surface:24 hour fcst: -18: 2:41050:d=2021032312:TSOIL:0-0.1 m below ground:24 hour fcst: -18: 3:73726:d=2021032312:SOILW:0-0.1 m below ground:24 hour fcst: -18: 4:99582:d=2021032312:SOILL:0-0.1 m below ground:24 hour fcst: -18: 5:125405:d=2021032312:TSOIL:0.1-0.4 m below ground:24 hour fcst: -18: 6:169418:d=2021032312:SOILW:0.1-0.4 m below ground:24 hour fcst: -18: 7:195437:d=2021032312:SOILL:0.1-0.4 m below ground:24 hour fcst: -18: 8:221414:d=2021032312:TSOIL:0.4-1 m below ground:24 hour fcst: -18: 9:262907:d=2021032312:SOILW:0.4-1 m below ground:24 hour fcst: -18: 10:289191:d=2021032312:SOILL:0.4-1 m below ground:24 hour fcst: -18: 11:315274:d=2021032312:TSOIL:1-2 m below ground:24 hour fcst: -18: 12:356209:d=2021032312:SOILW:1-2 m below ground:24 hour fcst: -18: 13:382139:d=2021032312:SOILL:1-2 m below ground:24 hour fcst: -18: 14:408280:d=2021032312:CNWAT:surface:24 hour fcst: -18: 15:424053:d=2021032312:WEASD:surface:24 hour fcst: -18: 16:460717:d=2021032312:SNOD:surface:24 hour fcst: -18: 17:501537:d=2021032312:PEVPR:surface:24 hour fcst: -18: 18:546078:d=2021032312:ICETK:surface:24 hour fcst: -18: 19:553926:d=2021032312:TMP:2 m above ground:24 hour fcst: -18: 20:608510:d=2021032312:SPFH:2 m above ground:24 hour fcst: -18: 21:689405:d=2021032312:DPT:2 m above ground:24 hour fcst: -18: 22:745522:d=2021032312:RH:2 m above ground:24 hour fcst: -18: 23:793007:d=2021032312:APTMP:2 m above ground:24 hour fcst: -18: 24:827739:d=2021032312:TMAX:2 m above ground:18-24 hour max fcst: -18: 25:881917:d=2021032312:TMIN:2 m above ground:18-24 hour min fcst: -18: 26:936202:d=2021032312:UGRD:10 m above ground:24 hour fcst: -18: 27:990775:d=2021032312:VGRD:10 m above ground:24 hour fcst: -18: 28:1046728:d=2021032312:ICEG:10 m above mean sea level:24 hour fcst: -18: 29:1048255:d=2021032312:CPOFP:surface:24 hour fcst: -18: 30:1101792:d=2021032312:CPRAT:surface:24 hour fcst: -18: 31:1162917:d=2021032312:PRATE:surface:24 hour fcst: -18: + product_functions.sh[10]rc=0 -18: + product_functions.sh[11](( rc == 0 )) -18: + product_functions.sh[11]mv pgb2file_f024_19_1p00.new pgb2file_f024_19_1p00 -18: + product_functions.sh[12]return 0 -18: + interp_atmos_master.sh[56]export err=0 -18: + interp_atmos_master.sh[56]err=0 -18: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -18: ++ interp_atmos_master.sh[62]wc -l -18: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_19_1p00 -match 'LAND|ICEC' -18: + interp_atmos_master.sh[62]var_count=0 -18: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -18: + interp_atmos_master.sh[73]exit 0 -13: + bash[8]'[' -z '' ']' -13: + bash[9]case "$-" in -13: + bash[12]__lmod_vx=x -13: + bash[16]'[' -n x ']' -13: + bash[16]set +x -13: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -13: Shell debugging restarted -13: + bash[224]unset __lmod_vx -13: + interp_atmos_master.sh[7]input_file=tmpfile_f024_14 -13: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_14 -13: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -13: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -13: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -13: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -13: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -13: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -13: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -13: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -13: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -13: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -13: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -13: + interp_atmos_master.sh[31]IFS=: -13: + interp_atmos_master.sh[31]read -ra grids -13: + interp_atmos_master.sh[33]output_grids= -13: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -13: + interp_atmos_master.sh[35]gridopt=grid0p25 -13: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_14_0p25' -13: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -13: + interp_atmos_master.sh[35]gridopt=grid0p50 -13: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_14_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_14_0p50' -13: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -13: + interp_atmos_master.sh[35]gridopt=grid1p00 -13: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_14_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_14_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_14_1p00' -13: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_14 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_14_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_14_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_14_1p00 -13: 1:0:d=2021032312:ABSV:650 mb:24 hour fcst: -13: 2:20904:d=2021032312:CLMR:650 mb:24 hour fcst: -13: 3:26126:d=2021032312:ICMR:650 mb:24 hour fcst: -13: 4:38562:d=2021032312:RWMR:650 mb:24 hour fcst: -13: 5:43132:d=2021032312:SNMR:650 mb:24 hour fcst: -13: 6:58199:d=2021032312:GRLE:650 mb:24 hour fcst: -13: 7:60477:d=2021032312:O3MR:650 mb:24 hour fcst: -13: 8:88238:d=2021032312:HGT:700 mb:24 hour fcst: -13: 9:112576:d=2021032312:TMP:700 mb:24 hour fcst: -13: 10:132183:d=2021032312:RH:700 mb:24 hour fcst: -13: 11:152474:d=2021032312:TCDC:700 mb:24 hour fcst: -13: 12:164006:d=2021032312:SPFH:700 mb:24 hour fcst: -13: 13:192618:d=2021032312:VVEL:700 mb:24 hour fcst: -13: 14:218136:d=2021032312:DZDT:700 mb:24 hour fcst: -13: 15:245301:d=2021032312:UGRD:700 mb:24 hour fcst: -13: 16:266136:d=2021032312:VGRD:700 mb:24 hour fcst: -13: 17:287753:d=2021032312:ABSV:700 mb:24 hour fcst: -13: 18:308788:d=2021032312:CLMR:700 mb:24 hour fcst: -13: 19:315737:d=2021032312:ICMR:700 mb:24 hour fcst: -13: 20:326879:d=2021032312:RWMR:700 mb:24 hour fcst: -13: 21:332398:d=2021032312:SNMR:700 mb:24 hour fcst: -13: 22:346626:d=2021032312:GRLE:700 mb:24 hour fcst: -13: 23:348875:d=2021032312:O3MR:700 mb:24 hour fcst: -13: 24:376250:d=2021032312:HGT:750 mb:24 hour fcst: -13: 25:400655:d=2021032312:TMP:750 mb:24 hour fcst: -13: 26:420547:d=2021032312:RH:750 mb:24 hour fcst: -13: 27:440999:d=2021032312:TCDC:750 mb:24 hour fcst: -13: 28:453332:d=2021032312:SPFH:750 mb:24 hour fcst: -13: 29:482460:d=2021032312:VVEL:750 mb:24 hour fcst: -13: 30:508163:d=2021032312:DZDT:750 mb:24 hour fcst: -13: 31:535334:d=2021032312:UGRD:750 mb:24 hour fcst: -13: 32:556178:d=2021032312:VGRD:750 mb:24 hour fcst: -13: + interp_atmos_master.sh[47]export err=0 -13: + interp_atmos_master.sh[47]err=0 -13: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -13: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -13: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_14_0p25 -13: + product_functions.sh[5]local filename=pgb2file_f024_14_0p25 -13: + product_functions.sh[6]wgrib2 pgb2file_f024_14_0p25 -not_if :RH: -grib pgb2file_f024_14_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_14_0p25.new -13: 1:0:d=2021032312:ABSV:650 mb:24 hour fcst: -13: 2:492417:d=2021032312:CLMR:650 mb:24 hour fcst: -13: 3:597382:d=2021032312:ICMR:650 mb:24 hour fcst: -13: 4:871581:d=2021032312:RWMR:650 mb:24 hour fcst: -13: 5:963431:d=2021032312:SNMR:650 mb:24 hour fcst: -13: 6:1290483:d=2021032312:GRLE:650 mb:24 hour fcst: -13: 7:1328558:d=2021032312:O3MR:650 mb:24 hour fcst: -13: 8:2084058:d=2021032312:HGT:700 mb:24 hour fcst: -13: 9:2712727:d=2021032312:TMP:700 mb:24 hour fcst: -13: 10:3163037:d=2021032312:RH:700 mb:24 hour fcst: -13: 11:3637860:d=2021032312:TCDC:700 mb:24 hour fcst: -13: 12:3862174:d=2021032312:SPFH:700 mb:24 hour fcst: -13: 13:4635334:d=2021032312:VVEL:700 mb:24 hour fcst: -13: 14:5318972:d=2021032312:DZDT:700 mb:24 hour fcst: -13: 15:6074135:d=2021032312:UGRD:700 mb:24 hour fcst: -13: 16:6576917:d=2021032312:VGRD:700 mb:24 hour fcst: -13: 17:7109551:d=2021032312:ABSV:700 mb:24 hour fcst: -13: 18:7608064:d=2021032312:CLMR:700 mb:24 hour fcst: -13: 19:7756233:d=2021032312:ICMR:700 mb:24 hour fcst: -13: 20:8010331:d=2021032312:RWMR:700 mb:24 hour fcst: -13: 21:8116666:d=2021032312:SNMR:700 mb:24 hour fcst: -13: 22:8421995:d=2021032312:GRLE:700 mb:24 hour fcst: -13: 23:8460947:d=2021032312:O3MR:700 mb:24 hour fcst: -13: 24:9194178:d=2021032312:HGT:750 mb:24 hour fcst: -13: 25:9824606:d=2021032312:TMP:750 mb:24 hour fcst: -13: 26:10287246:d=2021032312:RH:750 mb:24 hour fcst: -13: 27:10767663:d=2021032312:TCDC:750 mb:24 hour fcst: -13: 28:11007923:d=2021032312:SPFH:750 mb:24 hour fcst: -13: 29:11797353:d=2021032312:VVEL:750 mb:24 hour fcst: -13: 30:12483286:d=2021032312:DZDT:750 mb:24 hour fcst: -13: 31:13233709:d=2021032312:UGRD:750 mb:24 hour fcst: -13: 32:13738116:d=2021032312:VGRD:750 mb:24 hour fcst: -13: + product_functions.sh[10]rc=0 -13: + product_functions.sh[11](( rc == 0 )) -13: + product_functions.sh[11]mv pgb2file_f024_14_0p25.new pgb2file_f024_14_0p25 -13: + product_functions.sh[12]return 0 -13: + interp_atmos_master.sh[56]export err=0 -13: + interp_atmos_master.sh[56]err=0 -13: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -13: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_14_0p25 -match 'LAND|ICEC' -13: ++ interp_atmos_master.sh[62]wc -l -13: + interp_atmos_master.sh[62]var_count=0 -13: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -13: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -13: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_14_0p50 -13: + product_functions.sh[5]local filename=pgb2file_f024_14_0p50 -13: + product_functions.sh[6]wgrib2 pgb2file_f024_14_0p50 -not_if :RH: -grib pgb2file_f024_14_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_14_0p50.new -13: 1:0:d=2021032312:ABSV:650 mb:24 hour fcst: -13: 2:172387:d=2021032312:CLMR:650 mb:24 hour fcst: -13: 3:212072:d=2021032312:ICMR:650 mb:24 hour fcst: -13: 4:310191:d=2021032312:RWMR:650 mb:24 hour fcst: -13: 5:344236:d=2021032312:SNMR:650 mb:24 hour fcst: -13: 6:461120:d=2021032312:GRLE:650 mb:24 hour fcst: -13: 7:475503:d=2021032312:O3MR:650 mb:24 hour fcst: -13: 8:731428:d=2021032312:HGT:700 mb:24 hour fcst: -13: 9:943448:d=2021032312:TMP:700 mb:24 hour fcst: -13: 10:1097633:d=2021032312:RH:700 mb:24 hour fcst: -13: 11:1261426:d=2021032312:TCDC:700 mb:24 hour fcst: -13: 12:1346139:d=2021032312:SPFH:700 mb:24 hour fcst: -13: 13:1614532:d=2021032312:VVEL:700 mb:24 hour fcst: -13: 14:1848047:d=2021032312:DZDT:700 mb:24 hour fcst: -13: 15:2097451:d=2021032312:UGRD:700 mb:24 hour fcst: -13: 16:2270448:d=2021032312:VGRD:700 mb:24 hour fcst: -13: 17:2452507:d=2021032312:ABSV:700 mb:24 hour fcst: -13: 18:2624813:d=2021032312:CLMR:700 mb:24 hour fcst: -13: 19:2680392:d=2021032312:ICMR:700 mb:24 hour fcst: -13: 20:2769814:d=2021032312:RWMR:700 mb:24 hour fcst: -13: 21:2810129:d=2021032312:SNMR:700 mb:24 hour fcst: -13: 22:2920523:d=2021032312:GRLE:700 mb:24 hour fcst: -13: 23:2935420:d=2021032312:O3MR:700 mb:24 hour fcst: -13: 24:3189052:d=2021032312:HGT:750 mb:24 hour fcst: -13: 25:3401690:d=2021032312:TMP:750 mb:24 hour fcst: -13: 26:3561543:d=2021032312:RH:750 mb:24 hour fcst: -13: 27:3727116:d=2021032312:TCDC:750 mb:24 hour fcst: -13: 28:3817632:d=2021032312:SPFH:750 mb:24 hour fcst: -13: 29:4093488:d=2021032312:VVEL:750 mb:24 hour fcst: -13: 30:4327652:d=2021032312:DZDT:750 mb:24 hour fcst: -13: 31:4575562:d=2021032312:UGRD:750 mb:24 hour fcst: -13: 32:4749044:d=2021032312:VGRD:750 mb:24 hour fcst: -13: + product_functions.sh[10]rc=0 -13: + product_functions.sh[11](( rc == 0 )) -13: + product_functions.sh[11]mv pgb2file_f024_14_0p50.new pgb2file_f024_14_0p50 -13: + product_functions.sh[12]return 0 -13: + interp_atmos_master.sh[56]export err=0 -13: + interp_atmos_master.sh[56]err=0 -13: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -13: ++ interp_atmos_master.sh[62]wc -l -13: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_14_0p50 -match 'LAND|ICEC' -13: + interp_atmos_master.sh[62]var_count=0 -13: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -13: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -13: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_14_1p00 -13: + product_functions.sh[5]local filename=pgb2file_f024_14_1p00 -13: + product_functions.sh[6]wgrib2 pgb2file_f024_14_1p00 -not_if :RH: -grib pgb2file_f024_14_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_14_1p00.new -13: 1:0:d=2021032312:ABSV:650 mb:24 hour fcst: -13: 2:55092:d=2021032312:CLMR:650 mb:24 hour fcst: -13: 3:69364:d=2021032312:ICMR:650 mb:24 hour fcst: -13: 4:102148:d=2021032312:RWMR:650 mb:24 hour fcst: -13: 5:113891:d=2021032312:SNMR:650 mb:24 hour fcst: -13: 6:153252:d=2021032312:GRLE:650 mb:24 hour fcst: -13: 7:158836:d=2021032312:O3MR:650 mb:24 hour fcst: -13: 8:236663:d=2021032312:HGT:700 mb:24 hour fcst: -13: 9:302864:d=2021032312:TMP:700 mb:24 hour fcst: -13: 10:353861:d=2021032312:RH:700 mb:24 hour fcst: -13: 11:407300:d=2021032312:TCDC:700 mb:24 hour fcst: -13: 12:437469:d=2021032312:SPFH:700 mb:24 hour fcst: -13: 13:517846:d=2021032312:VVEL:700 mb:24 hour fcst: -13: 14:589307:d=2021032312:DZDT:700 mb:24 hour fcst: -13: 15:664370:d=2021032312:UGRD:700 mb:24 hour fcst: -13: 16:720130:d=2021032312:VGRD:700 mb:24 hour fcst: -13: 17:778234:d=2021032312:ABSV:700 mb:24 hour fcst: -13: 18:833639:d=2021032312:CLMR:700 mb:24 hour fcst: -13: 19:852926:d=2021032312:ICMR:700 mb:24 hour fcst: -13: 20:882399:d=2021032312:RWMR:700 mb:24 hour fcst: -13: 21:896596:d=2021032312:SNMR:700 mb:24 hour fcst: -13: 22:933811:d=2021032312:GRLE:700 mb:24 hour fcst: -13: 23:939602:d=2021032312:O3MR:700 mb:24 hour fcst: -13: 24:1016505:d=2021032312:HGT:750 mb:24 hour fcst: -13: 25:1082903:d=2021032312:TMP:750 mb:24 hour fcst: -13: 26:1135055:d=2021032312:RH:750 mb:24 hour fcst: -13: 27:1188837:d=2021032312:TCDC:750 mb:24 hour fcst: -13: 28:1220999:d=2021032312:SPFH:750 mb:24 hour fcst: -13: 29:1303097:d=2021032312:VVEL:750 mb:24 hour fcst: -13: 30:1374733:d=2021032312:DZDT:750 mb:24 hour fcst: -13: 31:1449320:d=2021032312:UGRD:750 mb:24 hour fcst: -13: 32:1505194:d=2021032312:VGRD:750 mb:24 hour fcst: -13: + product_functions.sh[10]rc=0 -13: + product_functions.sh[11](( rc == 0 )) -13: + product_functions.sh[11]mv pgb2file_f024_14_1p00.new pgb2file_f024_14_1p00 -13: + product_functions.sh[12]return 0 -13: + interp_atmos_master.sh[56]export err=0 -13: + interp_atmos_master.sh[56]err=0 -13: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -13: ++ interp_atmos_master.sh[62]wc -l -13: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_14_1p00 -match 'LAND|ICEC' -13: + interp_atmos_master.sh[62]var_count=0 -13: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -13: + interp_atmos_master.sh[73]exit 0 -10: + bash[8]'[' -z '' ']' -10: + bash[9]case "$-" in -10: + bash[12]__lmod_vx=x -10: + bash[16]'[' -n x ']' -10: + bash[16]set +x -10: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -10: Shell debugging restarted -10: + bash[224]unset __lmod_vx -10: + interp_atmos_master.sh[7]input_file=tmpfile_f024_11 -10: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_11 -10: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -10: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -10: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -10: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -10: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -10: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -10: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -10: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -10: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -10: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -10: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -10: + interp_atmos_master.sh[31]IFS=: -10: + interp_atmos_master.sh[31]read -ra grids -10: + interp_atmos_master.sh[33]output_grids= -10: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -10: + interp_atmos_master.sh[35]gridopt=grid0p25 -10: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_11_0p25' -10: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -10: + interp_atmos_master.sh[35]gridopt=grid0p50 -10: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_11_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_11_0p50' -10: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -10: + interp_atmos_master.sh[35]gridopt=grid1p00 -10: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_11_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_11_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_11_1p00' -10: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_11 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_11_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_11_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_11_1p00 -10: 1:0:d=2021032312:ICMR:350 mb:24 hour fcst: -10: 2:16814:d=2021032312:RWMR:350 mb:24 hour fcst: -10: 3:16993:d=2021032312:SNMR:350 mb:24 hour fcst: -10: 4:30272:d=2021032312:GRLE:350 mb:24 hour fcst: -10: 5:30779:d=2021032312:O3MR:350 mb:24 hour fcst: -10: 6:56355:d=2021032312:HGT:400 mb:24 hour fcst: -10: 7:78108:d=2021032312:TMP:400 mb:24 hour fcst: -10: 8:96292:d=2021032312:RH:400 mb:24 hour fcst: -10: 9:115493:d=2021032312:TCDC:400 mb:24 hour fcst: -10: 10:128610:d=2021032312:SPFH:400 mb:24 hour fcst: -10: 11:156478:d=2021032312:VVEL:400 mb:24 hour fcst: -10: 12:181434:d=2021032312:DZDT:400 mb:24 hour fcst: -10: 13:209827:d=2021032312:UGRD:400 mb:24 hour fcst: -10: 14:223902:d=2021032312:VGRD:400 mb:24 hour fcst: -10: 15:246612:d=2021032312:ABSV:400 mb:24 hour fcst: -10: 16:268574:d=2021032312:CLMR:400 mb:24 hour fcst: -10: 17:268948:d=2021032312:ICMR:400 mb:24 hour fcst: -10: 18:285933:d=2021032312:RWMR:400 mb:24 hour fcst: -10: 19:286221:d=2021032312:SNMR:400 mb:24 hour fcst: -10: 20:300144:d=2021032312:GRLE:400 mb:24 hour fcst: -10: 21:300956:d=2021032312:O3MR:400 mb:24 hour fcst: -10: 22:325540:d=2021032312:HGT:450 mb:24 hour fcst: -10: 23:349502:d=2021032312:TMP:450 mb:24 hour fcst: -10: 24:367878:d=2021032312:RH:450 mb:24 hour fcst: -10: 25:387288:d=2021032312:TCDC:450 mb:24 hour fcst: -10: 26:400015:d=2021032312:SPFH:450 mb:24 hour fcst: -10: 27:429492:d=2021032312:VVEL:450 mb:24 hour fcst: -10: 28:454650:d=2021032312:DZDT:450 mb:24 hour fcst: -10: 29:482799:d=2021032312:UGRD:450 mb:24 hour fcst: -10: 30:504483:d=2021032312:VGRD:450 mb:24 hour fcst: -10: 31:526811:d=2021032312:ABSV:450 mb:24 hour fcst: -10: + interp_atmos_master.sh[47]export err=0 -10: + interp_atmos_master.sh[47]err=0 -10: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -10: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -10: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_11_0p25 -10: + product_functions.sh[5]local filename=pgb2file_f024_11_0p25 -10: + product_functions.sh[6]wgrib2 pgb2file_f024_11_0p25 -not_if :RH: -grib pgb2file_f024_11_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_11_0p25.new -10: 1:0:d=2021032312:ICMR:350 mb:24 hour fcst: -10: 2:371516:d=2021032312:RWMR:350 mb:24 hour fcst: -10: 3:371695:d=2021032312:SNMR:350 mb:24 hour fcst: -10: 4:647476:d=2021032312:GRLE:350 mb:24 hour fcst: -10: 5:655151:d=2021032312:O3MR:350 mb:24 hour fcst: -10: 6:1325401:d=2021032312:HGT:400 mb:24 hour fcst: -10: 7:1861887:d=2021032312:TMP:400 mb:24 hour fcst: -10: 8:2272555:d=2021032312:RH:400 mb:24 hour fcst: -10: 9:2709435:d=2021032312:TCDC:400 mb:24 hour fcst: -10: 10:2948503:d=2021032312:SPFH:400 mb:24 hour fcst: -10: 11:3690322:d=2021032312:VVEL:400 mb:24 hour fcst: -10: 12:4352417:d=2021032312:DZDT:400 mb:24 hour fcst: -10: 13:5145120:d=2021032312:UGRD:400 mb:24 hour fcst: -10: 14:5692834:d=2021032312:VGRD:400 mb:24 hour fcst: -10: 15:6267438:d=2021032312:ABSV:400 mb:24 hour fcst: -10: 16:6803915:d=2021032312:CLMR:400 mb:24 hour fcst: -10: 17:6807766:d=2021032312:ICMR:400 mb:24 hour fcst: -10: 18:7183284:d=2021032312:RWMR:400 mb:24 hour fcst: -10: 19:7184765:d=2021032312:SNMR:400 mb:24 hour fcst: -10: 20:7477961:d=2021032312:GRLE:400 mb:24 hour fcst: -10: 21:7486689:d=2021032312:O3MR:400 mb:24 hour fcst: -10: 22:8121067:d=2021032312:HGT:450 mb:24 hour fcst: -10: 23:8743997:d=2021032312:TMP:450 mb:24 hour fcst: -10: 24:9163548:d=2021032312:RH:450 mb:24 hour fcst: -10: 25:9604602:d=2021032312:TCDC:450 mb:24 hour fcst: -10: 26:9841454:d=2021032312:SPFH:450 mb:24 hour fcst: -10: 27:10639351:d=2021032312:VVEL:450 mb:24 hour fcst: -10: 28:11306365:d=2021032312:DZDT:450 mb:24 hour fcst: -10: 29:12092876:d=2021032312:UGRD:450 mb:24 hour fcst: -10: 30:12626256:d=2021032312:VGRD:450 mb:24 hour fcst: -10: 31:13190220:d=2021032312:ABSV:450 mb:24 hour fcst: -10: + product_functions.sh[10]rc=0 -10: + product_functions.sh[11](( rc == 0 )) -10: + product_functions.sh[11]mv pgb2file_f024_11_0p25.new pgb2file_f024_11_0p25 -10: + product_functions.sh[12]return 0 -10: + interp_atmos_master.sh[56]export err=0 -10: + interp_atmos_master.sh[56]err=0 -10: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -10: ++ interp_atmos_master.sh[62]wc -l -10: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_11_0p25 -match 'LAND|ICEC' -10: + interp_atmos_master.sh[62]var_count=0 -10: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -10: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -10: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_11_0p50 -10: + product_functions.sh[5]local filename=pgb2file_f024_11_0p50 -10: + product_functions.sh[6]wgrib2 pgb2file_f024_11_0p50 -not_if :RH: -grib pgb2file_f024_11_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_11_0p50.new -10: 1:0:d=2021032312:ICMR:350 mb:24 hour fcst: -10: 2:135713:d=2021032312:RWMR:350 mb:24 hour fcst: -10: 3:135892:d=2021032312:SNMR:350 mb:24 hour fcst: -10: 4:235460:d=2021032312:GRLE:350 mb:24 hour fcst: -10: 5:238470:d=2021032312:O3MR:350 mb:24 hour fcst: -10: 6:465630:d=2021032312:HGT:400 mb:24 hour fcst: -10: 7:647340:d=2021032312:TMP:400 mb:24 hour fcst: -10: 8:788803:d=2021032312:RH:400 mb:24 hour fcst: -10: 9:941471:d=2021032312:TCDC:400 mb:24 hour fcst: -10: 10:1034050:d=2021032312:SPFH:400 mb:24 hour fcst: -10: 11:1292467:d=2021032312:VVEL:400 mb:24 hour fcst: -10: 12:1518009:d=2021032312:DZDT:400 mb:24 hour fcst: -10: 13:1776328:d=2021032312:UGRD:400 mb:24 hour fcst: -10: 14:1965173:d=2021032312:VGRD:400 mb:24 hour fcst: -10: 15:2160721:d=2021032312:ABSV:400 mb:24 hour fcst: -10: 16:2344980:d=2021032312:CLMR:400 mb:24 hour fcst: -10: 17:2346571:d=2021032312:ICMR:400 mb:24 hour fcst: -10: 18:2483279:d=2021032312:RWMR:400 mb:24 hour fcst: -10: 19:2484011:d=2021032312:SNMR:400 mb:24 hour fcst: -10: 20:2588462:d=2021032312:GRLE:400 mb:24 hour fcst: -10: 21:2592053:d=2021032312:O3MR:400 mb:24 hour fcst: -10: 22:2808141:d=2021032312:HGT:450 mb:24 hour fcst: -10: 23:3017504:d=2021032312:TMP:450 mb:24 hour fcst: -10: 24:3161378:d=2021032312:RH:450 mb:24 hour fcst: -10: 25:3315511:d=2021032312:TCDC:450 mb:24 hour fcst: -10: 26:3406296:d=2021032312:SPFH:450 mb:24 hour fcst: -10: 27:3684512:d=2021032312:VVEL:450 mb:24 hour fcst: -10: 28:3911833:d=2021032312:DZDT:450 mb:24 hour fcst: -10: 29:4168516:d=2021032312:UGRD:450 mb:24 hour fcst: -10: 30:4351401:d=2021032312:VGRD:450 mb:24 hour fcst: -10: 31:4543263:d=2021032312:ABSV:450 mb:24 hour fcst: -10: + product_functions.sh[10]rc=0 -10: + product_functions.sh[11](( rc == 0 )) -10: + product_functions.sh[11]mv pgb2file_f024_11_0p50.new pgb2file_f024_11_0p50 -10: + product_functions.sh[12]return 0 -10: + interp_atmos_master.sh[56]export err=0 -10: + interp_atmos_master.sh[56]err=0 -10: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -10: ++ interp_atmos_master.sh[62]wc -l -10: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_11_0p50 -match 'LAND|ICEC' -10: + interp_atmos_master.sh[62]var_count=0 -10: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -10: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -10: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_11_1p00 -10: + product_functions.sh[5]local filename=pgb2file_f024_11_1p00 -10: + product_functions.sh[6]wgrib2 pgb2file_f024_11_1p00 -not_if :RH: -grib pgb2file_f024_11_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_11_1p00.new -10: 1:0:d=2021032312:ICMR:350 mb:24 hour fcst: -10: 2:44444:d=2021032312:RWMR:350 mb:24 hour fcst: -10: 3:44623:d=2021032312:SNMR:350 mb:24 hour fcst: -10: 4:78417:d=2021032312:GRLE:350 mb:24 hour fcst: -10: 5:79622:d=2021032312:O3MR:350 mb:24 hour fcst: -10: 6:149543:d=2021032312:HGT:400 mb:24 hour fcst: -10: 7:207890:d=2021032312:TMP:400 mb:24 hour fcst: -10: 8:254897:d=2021032312:RH:400 mb:24 hour fcst: -10: 9:304923:d=2021032312:TCDC:400 mb:24 hour fcst: -10: 10:338326:d=2021032312:SPFH:400 mb:24 hour fcst: -10: 11:415784:d=2021032312:VVEL:400 mb:24 hour fcst: -10: 12:484770:d=2021032312:DZDT:400 mb:24 hour fcst: -10: 13:562365:d=2021032312:UGRD:400 mb:24 hour fcst: -10: 14:622291:d=2021032312:VGRD:400 mb:24 hour fcst: -10: 15:684122:d=2021032312:ABSV:400 mb:24 hour fcst: -10: 16:742589:d=2021032312:CLMR:400 mb:24 hour fcst: -10: 17:743341:d=2021032312:ICMR:400 mb:24 hour fcst: -10: 18:788604:d=2021032312:RWMR:400 mb:24 hour fcst: -10: 19:789014:d=2021032312:SNMR:400 mb:24 hour fcst: -10: 20:824455:d=2021032312:GRLE:400 mb:24 hour fcst: -10: 21:826229:d=2021032312:O3MR:400 mb:24 hour fcst: -10: 22:893290:d=2021032312:HGT:450 mb:24 hour fcst: -10: 23:958971:d=2021032312:TMP:450 mb:24 hour fcst: -10: 24:1006156:d=2021032312:RH:450 mb:24 hour fcst: -10: 25:1056651:d=2021032312:TCDC:450 mb:24 hour fcst: -10: 26:1089038:d=2021032312:SPFH:450 mb:24 hour fcst: -10: 27:1171526:d=2021032312:VVEL:450 mb:24 hour fcst: -10: 28:1241039:d=2021032312:DZDT:450 mb:24 hour fcst: -10: 29:1318026:d=2021032312:UGRD:450 mb:24 hour fcst: -10: 30:1376209:d=2021032312:VGRD:450 mb:24 hour fcst: -10: 31:1436663:d=2021032312:ABSV:450 mb:24 hour fcst: -10: + product_functions.sh[10]rc=0 -10: + product_functions.sh[11](( rc == 0 )) -10: + product_functions.sh[11]mv pgb2file_f024_11_1p00.new pgb2file_f024_11_1p00 -10: + product_functions.sh[12]return 0 -10: + interp_atmos_master.sh[56]export err=0 -10: + interp_atmos_master.sh[56]err=0 -10: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -10: ++ interp_atmos_master.sh[62]wc -l -10: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_11_1p00 -match 'LAND|ICEC' -10: + interp_atmos_master.sh[62]var_count=0 -10: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -10: + interp_atmos_master.sh[73]exit 0 -20: + bash[8]'[' -z '' ']' -20: + bash[9]case "$-" in -20: + bash[12]__lmod_vx=x -20: + bash[16]'[' -n x ']' -20: + bash[16]set +x -20: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -20: Shell debugging restarted -20: + bash[224]unset __lmod_vx -20: + interp_atmos_master.sh[7]input_file=tmpfile_f024_21 -20: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_21 -20: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -20: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -20: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -20: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -20: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -20: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -20: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -20: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -20: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -20: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -20: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -20: + interp_atmos_master.sh[31]IFS=: -20: + interp_atmos_master.sh[31]read -ra grids -20: + interp_atmos_master.sh[33]output_grids= -20: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -20: + interp_atmos_master.sh[35]gridopt=grid0p25 -20: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_21_0p25' -20: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -20: + interp_atmos_master.sh[35]gridopt=grid0p50 -20: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_21_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_21_0p50' -20: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -20: + interp_atmos_master.sh[35]gridopt=grid1p00 -20: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_21_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_21_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_21_1p00' -20: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_21 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_21_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_21_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_21_1p00 -20: 1:0:d=2021032312:LFTX:surface:24 hour fcst: -20: 2:13794:d=2021032312:CAPE:surface:24 hour fcst: -20: 3:28886:d=2021032312:CIN:surface:24 hour fcst: -20: 4:41872:d=2021032312:PWAT:entire atmosphere (considered as a single layer):24 hour fcst: -20: 5:68773:d=2021032312:CWAT:entire atmosphere (considered as a single layer):24 hour fcst: -20: 6:95238:d=2021032312:RH:entire atmosphere (considered as a single layer):24 hour fcst: -20: 7:110381:d=2021032312:TOZNE:entire atmosphere (considered as a single layer):24 hour fcst: -20: 8:124732:d=2021032312:LCDC:low cloud layer:24 hour fcst: -20: 9:142441:d=2021032312:LCDC:low cloud layer:18-24 hour ave fcst: -20: 10:162395:d=2021032312:MCDC:middle cloud layer:24 hour fcst: -20: 11:178231:d=2021032312:MCDC:middle cloud layer:18-24 hour ave fcst: -20: 12:195499:d=2021032312:HCDC:high cloud layer:24 hour fcst: -20: 13:210558:d=2021032312:HCDC:high cloud layer:18-24 hour ave fcst: -20: 14:228657:d=2021032312:TCDC:entire atmosphere (considered as a single layer):24 hour fcst: -20: 15:246160:d=2021032312:TCDC:entire atmosphere (considered as a single layer):18-24 hour ave fcst: -20: 16:266243:d=2021032312:HGT:cloud ceiling:24 hour fcst: -20: 17:298761:d=2021032312:PRES:convective cloud bottom level:24 hour fcst: -20: 18:312415:d=2021032312:PRES:low cloud bottom level:18-24 hour ave fcst: -20: 19:335255:d=2021032312:PRES:middle cloud bottom level:18-24 hour ave fcst: -20: 20:352982:d=2021032312:PRES:high cloud bottom level:18-24 hour ave fcst: -20: 21:371874:d=2021032312:PRES:convective cloud top level:24 hour fcst: -20: 22:387209:d=2021032312:PRES:low cloud top level:18-24 hour ave fcst: -20: 23:410223:d=2021032312:PRES:middle cloud top level:18-24 hour ave fcst: -20: 24:427934:d=2021032312:PRES:high cloud top level:18-24 hour ave fcst: -20: 25:447156:d=2021032312:TMP:low cloud top level:18-24 hour ave fcst: -20: 26:464312:d=2021032312:TMP:middle cloud top level:18-24 hour ave fcst: -20: 27:478358:d=2021032312:TMP:high cloud top level:18-24 hour ave fcst: -20: 28:494319:d=2021032312:TCDC:convective cloud layer:24 hour fcst: -20: 29:515137:d=2021032312:TCDC:boundary layer cloud layer:18-24 hour ave fcst: -20: 30:533052:d=2021032312:CWORK:entire atmosphere (considered as a single layer):18-24 hour ave fcst: -20: 31:546751:d=2021032312:DSWRF:surface:18-24 hour ave fcst: -20: + interp_atmos_master.sh[47]export err=0 -20: + interp_atmos_master.sh[47]err=0 -20: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -20: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -20: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_21_0p25 -20: + product_functions.sh[5]local filename=pgb2file_f024_21_0p25 -20: + product_functions.sh[6]wgrib2 pgb2file_f024_21_0p25 -not_if :RH: -grib pgb2file_f024_21_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_21_0p25.new -20: 1:0:d=2021032312:LFTX:surface:24 hour fcst: -20: 2:298611:d=2021032312:CAPE:surface:24 hour fcst: -20: 3:630696:d=2021032312:CIN:surface:24 hour fcst: -20: 4:885541:d=2021032312:PWAT:entire atmosphere (considered as a single layer):24 hour fcst: -20: 5:1616047:d=2021032312:CWAT:entire atmosphere (considered as a single layer):24 hour fcst: -20: 6:2278995:d=2021032312:RH:entire atmosphere (considered as a single layer):24 hour fcst: -20: 7:2615906:d=2021032312:TOZNE:entire atmosphere (considered as a single layer):24 hour fcst: -20: 8:2924577:d=2021032312:LCDC:low cloud layer:24 hour fcst: -20: 9:3263448:d=2021032312:LCDC:low cloud layer:18-24 hour ave fcst: -20: 10:3679387:d=2021032312:MCDC:middle cloud layer:24 hour fcst: -20: 11:3948708:d=2021032312:MCDC:middle cloud layer:18-24 hour ave fcst: -20: 12:4291366:d=2021032312:HCDC:high cloud layer:24 hour fcst: -20: 13:4539129:d=2021032312:HCDC:high cloud layer:18-24 hour ave fcst: -20: 14:4884994:d=2021032312:TCDC:entire atmosphere (considered as a single layer):24 hour fcst: -20: 15:5202112:d=2021032312:TCDC:entire atmosphere (considered as a single layer):18-24 hour ave fcst: -20: 16:5621103:d=2021032312:HGT:cloud ceiling:24 hour fcst: -20: 17:6298574:d=2021032312:PRES:convective cloud bottom level:24 hour fcst: -20: 18:6796352:d=2021032312:PRES:low cloud bottom level:18-24 hour ave fcst: -20: 19:7528999:d=2021032312:PRES:middle cloud bottom level:18-24 hour ave fcst: -20: 20:8088466:d=2021032312:PRES:high cloud bottom level:18-24 hour ave fcst: -20: 21:8690777:d=2021032312:PRES:convective cloud top level:24 hour fcst: -20: 22:9234757:d=2021032312:PRES:low cloud top level:18-24 hour ave fcst: -20: 23:9953040:d=2021032312:PRES:middle cloud top level:18-24 hour ave fcst: -20: 24:10483606:d=2021032312:PRES:high cloud top level:18-24 hour ave fcst: -20: 25:11109030:d=2021032312:TMP:low cloud top level:18-24 hour ave fcst: -20: 26:11647240:d=2021032312:TMP:middle cloud top level:18-24 hour ave fcst: -20: 27:12077767:d=2021032312:TMP:high cloud top level:18-24 hour ave fcst: -20: 28:12591750:d=2021032312:TCDC:convective cloud layer:24 hour fcst: -20: 29:13066816:d=2021032312:TCDC:boundary layer cloud layer:18-24 hour ave fcst: -20: 30:13419032:d=2021032312:CWORK:entire atmosphere (considered as a single layer):18-24 hour ave fcst: -20: 31:13740771:d=2021032312:DSWRF:surface:18-24 hour ave fcst: -20: + product_functions.sh[10]rc=0 -20: + product_functions.sh[11](( rc == 0 )) -20: + product_functions.sh[11]mv pgb2file_f024_21_0p25.new pgb2file_f024_21_0p25 -20: + product_functions.sh[12]return 0 -20: + interp_atmos_master.sh[56]export err=0 -20: + interp_atmos_master.sh[56]err=0 -20: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -20: ++ interp_atmos_master.sh[62]wc -l -20: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_21_0p25 -match 'LAND|ICEC' -20: + interp_atmos_master.sh[62]var_count=0 -20: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -20: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -20: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_21_0p50 -20: + product_functions.sh[5]local filename=pgb2file_f024_21_0p50 -20: + product_functions.sh[6]wgrib2 pgb2file_f024_21_0p50 -not_if :RH: -grib pgb2file_f024_21_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_21_0p50.new -20: 1:0:d=2021032312:LFTX:surface:24 hour fcst: -20: 2:95061:d=2021032312:CAPE:surface:24 hour fcst: -20: 3:210646:d=2021032312:CIN:surface:24 hour fcst: -20: 4:301305:d=2021032312:PWAT:entire atmosphere (considered as a single layer):24 hour fcst: -20: 5:552524:d=2021032312:CWAT:entire atmosphere (considered as a single layer):24 hour fcst: -20: 6:784070:d=2021032312:RH:entire atmosphere (considered as a single layer):24 hour fcst: -20: 7:896738:d=2021032312:TOZNE:entire atmosphere (considered as a single layer):24 hour fcst: -20: 8:998199:d=2021032312:LCDC:low cloud layer:24 hour fcst: -20: 9:1132847:d=2021032312:LCDC:low cloud layer:18-24 hour ave fcst: -20: 10:1288909:d=2021032312:MCDC:middle cloud layer:24 hour fcst: -20: 11:1397743:d=2021032312:MCDC:middle cloud layer:18-24 hour ave fcst: -20: 12:1525507:d=2021032312:HCDC:high cloud layer:24 hour fcst: -20: 13:1625712:d=2021032312:HCDC:high cloud layer:18-24 hour ave fcst: -20: 14:1758653:d=2021032312:TCDC:entire atmosphere (considered as a single layer):24 hour fcst: -20: 15:1886341:d=2021032312:TCDC:entire atmosphere (considered as a single layer):18-24 hour ave fcst: -20: 16:2043572:d=2021032312:HGT:cloud ceiling:24 hour fcst: -20: 17:2333199:d=2021032312:PRES:convective cloud bottom level:24 hour fcst: -20: 18:2484722:d=2021032312:PRES:low cloud bottom level:18-24 hour ave fcst: -20: 19:2726704:d=2021032312:PRES:middle cloud bottom level:18-24 hour ave fcst: -20: 20:2910127:d=2021032312:PRES:high cloud bottom level:18-24 hour ave fcst: -20: 21:3107673:d=2021032312:PRES:convective cloud top level:24 hour fcst: -20: 22:3279310:d=2021032312:PRES:low cloud top level:18-24 hour ave fcst: -20: 23:3519202:d=2021032312:PRES:middle cloud top level:18-24 hour ave fcst: -20: 24:3694790:d=2021032312:PRES:high cloud top level:18-24 hour ave fcst: -20: 25:3897120:d=2021032312:TMP:low cloud top level:18-24 hour ave fcst: -20: 26:4068232:d=2021032312:TMP:middle cloud top level:18-24 hour ave fcst: -20: 27:4206399:d=2021032312:TMP:high cloud top level:18-24 hour ave fcst: -20: 28:4369789:d=2021032312:TCDC:convective cloud layer:24 hour fcst: -20: 29:4554662:d=2021032312:TCDC:boundary layer cloud layer:18-24 hour ave fcst: -20: 30:4685844:d=2021032312:CWORK:entire atmosphere (considered as a single layer):18-24 hour ave fcst: -20: 31:4800122:d=2021032312:DSWRF:surface:18-24 hour ave fcst: -20: + product_functions.sh[10]rc=0 -20: + product_functions.sh[11](( rc == 0 )) -20: + product_functions.sh[11]mv pgb2file_f024_21_0p50.new pgb2file_f024_21_0p50 -20: + product_functions.sh[12]return 0 -20: + interp_atmos_master.sh[56]export err=0 -20: + interp_atmos_master.sh[56]err=0 -20: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -20: ++ interp_atmos_master.sh[62]wc -l -20: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_21_0p50 -match 'LAND|ICEC' -20: + interp_atmos_master.sh[62]var_count=0 -20: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -20: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -20: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_21_1p00 -20: + product_functions.sh[5]local filename=pgb2file_f024_21_1p00 -20: + product_functions.sh[6]wgrib2 pgb2file_f024_21_1p00 -not_if :RH: -grib pgb2file_f024_21_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_21_1p00.new -20: 1:0:d=2021032312:LFTX:surface:24 hour fcst: -20: 2:32865:d=2021032312:CAPE:surface:24 hour fcst: -20: 3:70993:d=2021032312:CIN:surface:24 hour fcst: -20: 4:102961:d=2021032312:PWAT:entire atmosphere (considered as a single layer):24 hour fcst: -20: 5:178517:d=2021032312:CWAT:entire atmosphere (considered as a single layer):24 hour fcst: -20: 6:250542:d=2021032312:RH:entire atmosphere (considered as a single layer):24 hour fcst: -20: 7:288367:d=2021032312:TOZNE:entire atmosphere (considered as a single layer):24 hour fcst: -20: 8:323098:d=2021032312:LCDC:low cloud layer:24 hour fcst: -20: 9:369878:d=2021032312:LCDC:low cloud layer:18-24 hour ave fcst: -20: 10:422490:d=2021032312:MCDC:middle cloud layer:24 hour fcst: -20: 11:461995:d=2021032312:MCDC:middle cloud layer:18-24 hour ave fcst: -20: 12:505902:d=2021032312:HCDC:high cloud layer:24 hour fcst: -20: 13:542649:d=2021032312:HCDC:high cloud layer:18-24 hour ave fcst: -20: 14:588619:d=2021032312:TCDC:entire atmosphere (considered as a single layer):24 hour fcst: -20: 15:633262:d=2021032312:TCDC:entire atmosphere (considered as a single layer):18-24 hour ave fcst: -20: 16:685644:d=2021032312:HGT:cloud ceiling:24 hour fcst: -20: 17:776077:d=2021032312:PRES:convective cloud bottom level:24 hour fcst: -20: 18:818594:d=2021032312:PRES:low cloud bottom level:18-24 hour ave fcst: -20: 19:887829:d=2021032312:PRES:middle cloud bottom level:18-24 hour ave fcst: -20: 20:941528:d=2021032312:PRES:high cloud bottom level:18-24 hour ave fcst: -20: 21:998805:d=2021032312:PRES:convective cloud top level:24 hour fcst: -20: 22:1047065:d=2021032312:PRES:low cloud top level:18-24 hour ave fcst: -20: 23:1116490:d=2021032312:PRES:middle cloud top level:18-24 hour ave fcst: -20: 24:1169030:d=2021032312:PRES:high cloud top level:18-24 hour ave fcst: -20: 25:1227149:d=2021032312:TMP:low cloud top level:18-24 hour ave fcst: -20: 26:1278018:d=2021032312:TMP:middle cloud top level:18-24 hour ave fcst: -20: 27:1319757:d=2021032312:TMP:high cloud top level:18-24 hour ave fcst: -20: 28:1367666:d=2021032312:TCDC:convective cloud layer:24 hour fcst: -20: 29:1425104:d=2021032312:TCDC:boundary layer cloud layer:18-24 hour ave fcst: -20: 30:1471202:d=2021032312:CWORK:entire atmosphere (considered as a single layer):18-24 hour ave fcst: -20: 31:1508146:d=2021032312:DSWRF:surface:18-24 hour ave fcst: -20: + product_functions.sh[10]rc=0 -20: + product_functions.sh[11](( rc == 0 )) -20: + product_functions.sh[11]mv pgb2file_f024_21_1p00.new pgb2file_f024_21_1p00 -20: + product_functions.sh[12]return 0 -20: + interp_atmos_master.sh[56]export err=0 -20: + interp_atmos_master.sh[56]err=0 -20: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -20: ++ interp_atmos_master.sh[62]wc -l -20: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_21_1p00 -match 'LAND|ICEC' -20: + interp_atmos_master.sh[62]var_count=0 -20: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -20: + interp_atmos_master.sh[73]exit 0 -14: + bash[8]'[' -z '' ']' -14: + bash[9]case "$-" in -14: + bash[12]__lmod_vx=x -14: + bash[16]'[' -n x ']' -14: + bash[16]set +x -14: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -14: Shell debugging restarted -14: + bash[224]unset __lmod_vx -14: + interp_atmos_master.sh[7]input_file=tmpfile_f024_15 -14: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_15 -14: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -14: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -14: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -14: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -14: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -14: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -14: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -14: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -14: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -14: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -14: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -14: + interp_atmos_master.sh[31]IFS=: -14: + interp_atmos_master.sh[31]read -ra grids -14: + interp_atmos_master.sh[33]output_grids= -14: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -14: + interp_atmos_master.sh[35]gridopt=grid0p25 -14: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_15_0p25' -14: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -14: + interp_atmos_master.sh[35]gridopt=grid0p50 -14: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_15_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_15_0p50' -14: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -14: + interp_atmos_master.sh[35]gridopt=grid1p00 -14: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_15_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_15_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_15_1p00' -14: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_15 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_15_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_15_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_15_1p00 -14: 1:0:d=2021032312:ABSV:750 mb:24 hour fcst: -14: 2:21111:d=2021032312:CLMR:750 mb:24 hour fcst: -14: 3:30487:d=2021032312:ICMR:750 mb:24 hour fcst: -14: 4:42179:d=2021032312:RWMR:750 mb:24 hour fcst: -14: 5:48109:d=2021032312:SNMR:750 mb:24 hour fcst: -14: 6:61625:d=2021032312:GRLE:750 mb:24 hour fcst: -14: 7:63559:d=2021032312:O3MR:750 mb:24 hour fcst: -14: 8:90822:d=2021032312:HGT:800 mb:24 hour fcst: -14: 9:115342:d=2021032312:TMP:800 mb:24 hour fcst: -14: 10:135744:d=2021032312:RH:800 mb:24 hour fcst: -14: 11:156221:d=2021032312:TCDC:800 mb:24 hour fcst: -14: 12:169486:d=2021032312:SPFH:800 mb:24 hour fcst: -14: 13:198910:d=2021032312:VVEL:800 mb:24 hour fcst: -14: 14:224610:d=2021032312:DZDT:800 mb:24 hour fcst: -14: 15:251654:d=2021032312:UGRD:800 mb:24 hour fcst: -14: 16:272586:d=2021032312:VGRD:800 mb:24 hour fcst: -14: 17:294338:d=2021032312:ABSV:800 mb:24 hour fcst: -14: 18:315555:d=2021032312:CLMR:800 mb:24 hour fcst: -14: 19:326904:d=2021032312:ICMR:800 mb:24 hour fcst: -14: 20:337857:d=2021032312:RWMR:800 mb:24 hour fcst: -14: 21:345569:d=2021032312:SNMR:800 mb:24 hour fcst: -14: 22:358488:d=2021032312:GRLE:800 mb:24 hour fcst: -14: 23:360800:d=2021032312:O3MR:800 mb:24 hour fcst: -14: 24:387811:d=2021032312:HGT:850 mb:24 hour fcst: -14: 25:412605:d=2021032312:TMP:850 mb:24 hour fcst: -14: 26:433293:d=2021032312:RH:850 mb:24 hour fcst: -14: 27:453583:d=2021032312:TCDC:850 mb:24 hour fcst: -14: 28:468301:d=2021032312:SPFH:850 mb:24 hour fcst: -14: 29:497975:d=2021032312:VVEL:850 mb:24 hour fcst: -14: 30:523255:d=2021032312:DZDT:850 mb:24 hour fcst: -14: 31:549815:d=2021032312:UGRD:850 mb:24 hour fcst: -14: 32:570970:d=2021032312:VGRD:850 mb:24 hour fcst: -14: + interp_atmos_master.sh[47]export err=0 -14: + interp_atmos_master.sh[47]err=0 -14: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -14: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -14: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_15_0p25 -14: + product_functions.sh[5]local filename=pgb2file_f024_15_0p25 -14: + product_functions.sh[6]wgrib2 pgb2file_f024_15_0p25 -not_if :RH: -grib pgb2file_f024_15_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_15_0p25.new -14: 1:0:d=2021032312:ABSV:750 mb:24 hour fcst: -14: 2:502574:d=2021032312:CLMR:750 mb:24 hour fcst: -14: 3:701309:d=2021032312:ICMR:750 mb:24 hour fcst: -14: 4:975467:d=2021032312:RWMR:750 mb:24 hour fcst: -14: 5:1084717:d=2021032312:SNMR:750 mb:24 hour fcst: -14: 6:1386072:d=2021032312:GRLE:750 mb:24 hour fcst: -14: 7:1420123:d=2021032312:O3MR:750 mb:24 hour fcst: -14: 8:2152736:d=2021032312:HGT:800 mb:24 hour fcst: -14: 9:2786749:d=2021032312:TMP:800 mb:24 hour fcst: -14: 10:3265817:d=2021032312:RH:800 mb:24 hour fcst: -14: 11:3748976:d=2021032312:TCDC:800 mb:24 hour fcst: -14: 12:4016525:d=2021032312:SPFH:800 mb:24 hour fcst: -14: 13:4815189:d=2021032312:VVEL:800 mb:24 hour fcst: -14: 14:5499771:d=2021032312:DZDT:800 mb:24 hour fcst: -14: 15:6242441:d=2021032312:UGRD:800 mb:24 hour fcst: -14: 16:6749382:d=2021032312:VGRD:800 mb:24 hour fcst: -14: 17:7287302:d=2021032312:ABSV:800 mb:24 hour fcst: -14: 18:7793323:d=2021032312:CLMR:800 mb:24 hour fcst: -14: 19:8041926:d=2021032312:ICMR:800 mb:24 hour fcst: -14: 20:8297329:d=2021032312:RWMR:800 mb:24 hour fcst: -14: 21:8437620:d=2021032312:SNMR:800 mb:24 hour fcst: -14: 22:8722188:d=2021032312:GRLE:800 mb:24 hour fcst: -14: 23:8764125:d=2021032312:O3MR:800 mb:24 hour fcst: -14: 24:9487362:d=2021032312:HGT:850 mb:24 hour fcst: -14: 25:10127329:d=2021032312:TMP:850 mb:24 hour fcst: -14: 26:10618728:d=2021032312:RH:850 mb:24 hour fcst: -14: 27:11088551:d=2021032312:TCDC:850 mb:24 hour fcst: -14: 28:11388519:d=2021032312:SPFH:850 mb:24 hour fcst: -14: 29:12192120:d=2021032312:VVEL:850 mb:24 hour fcst: -14: 30:12870668:d=2021032312:DZDT:850 mb:24 hour fcst: -14: 31:13603263:d=2021032312:UGRD:850 mb:24 hour fcst: -14: 32:14112965:d=2021032312:VGRD:850 mb:24 hour fcst: -14: + product_functions.sh[10]rc=0 -14: + product_functions.sh[11](( rc == 0 )) -14: + product_functions.sh[11]mv pgb2file_f024_15_0p25.new pgb2file_f024_15_0p25 -14: + product_functions.sh[12]return 0 -14: + interp_atmos_master.sh[56]export err=0 -14: + interp_atmos_master.sh[56]err=0 -14: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -14: ++ interp_atmos_master.sh[62]wc -l -14: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_15_0p25 -match 'LAND|ICEC' -14: + interp_atmos_master.sh[62]var_count=0 -14: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -14: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -14: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_15_0p50 -14: + product_functions.sh[5]local filename=pgb2file_f024_15_0p50 -14: + product_functions.sh[6]wgrib2 pgb2file_f024_15_0p50 -not_if :RH: -grib pgb2file_f024_15_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_15_0p50.new -14: 1:0:d=2021032312:ABSV:750 mb:24 hour fcst: -14: 2:173428:d=2021032312:CLMR:750 mb:24 hour fcst: -14: 3:247889:d=2021032312:ICMR:750 mb:24 hour fcst: -14: 4:345153:d=2021032312:RWMR:750 mb:24 hour fcst: -14: 5:385257:d=2021032312:SNMR:750 mb:24 hour fcst: -14: 6:491575:d=2021032312:GRLE:750 mb:24 hour fcst: -14: 7:504317:d=2021032312:O3MR:750 mb:24 hour fcst: -14: 8:757080:d=2021032312:HGT:800 mb:24 hour fcst: -14: 9:971294:d=2021032312:TMP:800 mb:24 hour fcst: -14: 10:1137145:d=2021032312:RH:800 mb:24 hour fcst: -14: 11:1303975:d=2021032312:TCDC:800 mb:24 hour fcst: -14: 12:1405166:d=2021032312:SPFH:800 mb:24 hour fcst: -14: 13:1685649:d=2021032312:VVEL:800 mb:24 hour fcst: -14: 14:1918981:d=2021032312:DZDT:800 mb:24 hour fcst: -14: 15:2162928:d=2021032312:UGRD:800 mb:24 hour fcst: -14: 16:2337461:d=2021032312:VGRD:800 mb:24 hour fcst: -14: 17:2521705:d=2021032312:ABSV:800 mb:24 hour fcst: -14: 18:2696253:d=2021032312:CLMR:800 mb:24 hour fcst: -14: 19:2789664:d=2021032312:ICMR:800 mb:24 hour fcst: -14: 20:2880537:d=2021032312:RWMR:800 mb:24 hour fcst: -14: 21:2932942:d=2021032312:SNMR:800 mb:24 hour fcst: -14: 22:3034591:d=2021032312:GRLE:800 mb:24 hour fcst: -14: 23:3050091:d=2021032312:O3MR:800 mb:24 hour fcst: -14: 24:3300106:d=2021032312:HGT:850 mb:24 hour fcst: -14: 25:3516695:d=2021032312:TMP:850 mb:24 hour fcst: -14: 26:3686717:d=2021032312:RH:850 mb:24 hour fcst: -14: 27:3851396:d=2021032312:TCDC:850 mb:24 hour fcst: -14: 28:3965161:d=2021032312:SPFH:850 mb:24 hour fcst: -14: 29:4247063:d=2021032312:VVEL:850 mb:24 hour fcst: -14: 30:4477883:d=2021032312:DZDT:850 mb:24 hour fcst: -14: 31:4718057:d=2021032312:UGRD:850 mb:24 hour fcst: -14: 32:4893676:d=2021032312:VGRD:850 mb:24 hour fcst: -14: + product_functions.sh[10]rc=0 -14: + product_functions.sh[11](( rc == 0 )) -14: + product_functions.sh[11]mv pgb2file_f024_15_0p50.new pgb2file_f024_15_0p50 -14: + product_functions.sh[12]return 0 -14: + interp_atmos_master.sh[56]export err=0 -14: + interp_atmos_master.sh[56]err=0 -14: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -14: ++ interp_atmos_master.sh[62]wc -l -14: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_15_0p50 -match 'LAND|ICEC' -14: + interp_atmos_master.sh[62]var_count=0 -14: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -14: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -14: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_15_1p00 -14: + product_functions.sh[5]local filename=pgb2file_f024_15_1p00 -14: + product_functions.sh[6]wgrib2 pgb2file_f024_15_1p00 -not_if :RH: -grib pgb2file_f024_15_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_15_1p00.new -14: 1:0:d=2021032312:ABSV:750 mb:24 hour fcst: -14: 2:55765:d=2021032312:CLMR:750 mb:24 hour fcst: -14: 3:80961:d=2021032312:ICMR:750 mb:24 hour fcst: -14: 4:112924:d=2021032312:RWMR:750 mb:24 hour fcst: -14: 5:127571:d=2021032312:SNMR:750 mb:24 hour fcst: -14: 6:163222:d=2021032312:GRLE:750 mb:24 hour fcst: -14: 7:168076:d=2021032312:O3MR:750 mb:24 hour fcst: -14: 8:244882:d=2021032312:HGT:800 mb:24 hour fcst: -14: 9:311525:d=2021032312:TMP:800 mb:24 hour fcst: -14: 10:365530:d=2021032312:RH:800 mb:24 hour fcst: -14: 11:419692:d=2021032312:TCDC:800 mb:24 hour fcst: -14: 12:454643:d=2021032312:SPFH:800 mb:24 hour fcst: -14: 13:538002:d=2021032312:VVEL:800 mb:24 hour fcst: -14: 14:609387:d=2021032312:DZDT:800 mb:24 hour fcst: -14: 15:683558:d=2021032312:UGRD:800 mb:24 hour fcst: -14: 16:739461:d=2021032312:VGRD:800 mb:24 hour fcst: -14: 17:798117:d=2021032312:ABSV:800 mb:24 hour fcst: -14: 18:854252:d=2021032312:CLMR:800 mb:24 hour fcst: -14: 19:885513:d=2021032312:ICMR:800 mb:24 hour fcst: -14: 20:915081:d=2021032312:RWMR:800 mb:24 hour fcst: -14: 21:934146:d=2021032312:SNMR:800 mb:24 hour fcst: -14: 22:968163:d=2021032312:GRLE:800 mb:24 hour fcst: -14: 23:973905:d=2021032312:O3MR:800 mb:24 hour fcst: -14: 24:1049976:d=2021032312:HGT:850 mb:24 hour fcst: -14: 25:1117405:d=2021032312:TMP:850 mb:24 hour fcst: -14: 26:1172338:d=2021032312:RH:850 mb:24 hour fcst: -14: 27:1225751:d=2021032312:TCDC:850 mb:24 hour fcst: -14: 28:1264490:d=2021032312:SPFH:850 mb:24 hour fcst: -14: 29:1348041:d=2021032312:VVEL:850 mb:24 hour fcst: -14: 30:1418478:d=2021032312:DZDT:850 mb:24 hour fcst: -14: 31:1491363:d=2021032312:UGRD:850 mb:24 hour fcst: -14: 32:1547625:d=2021032312:VGRD:850 mb:24 hour fcst: -14: + product_functions.sh[10]rc=0 -14: + product_functions.sh[11](( rc == 0 )) -14: + product_functions.sh[11]mv pgb2file_f024_15_1p00.new pgb2file_f024_15_1p00 -14: + product_functions.sh[12]return 0 -14: + interp_atmos_master.sh[56]export err=0 -14: + interp_atmos_master.sh[56]err=0 -14: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -14: ++ interp_atmos_master.sh[62]wc -l -14: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_15_1p00 -match 'LAND|ICEC' -14: + interp_atmos_master.sh[62]var_count=0 -14: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -14: + interp_atmos_master.sh[73]exit 0 - 5: + bash[8]'[' -z '' ']' - 5: + bash[9]case "$-" in - 5: + bash[12]__lmod_vx=x - 5: + bash[16]'[' -n x ']' - 5: + bash[16]set +x - 5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 5: Shell debugging restarted - 5: + bash[224]unset __lmod_vx - 5: + interp_atmos_master.sh[7]input_file=tmpfile_f024_6 - 5: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_6 - 5: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 - 5: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 5: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 5: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 5: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 5: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 5: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 5: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 5: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 5: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 5: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 5: + interp_atmos_master.sh[31]IFS=: - 5: + interp_atmos_master.sh[31]read -ra grids - 5: + interp_atmos_master.sh[33]output_grids= - 5: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 5: + interp_atmos_master.sh[35]gridopt=grid0p25 - 5: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_6_0p25' - 5: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 5: + interp_atmos_master.sh[35]gridopt=grid0p50 - 5: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_6_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_6_0p50' - 5: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 5: + interp_atmos_master.sh[35]gridopt=grid1p00 - 5: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_6_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_6_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_6_1p00' - 5: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_6 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_6_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_6_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_6_1p00 - 5: 1:0:d=2021032312:RH:15 mb:24 hour fcst: - 5: 2:5312:d=2021032312:SPFH:15 mb:24 hour fcst: - 5: 3:20785:d=2021032312:VVEL:15 mb:24 hour fcst: - 5: 4:47359:d=2021032312:DZDT:15 mb:24 hour fcst: - 5: 5:77738:d=2021032312:UGRD:15 mb:24 hour fcst: - 5: 6:89004:d=2021032312:VGRD:15 mb:24 hour fcst: - 5: 7:107297:d=2021032312:ABSV:15 mb:24 hour fcst: - 5: 8:124853:d=2021032312:O3MR:15 mb:24 hour fcst: - 5: 9:143904:d=2021032312:HGT:20 mb:24 hour fcst: - 5: 10:163322:d=2021032312:TMP:20 mb:24 hour fcst: - 5: 11:179377:d=2021032312:RH:20 mb:24 hour fcst: - 5: 12:185581:d=2021032312:SPFH:20 mb:24 hour fcst: - 5: 13:199991:d=2021032312:VVEL:20 mb:24 hour fcst: - 5: 14:227349:d=2021032312:DZDT:20 mb:24 hour fcst: - 5: 15:257633:d=2021032312:UGRD:20 mb:24 hour fcst: - 5: 16:276271:d=2021032312:VGRD:20 mb:24 hour fcst: - 5: 17:294809:d=2021032312:ABSV:20 mb:24 hour fcst: - 5: 18:312418:d=2021032312:O3MR:20 mb:24 hour fcst: - 5: 19:337344:d=2021032312:HGT:30 mb:24 hour fcst: - 5: 20:356659:d=2021032312:TMP:30 mb:24 hour fcst: - 5: 21:372665:d=2021032312:RH:30 mb:24 hour fcst: - 5: 22:380259:d=2021032312:SPFH:30 mb:24 hour fcst: - 5: 23:395562:d=2021032312:VVEL:30 mb:24 hour fcst: - 5: 24:424069:d=2021032312:DZDT:30 mb:24 hour fcst: - 5: 25:454048:d=2021032312:UGRD:30 mb:24 hour fcst: - 5: 26:472751:d=2021032312:VGRD:30 mb:24 hour fcst: - 5: 27:491331:d=2021032312:ABSV:30 mb:24 hour fcst: - 5: 28:509031:d=2021032312:O3MR:30 mb:24 hour fcst: - 5: 29:535861:d=2021032312:HGT:40 mb:24 hour fcst: - 5: 30:554711:d=2021032312:TMP:40 mb:24 hour fcst: - 5: 31:571046:d=2021032312:RH:40 mb:24 hour fcst: - 5: + interp_atmos_master.sh[47]export err=0 - 5: + interp_atmos_master.sh[47]err=0 - 5: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 5: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 5: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_6_0p25 - 5: + product_functions.sh[5]local filename=pgb2file_f024_6_0p25 - 5: + product_functions.sh[6]wgrib2 pgb2file_f024_6_0p25 -not_if :RH: -grib pgb2file_f024_6_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_6_0p25.new - 5: 1:0:d=2021032312:RH:15 mb:24 hour fcst: - 5: 2:117615:d=2021032312:SPFH:15 mb:24 hour fcst: - 5: 3:492067:d=2021032312:VVEL:15 mb:24 hour fcst: - 5: 4:1218684:d=2021032312:DZDT:15 mb:24 hour fcst: - 5: 5:2138997:d=2021032312:UGRD:15 mb:24 hour fcst: - 5: 6:2579096:d=2021032312:VGRD:15 mb:24 hour fcst: - 5: 7:3006469:d=2021032312:ABSV:15 mb:24 hour fcst: - 5: 8:3394983:d=2021032312:O3MR:15 mb:24 hour fcst: - 5: 9:3842475:d=2021032312:HGT:20 mb:24 hour fcst: - 5: 10:4300616:d=2021032312:TMP:20 mb:24 hour fcst: - 5: 11:4652648:d=2021032312:RH:20 mb:24 hour fcst: - 5: 12:4819190:d=2021032312:SPFH:20 mb:24 hour fcst: - 5: 13:5154345:d=2021032312:VVEL:20 mb:24 hour fcst: - 5: 14:5913872:d=2021032312:DZDT:20 mb:24 hour fcst: - 5: 15:6808849:d=2021032312:UGRD:20 mb:24 hour fcst: - 5: 16:7243941:d=2021032312:VGRD:20 mb:24 hour fcst: - 5: 17:7673539:d=2021032312:ABSV:20 mb:24 hour fcst: - 5: 18:8065963:d=2021032312:O3MR:20 mb:24 hour fcst: - 5: 19:8731699:d=2021032312:HGT:30 mb:24 hour fcst: - 5: 20:9180442:d=2021032312:TMP:30 mb:24 hour fcst: - 5: 21:9533263:d=2021032312:RH:30 mb:24 hour fcst: - 5: 22:9747003:d=2021032312:SPFH:30 mb:24 hour fcst: - 5: 23:10115688:d=2021032312:VVEL:30 mb:24 hour fcst: - 5: 24:10923292:d=2021032312:DZDT:30 mb:24 hour fcst: - 5: 25:11818507:d=2021032312:UGRD:30 mb:24 hour fcst: - 5: 26:12252910:d=2021032312:VGRD:30 mb:24 hour fcst: - 5: 27:12684912:d=2021032312:ABSV:30 mb:24 hour fcst: - 5: 28:13080935:d=2021032312:O3MR:30 mb:24 hour fcst: - 5: 29:13814712:d=2021032312:HGT:40 mb:24 hour fcst: - 5: 30:14255668:d=2021032312:TMP:40 mb:24 hour fcst: - 5: 31:14619090:d=2021032312:RH:40 mb:24 hour fcst: - 5: + product_functions.sh[10]rc=0 - 5: + product_functions.sh[11](( rc == 0 )) - 5: + product_functions.sh[11]mv pgb2file_f024_6_0p25.new pgb2file_f024_6_0p25 - 5: + product_functions.sh[12]return 0 - 5: + interp_atmos_master.sh[56]export err=0 - 5: + interp_atmos_master.sh[56]err=0 - 5: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 5: ++ interp_atmos_master.sh[62]wc -l - 5: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_6_0p25 -match 'LAND|ICEC' - 5: + interp_atmos_master.sh[62]var_count=0 - 5: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 5: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 5: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_6_0p50 - 5: + product_functions.sh[5]local filename=pgb2file_f024_6_0p50 - 5: + product_functions.sh[6]wgrib2 pgb2file_f024_6_0p50 -not_if :RH: -grib pgb2file_f024_6_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_6_0p50.new - 5: 1:0:d=2021032312:RH:15 mb:24 hour fcst: - 5: 2:41615:d=2021032312:SPFH:15 mb:24 hour fcst: - 5: 3:162636:d=2021032312:VVEL:15 mb:24 hour fcst: - 5: 4:415200:d=2021032312:DZDT:15 mb:24 hour fcst: - 5: 5:711790:d=2021032312:UGRD:15 mb:24 hour fcst: - 5: 6:864855:d=2021032312:VGRD:15 mb:24 hour fcst: - 5: 7:1013016:d=2021032312:ABSV:15 mb:24 hour fcst: - 5: 8:1146542:d=2021032312:O3MR:15 mb:24 hour fcst: - 5: 9:1299291:d=2021032312:HGT:20 mb:24 hour fcst: - 5: 10:1454988:d=2021032312:TMP:20 mb:24 hour fcst: - 5: 11:1575103:d=2021032312:RH:20 mb:24 hour fcst: - 5: 12:1629040:d=2021032312:SPFH:20 mb:24 hour fcst: - 5: 13:1738412:d=2021032312:VVEL:20 mb:24 hour fcst: - 5: 14:1999406:d=2021032312:DZDT:20 mb:24 hour fcst: - 5: 15:2291983:d=2021032312:UGRD:20 mb:24 hour fcst: - 5: 16:2441205:d=2021032312:VGRD:20 mb:24 hour fcst: - 5: 17:2590207:d=2021032312:ABSV:20 mb:24 hour fcst: - 5: 18:2724662:d=2021032312:O3MR:20 mb:24 hour fcst: - 5: 19:2948970:d=2021032312:HGT:30 mb:24 hour fcst: - 5: 20:3102080:d=2021032312:TMP:30 mb:24 hour fcst: - 5: 21:3222010:d=2021032312:RH:30 mb:24 hour fcst: - 5: 22:3284101:d=2021032312:SPFH:30 mb:24 hour fcst: - 5: 23:3403576:d=2021032312:VVEL:30 mb:24 hour fcst: - 5: 24:3678500:d=2021032312:DZDT:30 mb:24 hour fcst: - 5: 25:3967395:d=2021032312:UGRD:30 mb:24 hour fcst: - 5: 26:4116211:d=2021032312:VGRD:30 mb:24 hour fcst: - 5: 27:4266060:d=2021032312:ABSV:30 mb:24 hour fcst: - 5: 28:4402730:d=2021032312:O3MR:30 mb:24 hour fcst: - 5: 29:4653808:d=2021032312:HGT:40 mb:24 hour fcst: - 5: 30:4804176:d=2021032312:TMP:40 mb:24 hour fcst: - 5: 31:4928261:d=2021032312:RH:40 mb:24 hour fcst: - 5: + product_functions.sh[10]rc=0 - 5: + product_functions.sh[11](( rc == 0 )) - 5: + product_functions.sh[11]mv pgb2file_f024_6_0p50.new pgb2file_f024_6_0p50 - 5: + product_functions.sh[12]return 0 - 5: + interp_atmos_master.sh[56]export err=0 - 5: + interp_atmos_master.sh[56]err=0 - 5: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 5: ++ interp_atmos_master.sh[62]wc -l - 5: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_6_0p50 -match 'LAND|ICEC' - 5: + interp_atmos_master.sh[62]var_count=0 - 5: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 5: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 5: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_6_1p00 - 5: + product_functions.sh[5]local filename=pgb2file_f024_6_1p00 - 5: + product_functions.sh[6]wgrib2 pgb2file_f024_6_1p00 -not_if :RH: -grib pgb2file_f024_6_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_6_1p00.new - 5: 1:0:d=2021032312:RH:15 mb:24 hour fcst: - 5: 2:13852:d=2021032312:SPFH:15 mb:24 hour fcst: - 5: 3:52740:d=2021032312:VVEL:15 mb:24 hour fcst: - 5: 4:128309:d=2021032312:DZDT:15 mb:24 hour fcst: - 5: 5:215129:d=2021032312:UGRD:15 mb:24 hour fcst: - 5: 6:265010:d=2021032312:VGRD:15 mb:24 hour fcst: - 5: 7:313348:d=2021032312:ABSV:15 mb:24 hour fcst: - 5: 8:357834:d=2021032312:O3MR:15 mb:24 hour fcst: - 5: 9:407227:d=2021032312:HGT:20 mb:24 hour fcst: - 5: 10:457166:d=2021032312:TMP:20 mb:24 hour fcst: - 5: 11:497630:d=2021032312:RH:20 mb:24 hour fcst: - 5: 12:513781:d=2021032312:SPFH:20 mb:24 hour fcst: - 5: 13:549277:d=2021032312:VVEL:20 mb:24 hour fcst: - 5: 14:627142:d=2021032312:DZDT:20 mb:24 hour fcst: - 5: 15:712738:d=2021032312:UGRD:20 mb:24 hour fcst: - 5: 16:761239:d=2021032312:VGRD:20 mb:24 hour fcst: - 5: 17:809720:d=2021032312:ABSV:20 mb:24 hour fcst: - 5: 18:854626:d=2021032312:O3MR:20 mb:24 hour fcst: - 5: 19:922893:d=2021032312:HGT:30 mb:24 hour fcst: - 5: 20:972282:d=2021032312:TMP:30 mb:24 hour fcst: - 5: 21:1012487:d=2021032312:RH:30 mb:24 hour fcst: - 5: 22:1031147:d=2021032312:SPFH:30 mb:24 hour fcst: - 5: 23:1069821:d=2021032312:VVEL:30 mb:24 hour fcst: - 5: 24:1151689:d=2021032312:DZDT:30 mb:24 hour fcst: - 5: 25:1236500:d=2021032312:UGRD:30 mb:24 hour fcst: - 5: 26:1284857:d=2021032312:VGRD:30 mb:24 hour fcst: - 5: 27:1333595:d=2021032312:ABSV:30 mb:24 hour fcst: - 5: 28:1378997:d=2021032312:O3MR:30 mb:24 hour fcst: - 5: 29:1454462:d=2021032312:HGT:40 mb:24 hour fcst: - 5: 30:1503094:d=2021032312:TMP:40 mb:24 hour fcst: - 5: 31:1544668:d=2021032312:RH:40 mb:24 hour fcst: - 5: + product_functions.sh[10]rc=0 - 5: + product_functions.sh[11](( rc == 0 )) - 5: + product_functions.sh[11]mv pgb2file_f024_6_1p00.new pgb2file_f024_6_1p00 - 5: + product_functions.sh[12]return 0 - 5: + interp_atmos_master.sh[56]export err=0 - 5: + interp_atmos_master.sh[56]err=0 - 5: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 5: ++ interp_atmos_master.sh[62]wc -l - 5: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_6_1p00 -match 'LAND|ICEC' - 5: + interp_atmos_master.sh[62]var_count=0 - 5: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 5: + interp_atmos_master.sh[73]exit 0 -11: + bash[8]'[' -z '' ']' -11: + bash[9]case "$-" in -11: + bash[12]__lmod_vx=x -11: + bash[16]'[' -n x ']' -11: + bash[16]set +x -11: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -11: Shell debugging restarted -11: + bash[224]unset __lmod_vx -11: + interp_atmos_master.sh[7]input_file=tmpfile_f024_12 -11: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f024_12 -11: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -11: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -11: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -11: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -11: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -11: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -11: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -11: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -11: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -11: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -11: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -11: + interp_atmos_master.sh[31]IFS=: -11: + interp_atmos_master.sh[31]read -ra grids -11: + interp_atmos_master.sh[33]output_grids= -11: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -11: + interp_atmos_master.sh[35]gridopt=grid0p25 -11: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_12_0p25' -11: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -11: + interp_atmos_master.sh[35]gridopt=grid0p50 -11: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_12_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_12_0p50' -11: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -11: + interp_atmos_master.sh[35]gridopt=grid1p00 -11: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_12_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_12_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_12_1p00' -11: + interp_atmos_master.sh[40]wgrib2 tmpfile_f024_12 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f024_12_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2file_f024_12_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2file_f024_12_1p00 -11: 1:0:d=2021032312:CLMR:450 mb:24 hour fcst: -11: 2:939:d=2021032312:ICMR:450 mb:24 hour fcst: -11: 3:17500:d=2021032312:RWMR:450 mb:24 hour fcst: -11: 4:18019:d=2021032312:SNMR:450 mb:24 hour fcst: -11: 5:32897:d=2021032312:GRLE:450 mb:24 hour fcst: -11: 6:34166:d=2021032312:O3MR:450 mb:24 hour fcst: -11: 7:58263:d=2021032312:HGT:500 mb:24 hour fcst: -11: 8:81990:d=2021032312:TMP:500 mb:24 hour fcst: -11: 9:100531:d=2021032312:RH:500 mb:24 hour fcst: -11: 10:120106:d=2021032312:TCDC:500 mb:24 hour fcst: -11: 11:132308:d=2021032312:SPFH:500 mb:24 hour fcst: -11: 12:160655:d=2021032312:VVEL:500 mb:24 hour fcst: -11: 13:185728:d=2021032312:DZDT:500 mb:24 hour fcst: -11: 14:213451:d=2021032312:UGRD:500 mb:24 hour fcst: -11: 15:234888:d=2021032312:VGRD:500 mb:24 hour fcst: -11: 16:256928:d=2021032312:ABSV:500 mb:24 hour fcst: -11: 17:278097:d=2021032312:CLMR:500 mb:24 hour fcst: -11: 18:279741:d=2021032312:ICMR:500 mb:24 hour fcst: -11: 19:297129:d=2021032312:RWMR:500 mb:24 hour fcst: -11: 20:298070:d=2021032312:SNMR:500 mb:24 hour fcst: -11: 21:312344:d=2021032312:GRLE:500 mb:24 hour fcst: -11: 22:314252:d=2021032312:O3MR:500 mb:24 hour fcst: -11: 23:337670:d=2021032312:HGT:550 mb:24 hour fcst: -11: 24:361237:d=2021032312:TMP:550 mb:24 hour fcst: -11: 25:379910:d=2021032312:RH:550 mb:24 hour fcst: -11: 26:399758:d=2021032312:TCDC:550 mb:24 hour fcst: -11: 27:411937:d=2021032312:SPFH:550 mb:24 hour fcst: -11: 28:441528:d=2021032312:VVEL:550 mb:24 hour fcst: -11: 29:466775:d=2021032312:DZDT:550 mb:24 hour fcst: -11: 30:494312:d=2021032312:UGRD:550 mb:24 hour fcst: -11: 31:515485:d=2021032312:VGRD:550 mb:24 hour fcst: -11: + interp_atmos_master.sh[47]export err=0 -11: + interp_atmos_master.sh[47]err=0 -11: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -11: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -11: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_12_0p25 -11: + product_functions.sh[5]local filename=pgb2file_f024_12_0p25 -11: + product_functions.sh[6]wgrib2 pgb2file_f024_12_0p25 -not_if :RH: -grib pgb2file_f024_12_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_12_0p25.new -11: 1:0:d=2021032312:CLMR:450 mb:24 hour fcst: -11: 2:16974:d=2021032312:ICMR:450 mb:24 hour fcst: -11: 3:386860:d=2021032312:RWMR:450 mb:24 hour fcst: -11: 4:392605:d=2021032312:SNMR:450 mb:24 hour fcst: -11: 5:713683:d=2021032312:GRLE:450 mb:24 hour fcst: -11: 6:729933:d=2021032312:O3MR:450 mb:24 hour fcst: -11: 7:1347251:d=2021032312:HGT:500 mb:24 hour fcst: -11: 8:1961038:d=2021032312:TMP:500 mb:24 hour fcst: -11: 9:2390485:d=2021032312:RH:500 mb:24 hour fcst: -11: 10:2838170:d=2021032312:TCDC:500 mb:24 hour fcst: -11: 11:3069441:d=2021032312:SPFH:500 mb:24 hour fcst: -11: 12:3827462:d=2021032312:VVEL:500 mb:24 hour fcst: -11: 13:4495783:d=2021032312:DZDT:500 mb:24 hour fcst: -11: 14:5274399:d=2021032312:UGRD:500 mb:24 hour fcst: -11: 15:5797317:d=2021032312:VGRD:500 mb:24 hour fcst: -11: 16:6347855:d=2021032312:ABSV:500 mb:24 hour fcst: -11: 17:6859413:d=2021032312:CLMR:500 mb:24 hour fcst: -11: 18:6892210:d=2021032312:ICMR:500 mb:24 hour fcst: -11: 19:7285464:d=2021032312:RWMR:500 mb:24 hour fcst: -11: 20:7300570:d=2021032312:SNMR:500 mb:24 hour fcst: -11: 21:7601195:d=2021032312:GRLE:500 mb:24 hour fcst: -11: 22:7627824:d=2021032312:O3MR:500 mb:24 hour fcst: -11: 23:8227574:d=2021032312:HGT:550 mb:24 hour fcst: -11: 24:8826997:d=2021032312:TMP:550 mb:24 hour fcst: -11: 25:9248384:d=2021032312:RH:550 mb:24 hour fcst: -11: 26:9700948:d=2021032312:TCDC:550 mb:24 hour fcst: -11: 27:9928851:d=2021032312:SPFH:550 mb:24 hour fcst: -11: 28:10732468:d=2021032312:VVEL:550 mb:24 hour fcst: -11: 29:11401621:d=2021032312:DZDT:550 mb:24 hour fcst: -11: 30:12171229:d=2021032312:UGRD:550 mb:24 hour fcst: -11: 31:12683979:d=2021032312:VGRD:550 mb:24 hour fcst: -11: + product_functions.sh[10]rc=0 -11: + product_functions.sh[11](( rc == 0 )) -11: + product_functions.sh[11]mv pgb2file_f024_12_0p25.new pgb2file_f024_12_0p25 -11: + product_functions.sh[12]return 0 -11: + interp_atmos_master.sh[56]export err=0 -11: + interp_atmos_master.sh[56]err=0 -11: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -11: ++ interp_atmos_master.sh[62]wc -l -11: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_12_0p25 -match 'LAND|ICEC' -11: + interp_atmos_master.sh[62]var_count=0 -11: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -11: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -11: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_12_0p50 -11: + product_functions.sh[5]local filename=pgb2file_f024_12_0p50 -11: + product_functions.sh[6]wgrib2 pgb2file_f024_12_0p50 -not_if :RH: -grib pgb2file_f024_12_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_12_0p50.new -11: 1:0:d=2021032312:CLMR:450 mb:24 hour fcst: -11: 2:6546:d=2021032312:ICMR:450 mb:24 hour fcst: -11: 3:139214:d=2021032312:RWMR:450 mb:24 hour fcst: -11: 4:141573:d=2021032312:SNMR:450 mb:24 hour fcst: -11: 5:257648:d=2021032312:GRLE:450 mb:24 hour fcst: -11: 6:263908:d=2021032312:O3MR:450 mb:24 hour fcst: -11: 7:473495:d=2021032312:HGT:500 mb:24 hour fcst: -11: 8:679959:d=2021032312:TMP:500 mb:24 hour fcst: -11: 9:825729:d=2021032312:RH:500 mb:24 hour fcst: -11: 10:982031:d=2021032312:TCDC:500 mb:24 hour fcst: -11: 11:1071460:d=2021032312:SPFH:500 mb:24 hour fcst: -11: 12:1335301:d=2021032312:VVEL:500 mb:24 hour fcst: -11: 13:1563411:d=2021032312:DZDT:500 mb:24 hour fcst: -11: 14:1817682:d=2021032312:UGRD:500 mb:24 hour fcst: -11: 15:1996885:d=2021032312:VGRD:500 mb:24 hour fcst: -11: 16:2184824:d=2021032312:ABSV:500 mb:24 hour fcst: -11: 17:2359575:d=2021032312:CLMR:500 mb:24 hour fcst: -11: 18:2371977:d=2021032312:ICMR:500 mb:24 hour fcst: -11: 19:2513579:d=2021032312:RWMR:500 mb:24 hour fcst: -11: 20:2519515:d=2021032312:SNMR:500 mb:24 hour fcst: -11: 21:2627488:d=2021032312:GRLE:500 mb:24 hour fcst: -11: 22:2638156:d=2021032312:O3MR:500 mb:24 hour fcst: -11: 23:2840456:d=2021032312:HGT:550 mb:24 hour fcst: -11: 24:3043409:d=2021032312:TMP:550 mb:24 hour fcst: -11: 25:3190569:d=2021032312:RH:550 mb:24 hour fcst: -11: 26:3348715:d=2021032312:TCDC:550 mb:24 hour fcst: -11: 27:3435604:d=2021032312:SPFH:550 mb:24 hour fcst: -11: 28:3712732:d=2021032312:VVEL:550 mb:24 hour fcst: -11: 29:3941714:d=2021032312:DZDT:550 mb:24 hour fcst: -11: 30:4193679:d=2021032312:UGRD:550 mb:24 hour fcst: -11: 31:4369485:d=2021032312:VGRD:550 mb:24 hour fcst: -11: + product_functions.sh[10]rc=0 -11: + product_functions.sh[11](( rc == 0 )) -11: + product_functions.sh[11]mv pgb2file_f024_12_0p50.new pgb2file_f024_12_0p50 -11: + product_functions.sh[12]return 0 -11: + interp_atmos_master.sh[56]export err=0 -11: + interp_atmos_master.sh[56]err=0 -11: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -11: ++ interp_atmos_master.sh[62]wc -l -11: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_12_0p50 -match 'LAND|ICEC' -11: + interp_atmos_master.sh[62]var_count=0 -11: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -11: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -11: + interp_atmos_master.sh[55]trim_rh pgb2file_f024_12_1p00 -11: + product_functions.sh[5]local filename=pgb2file_f024_12_1p00 -11: + product_functions.sh[6]wgrib2 pgb2file_f024_12_1p00 -not_if :RH: -grib pgb2file_f024_12_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f024_12_1p00.new -11: 1:0:d=2021032312:CLMR:450 mb:24 hour fcst: -11: 2:2467:d=2021032312:ICMR:450 mb:24 hour fcst: -11: 3:47100:d=2021032312:RWMR:450 mb:24 hour fcst: -11: 4:48145:d=2021032312:SNMR:450 mb:24 hour fcst: -11: 5:87275:d=2021032312:GRLE:450 mb:24 hour fcst: -11: 6:90108:d=2021032312:O3MR:450 mb:24 hour fcst: -11: 7:155443:d=2021032312:HGT:500 mb:24 hour fcst: -11: 8:219923:d=2021032312:TMP:500 mb:24 hour fcst: -11: 9:267970:d=2021032312:RH:500 mb:24 hour fcst: -11: 10:319226:d=2021032312:TCDC:500 mb:24 hour fcst: -11: 11:350717:d=2021032312:SPFH:500 mb:24 hour fcst: -11: 12:429893:d=2021032312:VVEL:500 mb:24 hour fcst: -11: 13:499479:d=2021032312:DZDT:500 mb:24 hour fcst: -11: 14:575885:d=2021032312:UGRD:500 mb:24 hour fcst: -11: 15:632810:d=2021032312:VGRD:500 mb:24 hour fcst: -11: 16:692402:d=2021032312:ABSV:500 mb:24 hour fcst: -11: 17:748532:d=2021032312:CLMR:500 mb:24 hour fcst: -11: 18:753105:d=2021032312:ICMR:500 mb:24 hour fcst: -11: 19:800019:d=2021032312:RWMR:500 mb:24 hour fcst: -11: 20:802375:d=2021032312:SNMR:500 mb:24 hour fcst: -11: 21:838982:d=2021032312:GRLE:500 mb:24 hour fcst: -11: 22:843317:d=2021032312:O3MR:500 mb:24 hour fcst: -11: 23:906955:d=2021032312:HGT:550 mb:24 hour fcst: -11: 24:970905:d=2021032312:TMP:550 mb:24 hour fcst: -11: 25:1019039:d=2021032312:RH:550 mb:24 hour fcst: -11: 26:1070859:d=2021032312:TCDC:550 mb:24 hour fcst: -11: 27:1102057:d=2021032312:SPFH:550 mb:24 hour fcst: -11: 28:1185147:d=2021032312:VVEL:550 mb:24 hour fcst: -11: 29:1255326:d=2021032312:DZDT:550 mb:24 hour fcst: -11: 30:1331161:d=2021032312:UGRD:550 mb:24 hour fcst: -11: 31:1387300:d=2021032312:VGRD:550 mb:24 hour fcst: -11: + product_functions.sh[10]rc=0 -11: + product_functions.sh[11](( rc == 0 )) -11: + product_functions.sh[11]mv pgb2file_f024_12_1p00.new pgb2file_f024_12_1p00 -11: + product_functions.sh[12]return 0 -11: + interp_atmos_master.sh[56]export err=0 -11: + interp_atmos_master.sh[56]err=0 -11: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -11: ++ interp_atmos_master.sh[62]wc -l -11: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f024_12_1p00 -match 'LAND|ICEC' -11: + interp_atmos_master.sh[62]var_count=0 -11: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -11: + interp_atmos_master.sh[73]exit 0 -+ run_mpmd.sh[113]exit 0 -+ run_mpmd.sh[1]postamble run_mpmd.sh 1753755885 0 -+ preamble.sh[62]set +x -End run_mpmd.sh at 02:25:16 with error code 0 (time elapsed: 00:00:31) -+ exglobal_atmos_products.sh[142]true -+ exglobal_atmos_products.sh[143]export err=0 -+ exglobal_atmos_products.sh[143]err=0 -+ exglobal_atmos_products.sh[144][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[149]mv mpmd.out mpmd_1.out -+ exglobal_atmos_products.sh[153]echo 'Concatenating processor-specific grib2 files into a single product file' -Concatenating processor-specific grib2 files into a single product file -+ exglobal_atmos_products.sh[154](( iproc = 1 )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_1_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_1_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_1_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_1_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_1_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_1_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_1 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_2_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_2_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_2_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_2_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_2_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_2_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_2 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_3_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_3_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_3_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_3_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_3_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_3_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_3 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_4_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_4_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_4_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_4_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_4_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_4_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_4 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_5_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_5_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_5_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_5_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_5_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_5_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_5 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_6_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_6_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_6_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_6_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_6_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_6_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_6 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_7_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_7_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_7_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_7_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_7_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_7_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_7 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_8_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_8_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_8_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_8_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_8_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_8_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_8 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_9_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_9_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_9_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_9_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_9_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_9_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_9 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_10_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_10_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_10_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_10_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_10_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_10_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_10 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_11_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_11_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_11_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_11_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_11_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_11_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_11 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_12_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_12_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_12_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_12_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_12_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_12_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_12 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_13_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_13_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_13_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_13_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_13_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_13_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_13 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_14_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_14_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_14_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_14_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_14_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_14_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_14 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_15_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_15_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_15_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_15_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_15_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_15_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_15 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_16_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_16_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_16_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_16_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_16_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_16_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_16 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_17_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_17_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_17_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_17_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_17_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_17_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_17 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_18_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_18_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_18_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_18_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_18_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_18_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_18 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_19_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_19_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_19_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_19_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_19_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_19_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_19 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_20_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_20_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_20_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_20_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_20_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_20_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_20 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_21_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_21_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_21_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_21_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_21_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_21_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_21 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_22_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_22_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_22_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_22_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_22_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_22_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_22 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_23_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_23_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_23_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_23_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_23_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_23_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_23 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_24_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_24_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_24_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_24_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f024_24_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f024_24_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f024_24 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[164]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[165]prod_dir=COMOUT_ATMOS_GRIB_0p25 -+ exglobal_atmos_products.sh[166]cpfs pgb2file_f024_0p25 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f024 -+ cpfs[3]'[' 2 -ne 2 ']' -+ cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f024 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f024 = ./ ']' -+ cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f024 ']' -+ cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f024 -+ cpfs[16]cp pgb2file_f024_0p25 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f024.cptmp -+ cpfs[18]'[' 0 -ne 0 ']' -+ cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f024.cptmp -+ cpfs[23]'[' 0 -ne 0 ']' -+ cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f024.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f024 -+ cpfs[28]'[' 0 -ne 0 ']' -+ exglobal_atmos_products.sh[167]wgrib2 -s pgb2file_f024_0p25 -+ exglobal_atmos_products.sh[164]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[165]prod_dir=COMOUT_ATMOS_GRIB_0p50 -+ exglobal_atmos_products.sh[166]cpfs pgb2file_f024_0p50 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50/gfs.t12z.pgrb2.0p50.f024 -+ cpfs[3]'[' 2 -ne 2 ']' -+ cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50/gfs.t12z.pgrb2.0p50.f024 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50/gfs.t12z.pgrb2.0p50.f024 = ./ ']' -+ cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50/gfs.t12z.pgrb2.0p50.f024 ']' -+ cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50/gfs.t12z.pgrb2.0p50.f024 -+ cpfs[16]cp pgb2file_f024_0p50 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50/gfs.t12z.pgrb2.0p50.f024.cptmp -+ cpfs[18]'[' 0 -ne 0 ']' -+ cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50/gfs.t12z.pgrb2.0p50.f024.cptmp -+ cpfs[23]'[' 0 -ne 0 ']' -+ cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50/gfs.t12z.pgrb2.0p50.f024.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50/gfs.t12z.pgrb2.0p50.f024 -+ cpfs[28]'[' 0 -ne 0 ']' -+ exglobal_atmos_products.sh[167]wgrib2 -s pgb2file_f024_0p50 -+ exglobal_atmos_products.sh[164]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[165]prod_dir=COMOUT_ATMOS_GRIB_1p00 -+ exglobal_atmos_products.sh[166]cpfs pgb2file_f024_1p00 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.pgrb2.1p00.f024 -+ cpfs[3]'[' 2 -ne 2 ']' -+ cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.pgrb2.1p00.f024 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.pgrb2.1p00.f024 = ./ ']' -+ cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.pgrb2.1p00.f024 ']' -+ cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.pgrb2.1p00.f024 -+ cpfs[16]cp pgb2file_f024_1p00 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.pgrb2.1p00.f024.cptmp -+ cpfs[18]'[' 0 -ne 0 ']' -+ cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.pgrb2.1p00.f024.cptmp -+ cpfs[23]'[' 0 -ne 0 ']' -+ cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.pgrb2.1p00.f024.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.pgrb2.1p00.f024 -+ cpfs[28]'[' 0 -ne 0 ']' -+ exglobal_atmos_products.sh[167]wgrib2 -s pgb2file_f024_1p00 -+ exglobal_atmos_products.sh[170]echo 'Finished processing nset = 1' -Finished processing nset = 1 -+ exglobal_atmos_products.sh[154](( nset++ )) -+ exglobal_atmos_products.sh[154](( nset <= downset )) -+ exglobal_atmos_products.sh[74]echo 'Begin processing nset = 2' -Begin processing nset = 2 -+ exglobal_atmos_products.sh[77]nproc=24 -+ exglobal_atmos_products.sh[80][[ 2 == 1 ]] -+ exglobal_atmos_products.sh[82][[ 2 == 2 ]] -+ exglobal_atmos_products.sh[83]grp=b -+ exglobal_atmos_products.sh[87]tmpfile=tmpfileb_f024 -++ exglobal_atmos_products.sh[90]wgrib2 tmpfileb_f024 -++ exglobal_atmos_products.sh[90]wc -l -+ exglobal_atmos_products.sh[90]ncount=349 -+ exglobal_atmos_products.sh[91][[ 24 -gt 349 ]] -+ exglobal_atmos_products.sh[95]inv=14 -+ exglobal_atmos_products.sh[96]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601/poescript -+ exglobal_atmos_products.sh[98]last=0 -+ exglobal_atmos_products.sh[134](( iproc = 1 )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=1 -+ exglobal_atmos_products.sh[101]last=14 -+ exglobal_atmos_products.sh[102][[ 14 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 14 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 1 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 1:14 -grib tmpfileb_f024_1 -1:0:d=2021032312:HGT:1 mb:24 hour fcst: -2:18247:d=2021032312:TMP:1 mb:24 hour fcst: -3:34112:d=2021032312:RH:1 mb:24 hour fcst: -4:43938:d=2021032312:UGRD:1 mb:24 hour fcst: -5:62539:d=2021032312:VGRD:1 mb:24 hour fcst: -6:80240:d=2021032312:ABSV:1 mb:24 hour fcst: -7:96594:d=2021032312:O3MR:1 mb:24 hour fcst: -8:117662:d=2021032312:HGT:2 mb:24 hour fcst: -9:137974:d=2021032312:TMP:2 mb:24 hour fcst: -10:154578:d=2021032312:RH:2 mb:24 hour fcst: -11:163159:d=2021032312:UGRD:2 mb:24 hour fcst: -12:174744:d=2021032312:VGRD:2 mb:24 hour fcst: -13:193090:d=2021032312:ABSV:2 mb:24 hour fcst: -14:210173:d=2021032312:O3MR:2 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_1 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_1 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_1 pgb2bfile_f024_1 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 14 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=15 -+ exglobal_atmos_products.sh[101]last=28 -+ exglobal_atmos_products.sh[102][[ 28 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 28 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 2 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 15:28 -grib tmpfileb_f024_2 -15:232991:d=2021032312:HGT:3 mb:24 hour fcst: -16:250827:d=2021032312:TMP:3 mb:24 hour fcst: -17:267300:d=2021032312:RH:3 mb:24 hour fcst: -18:273330:d=2021032312:UGRD:3 mb:24 hour fcst: -19:284713:d=2021032312:VGRD:3 mb:24 hour fcst: -20:302850:d=2021032312:ABSV:3 mb:24 hour fcst: -21:319864:d=2021032312:O3MR:3 mb:24 hour fcst: -22:343237:d=2021032312:HGT:5 mb:24 hour fcst: -23:360912:d=2021032312:TMP:5 mb:24 hour fcst: -24:377541:d=2021032312:RH:5 mb:24 hour fcst: -25:386583:d=2021032312:UGRD:5 mb:24 hour fcst: -26:398175:d=2021032312:VGRD:5 mb:24 hour fcst: -27:416745:d=2021032312:ABSV:5 mb:24 hour fcst: -28:434366:d=2021032312:O3MR:5 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_2 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_2 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_2 pgb2bfile_f024_2 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 28 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=29 -+ exglobal_atmos_products.sh[101]last=42 -+ exglobal_atmos_products.sh[102][[ 42 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 42 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 3 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 29:42 -grib tmpfileb_f024_3 -29:457026:d=2021032312:HGT:7 mb:24 hour fcst: -30:474785:d=2021032312:TMP:7 mb:24 hour fcst: -31:490974:d=2021032312:RH:7 mb:24 hour fcst: -32:497093:d=2021032312:UGRD:7 mb:24 hour fcst: -33:508507:d=2021032312:VGRD:7 mb:24 hour fcst: -34:527199:d=2021032312:ABSV:7 mb:24 hour fcst: -35:545008:d=2021032312:O3MR:7 mb:24 hour fcst: -36:568782:d=2021032312:TCDC:70 mb:24 hour fcst: -37:568961:d=2021032312:CLMR:70 mb:24 hour fcst: -38:569140:d=2021032312:ICMR:70 mb:24 hour fcst: -39:569319:d=2021032312:RWMR:70 mb:24 hour fcst: -40:569498:d=2021032312:SNMR:70 mb:24 hour fcst: -41:569677:d=2021032312:GRLE:70 mb:24 hour fcst: -42:569856:d=2021032312:HGT:125 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_3 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_3 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_3 pgb2bfile_f024_3 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 42 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=43 -+ exglobal_atmos_products.sh[101]last=56 -+ exglobal_atmos_products.sh[102][[ 56 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 56 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 4 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 43:56 -grib tmpfileb_f024_4 -43:589944:d=2021032312:TMP:125 mb:24 hour fcst: -44:606985:d=2021032312:RH:125 mb:24 hour fcst: -45:617630:d=2021032312:TCDC:125 mb:24 hour fcst: -46:619994:d=2021032312:VVEL:125 mb:24 hour fcst: -47:646438:d=2021032312:DZDT:125 mb:24 hour fcst: -48:671592:d=2021032312:UGRD:125 mb:24 hour fcst: -49:691127:d=2021032312:VGRD:125 mb:24 hour fcst: -50:710779:d=2021032312:ABSV:125 mb:24 hour fcst: -51:729106:d=2021032312:CLMR:125 mb:24 hour fcst: -52:729285:d=2021032312:ICMR:125 mb:24 hour fcst: -53:733679:d=2021032312:RWMR:125 mb:24 hour fcst: -54:733858:d=2021032312:SNMR:125 mb:24 hour fcst: -55:737339:d=2021032312:GRLE:125 mb:24 hour fcst: -56:737824:d=2021032312:HGT:175 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_4 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_4 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_4 pgb2bfile_f024_4 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 56 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=57 -+ exglobal_atmos_products.sh[101]last=70 -+ exglobal_atmos_products.sh[102][[ 70 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 70 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 5 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 57:70 -grib tmpfileb_f024_5 -57:758470:d=2021032312:TMP:175 mb:24 hour fcst: -58:775998:d=2021032312:RH:175 mb:24 hour fcst: -59:789153:d=2021032312:TCDC:175 mb:24 hour fcst: -60:793770:d=2021032312:VVEL:175 mb:24 hour fcst: -61:822255:d=2021032312:DZDT:175 mb:24 hour fcst: -62:848389:d=2021032312:UGRD:175 mb:24 hour fcst: -63:861350:d=2021032312:VGRD:175 mb:24 hour fcst: -64:882120:d=2021032312:ABSV:175 mb:24 hour fcst: -65:901519:d=2021032312:CLMR:175 mb:24 hour fcst: -66:901698:d=2021032312:ICMR:175 mb:24 hour fcst: -67:908216:d=2021032312:RWMR:175 mb:24 hour fcst: -68:908395:d=2021032312:SNMR:175 mb:24 hour fcst: -69:912104:d=2021032312:GRLE:175 mb:24 hour fcst: -70:912601:d=2021032312:HGT:225 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_5 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_5 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_5 pgb2bfile_f024_5 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 70 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=71 -+ exglobal_atmos_products.sh[101]last=84 -+ exglobal_atmos_products.sh[102][[ 84 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 84 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 6 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 71:84 -grib tmpfileb_f024_6 -71:934057:d=2021032312:TMP:225 mb:24 hour fcst: -72:952496:d=2021032312:RH:225 mb:24 hour fcst: -73:969124:d=2021032312:TCDC:225 mb:24 hour fcst: -74:977789:d=2021032312:VVEL:225 mb:24 hour fcst: -75:1005354:d=2021032312:DZDT:225 mb:24 hour fcst: -76:1032154:d=2021032312:UGRD:225 mb:24 hour fcst: -77:1046134:d=2021032312:VGRD:225 mb:24 hour fcst: -78:1060506:d=2021032312:ABSV:225 mb:24 hour fcst: -79:1081223:d=2021032312:CLMR:225 mb:24 hour fcst: -80:1081402:d=2021032312:ICMR:225 mb:24 hour fcst: -81:1092874:d=2021032312:RWMR:225 mb:24 hour fcst: -82:1093053:d=2021032312:SNMR:225 mb:24 hour fcst: -83:1099978:d=2021032312:GRLE:225 mb:24 hour fcst: -84:1100289:d=2021032312:HGT:275 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_6 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_6 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_6 pgb2bfile_f024_6 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 84 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=85 -+ exglobal_atmos_products.sh[101]last=98 -+ exglobal_atmos_products.sh[102][[ 98 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 98 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 7 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 85:98 -grib tmpfileb_f024_7 -85:1122468:d=2021032312:TMP:275 mb:24 hour fcst: -86:1141088:d=2021032312:RH:275 mb:24 hour fcst: -87:1159588:d=2021032312:TCDC:275 mb:24 hour fcst: -88:1171412:d=2021032312:VVEL:275 mb:24 hour fcst: -89:1194702:d=2021032312:DZDT:275 mb:24 hour fcst: -90:1222297:d=2021032312:UGRD:275 mb:24 hour fcst: -91:1236625:d=2021032312:VGRD:275 mb:24 hour fcst: -92:1251782:d=2021032312:ABSV:275 mb:24 hour fcst: -93:1273840:d=2021032312:CLMR:275 mb:24 hour fcst: -94:1274019:d=2021032312:ICMR:275 mb:24 hour fcst: -95:1288544:d=2021032312:RWMR:275 mb:24 hour fcst: -96:1288723:d=2021032312:SNMR:275 mb:24 hour fcst: -97:1298719:d=2021032312:GRLE:275 mb:24 hour fcst: -98:1299078:d=2021032312:HGT:325 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_7 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_7 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_7 pgb2bfile_f024_7 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 98 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=99 -+ exglobal_atmos_products.sh[101]last=112 -+ exglobal_atmos_products.sh[102][[ 112 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 112 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 8 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 99:112 -grib tmpfileb_f024_8 -99:1321267:d=2021032312:TMP:325 mb:24 hour fcst: -100:1339633:d=2021032312:RH:325 mb:24 hour fcst: -101:1358656:d=2021032312:TCDC:325 mb:24 hour fcst: -102:1371711:d=2021032312:VVEL:325 mb:24 hour fcst: -103:1395969:d=2021032312:DZDT:325 mb:24 hour fcst: -104:1424213:d=2021032312:UGRD:325 mb:24 hour fcst: -105:1438852:d=2021032312:VGRD:325 mb:24 hour fcst: -106:1454140:d=2021032312:ABSV:325 mb:24 hour fcst: -107:1476462:d=2021032312:CLMR:325 mb:24 hour fcst: -108:1476641:d=2021032312:ICMR:325 mb:24 hour fcst: -109:1493144:d=2021032312:RWMR:325 mb:24 hour fcst: -110:1493323:d=2021032312:SNMR:325 mb:24 hour fcst: -111:1505110:d=2021032312:GRLE:325 mb:24 hour fcst: -112:1505546:d=2021032312:HGT:375 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_8 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_8 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_8 pgb2bfile_f024_8 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 112 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=113 -+ exglobal_atmos_products.sh[101]last=126 -+ exglobal_atmos_products.sh[102][[ 126 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 126 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 9 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 113:126 -grib tmpfileb_f024_9 -113:1527438:d=2021032312:TMP:375 mb:24 hour fcst: -114:1545738:d=2021032312:RH:375 mb:24 hour fcst: -115:1564902:d=2021032312:TCDC:375 mb:24 hour fcst: -116:1578171:d=2021032312:VVEL:375 mb:24 hour fcst: -117:1602998:d=2021032312:DZDT:375 mb:24 hour fcst: -118:1631141:d=2021032312:UGRD:375 mb:24 hour fcst: -119:1645588:d=2021032312:VGRD:375 mb:24 hour fcst: -120:1668526:d=2021032312:ABSV:375 mb:24 hour fcst: -121:1690742:d=2021032312:CLMR:375 mb:24 hour fcst: -122:1690921:d=2021032312:ICMR:375 mb:24 hour fcst: -123:1707849:d=2021032312:RWMR:375 mb:24 hour fcst: -124:1708028:d=2021032312:SNMR:375 mb:24 hour fcst: -125:1721061:d=2021032312:GRLE:375 mb:24 hour fcst: -126:1721673:d=2021032312:HGT:425 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_9 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_9 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_9 pgb2bfile_f024_9 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 126 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=127 -+ exglobal_atmos_products.sh[101]last=140 -+ exglobal_atmos_products.sh[102][[ 140 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 140 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 10 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 127:140 -grib tmpfileb_f024_10 -127:1745723:d=2021032312:TMP:425 mb:24 hour fcst: -128:1764252:d=2021032312:RH:425 mb:24 hour fcst: -129:1783621:d=2021032312:TCDC:425 mb:24 hour fcst: -130:1796387:d=2021032312:VVEL:425 mb:24 hour fcst: -131:1821466:d=2021032312:DZDT:425 mb:24 hour fcst: -132:1849662:d=2021032312:UGRD:425 mb:24 hour fcst: -133:1871495:d=2021032312:VGRD:425 mb:24 hour fcst: -134:1893945:d=2021032312:ABSV:425 mb:24 hour fcst: -135:1915556:d=2021032312:CLMR:425 mb:24 hour fcst: -136:1916212:d=2021032312:ICMR:425 mb:24 hour fcst: -137:1933014:d=2021032312:RWMR:425 mb:24 hour fcst: -138:1933368:d=2021032312:SNMR:425 mb:24 hour fcst: -139:1947720:d=2021032312:GRLE:425 mb:24 hour fcst: -140:1948594:d=2021032312:HGT:475 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_10 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_10 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_10 pgb2bfile_f024_10 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 140 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=141 -+ exglobal_atmos_products.sh[101]last=154 -+ exglobal_atmos_products.sh[102][[ 154 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 154 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 11 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 141:154 -grib tmpfileb_f024_11 -141:1972442:d=2021032312:TMP:475 mb:24 hour fcst: -142:1990916:d=2021032312:RH:475 mb:24 hour fcst: -143:2010414:d=2021032312:TCDC:475 mb:24 hour fcst: -144:2022788:d=2021032312:VVEL:475 mb:24 hour fcst: -145:2047767:d=2021032312:DZDT:475 mb:24 hour fcst: -146:2075772:d=2021032312:UGRD:475 mb:24 hour fcst: -147:2097374:d=2021032312:VGRD:475 mb:24 hour fcst: -148:2119530:d=2021032312:ABSV:475 mb:24 hour fcst: -149:2140840:d=2021032312:CLMR:475 mb:24 hour fcst: -150:2142214:d=2021032312:ICMR:475 mb:24 hour fcst: -151:2158398:d=2021032312:RWMR:475 mb:24 hour fcst: -152:2159095:d=2021032312:SNMR:475 mb:24 hour fcst: -153:2173050:d=2021032312:GRLE:475 mb:24 hour fcst: -154:2174621:d=2021032312:HGT:525 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_11 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_11 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_11 pgb2bfile_f024_11 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 154 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=155 -+ exglobal_atmos_products.sh[101]last=168 -+ exglobal_atmos_products.sh[102][[ 168 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 168 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 12 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 155:168 -grib tmpfileb_f024_12 -155:2198282:d=2021032312:TMP:525 mb:24 hour fcst: -156:2216912:d=2021032312:RH:525 mb:24 hour fcst: -157:2236627:d=2021032312:TCDC:525 mb:24 hour fcst: -158:2248663:d=2021032312:VVEL:525 mb:24 hour fcst: -159:2273739:d=2021032312:DZDT:525 mb:24 hour fcst: -160:2301361:d=2021032312:UGRD:525 mb:24 hour fcst: -161:2322492:d=2021032312:VGRD:525 mb:24 hour fcst: -162:2344394:d=2021032312:ABSV:525 mb:24 hour fcst: -163:2365425:d=2021032312:CLMR:525 mb:24 hour fcst: -164:2367486:d=2021032312:ICMR:525 mb:24 hour fcst: -165:2384474:d=2021032312:RWMR:525 mb:24 hour fcst: -166:2385639:d=2021032312:SNMR:525 mb:24 hour fcst: -167:2399992:d=2021032312:GRLE:525 mb:24 hour fcst: -168:2401534:d=2021032312:HGT:575 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_12 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_12 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_12 pgb2bfile_f024_12 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 168 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=169 -+ exglobal_atmos_products.sh[101]last=182 -+ exglobal_atmos_products.sh[102][[ 182 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 182 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 13 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 169:182 -grib tmpfileb_f024_13 -169:2425065:d=2021032312:TMP:575 mb:24 hour fcst: -170:2443953:d=2021032312:RH:575 mb:24 hour fcst: -171:2463888:d=2021032312:TCDC:575 mb:24 hour fcst: -172:2476200:d=2021032312:VVEL:575 mb:24 hour fcst: -173:2501329:d=2021032312:DZDT:575 mb:24 hour fcst: -174:2528739:d=2021032312:UGRD:575 mb:24 hour fcst: -175:2549667:d=2021032312:VGRD:575 mb:24 hour fcst: -176:2571316:d=2021032312:ABSV:575 mb:24 hour fcst: -177:2592187:d=2021032312:CLMR:575 mb:24 hour fcst: -178:2596323:d=2021032312:ICMR:575 mb:24 hour fcst: -179:2612260:d=2021032312:RWMR:575 mb:24 hour fcst: -180:2614718:d=2021032312:SNMR:575 mb:24 hour fcst: -181:2629260:d=2021032312:GRLE:575 mb:24 hour fcst: -182:2631471:d=2021032312:HGT:625 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_13 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_13 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_13 pgb2bfile_f024_13 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 182 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=183 -+ exglobal_atmos_products.sh[101]last=196 -+ exglobal_atmos_products.sh[102][[ 196 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 196 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 14 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 183:196 -grib tmpfileb_f024_14 -183:2655640:d=2021032312:TMP:625 mb:24 hour fcst: -184:2674770:d=2021032312:RH:625 mb:24 hour fcst: -185:2694945:d=2021032312:TCDC:625 mb:24 hour fcst: -186:2706785:d=2021032312:VVEL:625 mb:24 hour fcst: -187:2732271:d=2021032312:DZDT:625 mb:24 hour fcst: -188:2759687:d=2021032312:UGRD:625 mb:24 hour fcst: -189:2780716:d=2021032312:VGRD:625 mb:24 hour fcst: -190:2802164:d=2021032312:ABSV:625 mb:24 hour fcst: -191:2822974:d=2021032312:CLMR:625 mb:24 hour fcst: -192:2828032:d=2021032312:ICMR:625 mb:24 hour fcst: -193:2841233:d=2021032312:RWMR:625 mb:24 hour fcst: -194:2845487:d=2021032312:SNMR:625 mb:24 hour fcst: -195:2860963:d=2021032312:GRLE:625 mb:24 hour fcst: -196:2863462:d=2021032312:HGT:675 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_14 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_14 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_14 pgb2bfile_f024_14 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 196 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=197 -+ exglobal_atmos_products.sh[101]last=210 -+ exglobal_atmos_products.sh[102][[ 210 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 210 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 15 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 197:210 -grib tmpfileb_f024_15 -197:2887769:d=2021032312:TMP:675 mb:24 hour fcst: -198:2907217:d=2021032312:RH:675 mb:24 hour fcst: -199:2927509:d=2021032312:TCDC:675 mb:24 hour fcst: -200:2939223:d=2021032312:VVEL:675 mb:24 hour fcst: -201:2964865:d=2021032312:DZDT:675 mb:24 hour fcst: -202:2992277:d=2021032312:UGRD:675 mb:24 hour fcst: -203:3013365:d=2021032312:VGRD:675 mb:24 hour fcst: -204:3035044:d=2021032312:ABSV:675 mb:24 hour fcst: -205:3056084:d=2021032312:CLMR:675 mb:24 hour fcst: -206:3061776:d=2021032312:ICMR:675 mb:24 hour fcst: -207:3073671:d=2021032312:RWMR:675 mb:24 hour fcst: -208:3078601:d=2021032312:SNMR:675 mb:24 hour fcst: -209:3093225:d=2021032312:GRLE:675 mb:24 hour fcst: -210:3095348:d=2021032312:HGT:725 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_15 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_15 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_15 pgb2bfile_f024_15 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 210 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=211 -+ exglobal_atmos_products.sh[101]last=224 -+ exglobal_atmos_products.sh[102][[ 224 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 224 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 16 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 211:224 -grib tmpfileb_f024_16 -211:3119516:d=2021032312:TMP:725 mb:24 hour fcst: -212:3139236:d=2021032312:RH:725 mb:24 hour fcst: -213:3159578:d=2021032312:TCDC:725 mb:24 hour fcst: -214:3171272:d=2021032312:VVEL:725 mb:24 hour fcst: -215:3196960:d=2021032312:DZDT:725 mb:24 hour fcst: -216:3224225:d=2021032312:UGRD:725 mb:24 hour fcst: -217:3245048:d=2021032312:VGRD:725 mb:24 hour fcst: -218:3266678:d=2021032312:ABSV:725 mb:24 hour fcst: -219:3287563:d=2021032312:CLMR:725 mb:24 hour fcst: -220:3295499:d=2021032312:ICMR:725 mb:24 hour fcst: -221:3307383:d=2021032312:RWMR:725 mb:24 hour fcst: -222:3313115:d=2021032312:SNMR:725 mb:24 hour fcst: -223:3326930:d=2021032312:GRLE:725 mb:24 hour fcst: -224:3328864:d=2021032312:HGT:775 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_16 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_16 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_16 pgb2bfile_f024_16 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 224 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=225 -+ exglobal_atmos_products.sh[101]last=238 -+ exglobal_atmos_products.sh[102][[ 238 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 238 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 17 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 225:238 -grib tmpfileb_f024_17 -225:3353336:d=2021032312:TMP:775 mb:24 hour fcst: -226:3373429:d=2021032312:RH:775 mb:24 hour fcst: -227:3393962:d=2021032312:TCDC:775 mb:24 hour fcst: -228:3406890:d=2021032312:VVEL:775 mb:24 hour fcst: -229:3432580:d=2021032312:DZDT:775 mb:24 hour fcst: -230:3459522:d=2021032312:UGRD:775 mb:24 hour fcst: -231:3480382:d=2021032312:VGRD:775 mb:24 hour fcst: -232:3502067:d=2021032312:ABSV:775 mb:24 hour fcst: -233:3523026:d=2021032312:CLMR:775 mb:24 hour fcst: -234:3533536:d=2021032312:ICMR:775 mb:24 hour fcst: -235:3545015:d=2021032312:RWMR:775 mb:24 hour fcst: -236:3551923:d=2021032312:SNMR:775 mb:24 hour fcst: -237:3565067:d=2021032312:GRLE:775 mb:24 hour fcst: -238:3567399:d=2021032312:HGT:825 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_17 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_17 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_17 pgb2bfile_f024_17 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 238 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=239 -+ exglobal_atmos_products.sh[101]last=252 -+ exglobal_atmos_products.sh[102][[ 252 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 252 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 18 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 239:252 -grib tmpfileb_f024_18 -239:3591884:d=2021032312:TMP:825 mb:24 hour fcst: -240:3612703:d=2021032312:RH:825 mb:24 hour fcst: -241:3633349:d=2021032312:TCDC:825 mb:24 hour fcst: -242:3647466:d=2021032312:VVEL:825 mb:24 hour fcst: -243:3673049:d=2021032312:DZDT:825 mb:24 hour fcst: -244:3699798:d=2021032312:UGRD:825 mb:24 hour fcst: -245:3720956:d=2021032312:VGRD:825 mb:24 hour fcst: -246:3742769:d=2021032312:ABSV:825 mb:24 hour fcst: -247:3763799:d=2021032312:CLMR:825 mb:24 hour fcst: -248:3777184:d=2021032312:ICMR:825 mb:24 hour fcst: -249:3787606:d=2021032312:RWMR:825 mb:24 hour fcst: -250:3797265:d=2021032312:SNMR:825 mb:24 hour fcst: -251:3809561:d=2021032312:GRLE:825 mb:24 hour fcst: -252:3812339:d=2021032312:HGT:875 mb:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_18 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_18 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_18 pgb2bfile_f024_18 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 252 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=253 -+ exglobal_atmos_products.sh[101]last=266 -+ exglobal_atmos_products.sh[102][[ 266 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 266 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 19 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 253:266 -grib tmpfileb_f024_19 -253:3837270:d=2021032312:TMP:875 mb:24 hour fcst: -254:3858035:d=2021032312:RH:875 mb:24 hour fcst: -255:3878036:d=2021032312:TCDC:875 mb:24 hour fcst: -256:3892119:d=2021032312:VVEL:875 mb:24 hour fcst: -257:3917240:d=2021032312:DZDT:875 mb:24 hour fcst: -258:3943633:d=2021032312:UGRD:875 mb:24 hour fcst: -259:3964861:d=2021032312:VGRD:875 mb:24 hour fcst: -260:3986661:d=2021032312:ABSV:875 mb:24 hour fcst: -261:4007870:d=2021032312:CLMR:875 mb:24 hour fcst: -262:4020473:d=2021032312:ICMR:875 mb:24 hour fcst: -263:4030372:d=2021032312:RWMR:875 mb:24 hour fcst: -264:4042132:d=2021032312:SNMR:875 mb:24 hour fcst: -265:4053156:d=2021032312:GRLE:875 mb:24 hour fcst: -266:4055789:d=2021032312:SOILL:0-0.1 m below ground:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_19 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_19 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_19 pgb2bfile_f024_19 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 266 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=267 -+ exglobal_atmos_products.sh[101]last=280 -+ exglobal_atmos_products.sh[102][[ 280 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 280 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 20 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 267:280 -grib tmpfileb_f024_20 -267:4064586:d=2021032312:SOILL:0.1-0.4 m below ground:24 hour fcst: -268:4073480:d=2021032312:SOILL:0.4-1 m below ground:24 hour fcst: -269:4082384:d=2021032312:SOILL:1-2 m below ground:24 hour fcst: -270:4091460:d=2021032312:CNWAT:surface:24 hour fcst: -271:4097045:d=2021032312:ICETK:surface:24 hour fcst: -272:4100459:d=2021032312:DUVB:surface:18-24 hour ave fcst: -273:4121149:d=2021032312:CDUVB:surface:18-24 hour ave fcst: -274:4136997:d=2021032312:TMP:305 m above mean sea level:24 hour fcst: -275:4155331:d=2021032312:UGRD:305 m above mean sea level:24 hour fcst: -276:4174503:d=2021032312:VGRD:305 m above mean sea level:24 hour fcst: -277:4194089:d=2021032312:TMP:457 m above mean sea level:24 hour fcst: -278:4213595:d=2021032312:UGRD:457 m above mean sea level:24 hour fcst: -279:4233825:d=2021032312:VGRD:457 m above mean sea level:24 hour fcst: -280:4254380:d=2021032312:TMP:610 m above mean sea level:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_20 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_20 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_20 pgb2bfile_f024_20 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 280 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=281 -+ exglobal_atmos_products.sh[101]last=294 -+ exglobal_atmos_products.sh[102][[ 294 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 294 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 21 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 281:294 -grib tmpfileb_f024_21 -281:4274643:d=2021032312:UGRD:610 m above mean sea level:24 hour fcst: -282:4295495:d=2021032312:VGRD:610 m above mean sea level:24 hour fcst: -283:4316686:d=2021032312:TMP:914 m above mean sea level:24 hour fcst: -284:4337576:d=2021032312:UGRD:914 m above mean sea level:24 hour fcst: -285:4358949:d=2021032312:VGRD:914 m above mean sea level:24 hour fcst: -286:4380801:d=2021032312:TMP:4572 m above mean sea level:24 hour fcst: -287:4401975:d=2021032312:UGRD:4572 m above mean sea level:24 hour fcst: -288:4425369:d=2021032312:VGRD:4572 m above mean sea level:24 hour fcst: -289:4449339:d=2021032312:TMP:60-30 mb above ground:24 hour fcst: -290:4470160:d=2021032312:RH:60-30 mb above ground:24 hour fcst: -291:4488858:d=2021032312:SPFH:60-30 mb above ground:24 hour fcst: -292:4517605:d=2021032312:UGRD:60-30 mb above ground:24 hour fcst: -293:4539232:d=2021032312:VGRD:60-30 mb above ground:24 hour fcst: -294:4561296:d=2021032312:TMP:90-60 mb above ground:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_21 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_21 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_21 pgb2bfile_f024_21 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 294 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=295 -+ exglobal_atmos_products.sh[101]last=308 -+ exglobal_atmos_products.sh[102][[ 308 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 308 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 22 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 295:308 -grib tmpfileb_f024_22 -295:4582142:d=2021032312:RH:90-60 mb above ground:24 hour fcst: -296:4601255:d=2021032312:SPFH:90-60 mb above ground:24 hour fcst: -297:4630170:d=2021032312:UGRD:90-60 mb above ground:24 hour fcst: -298:4651556:d=2021032312:VGRD:90-60 mb above ground:24 hour fcst: -299:4673567:d=2021032312:TMP:120-90 mb above ground:24 hour fcst: -300:4694511:d=2021032312:RH:120-90 mb above ground:24 hour fcst: -301:4713908:d=2021032312:SPFH:120-90 mb above ground:24 hour fcst: -302:4742806:d=2021032312:UGRD:120-90 mb above ground:24 hour fcst: -303:4763935:d=2021032312:VGRD:120-90 mb above ground:24 hour fcst: -304:4785748:d=2021032312:TMP:150-120 mb above ground:24 hour fcst: -305:4806688:d=2021032312:RH:150-120 mb above ground:24 hour fcst: -306:4826382:d=2021032312:SPFH:150-120 mb above ground:24 hour fcst: -307:4855516:d=2021032312:UGRD:150-120 mb above ground:24 hour fcst: -308:4876551:d=2021032312:VGRD:150-120 mb above ground:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_22 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_22 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_22 pgb2bfile_f024_22 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 308 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=309 -+ exglobal_atmos_products.sh[101]last=322 -+ exglobal_atmos_products.sh[102][[ 322 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 322 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 23 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 309:322 -grib tmpfileb_f024_23 -309:4898254:d=2021032312:TMP:180-150 mb above ground:24 hour fcst: -310:4919366:d=2021032312:RH:180-150 mb above ground:24 hour fcst: -311:4939380:d=2021032312:SPFH:180-150 mb above ground:24 hour fcst: -312:4968614:d=2021032312:UGRD:180-150 mb above ground:24 hour fcst: -313:4989680:d=2021032312:VGRD:180-150 mb above ground:24 hour fcst: -314:5011352:d=2021032312:UGRD:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -315:5022034:d=2021032312:VGRD:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -316:5032576:d=2021032312:TMP:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -317:5047152:d=2021032312:HGT:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -318:5063868:d=2021032312:PRES:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -319:5080710:d=2021032312:VWSH:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -320:5091613:d=2021032312:UGRD:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -321:5102677:d=2021032312:VGRD:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -322:5117176:d=2021032312:TMP:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_23 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_23 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_23 pgb2bfile_f024_23 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 322 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=323 -+ exglobal_atmos_products.sh[101]last=336 -+ exglobal_atmos_products.sh[102][[ 336 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 336 tmpfileb_f024 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 24 -eq 24 ]] -+ exglobal_atmos_products.sh[118]last=349 -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f024 -for 323:349 -grib tmpfileb_f024_24 -323:5132541:d=2021032312:HGT:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -324:5150245:d=2021032312:PRES:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -325:5167935:d=2021032312:VWSH:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -326:5179301:d=2021032312:UGRD:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -327:5190645:d=2021032312:VGRD:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -328:5201959:d=2021032312:TMP:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -329:5217858:d=2021032312:HGT:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -330:5236594:d=2021032312:PRES:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -331:5255352:d=2021032312:VWSH:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -332:5267641:d=2021032312:UGRD:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -333:5277798:d=2021032312:VGRD:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -334:5289089:d=2021032312:TMP:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -335:5304883:d=2021032312:HGT:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -336:5323409:d=2021032312:PRES:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -337:5341862:d=2021032312:VWSH:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -338:5354170:d=2021032312:UGRD:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -339:5365274:d=2021032312:VGRD:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -340:5376237:d=2021032312:TMP:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -341:5391204:d=2021032312:HGT:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -342:5409060:d=2021032312:PRES:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -343:5426874:d=2021032312:VWSH:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -344:5439014:d=2021032312:UGRD:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -345:5450079:d=2021032312:VGRD:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -346:5461107:d=2021032312:TMP:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -347:5476164:d=2021032312:HGT:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -348:5493998:d=2021032312:PRES:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -349:5511674:d=2021032312:VWSH:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f024_24 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f024_24 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_24 pgb2bfile_f024_24 0p25:0p50:1p00' -+ exglobal_atmos_products.sh[133][[ 349 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( pproc = iproc+1 )) -+ exglobal_atmos_products.sh[134](( pproc < nproc )) -+ exglobal_atmos_products.sh[137]break -+ exglobal_atmos_products.sh[142]/work2/noaa/global/mterry/global-workflow_forked/ush/run_mpmd.sh /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601/poescript -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ run_mpmd.sh[31]source /work2/noaa/global/mterry/global-workflow_forked/ush/preamble.sh -++ preamble.sh[20]set +x -Begin run_mpmd.sh at Tue Jul 29 02:25:35 UTC 2025 -+ run_mpmd.sh[33]cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601/poescript -+ run_mpmd.sh[36][[ YES != \Y\E\S ]] -+ run_mpmd.sh[46]export OMP_NUM_THREADS=1 -+ run_mpmd.sh[46]OMP_NUM_THREADS=1 -++ run_mpmd.sh[49]wc -l -+ run_mpmd.sh[49]nprocs=24 -+ run_mpmd.sh[52]mpmd_cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601/mpmd_cmdfile -+ run_mpmd.sh[53][[ -s /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601/mpmd_cmdfile ]] -+ run_mpmd.sh[55]cat - INFO: Executing MPMD job, STDOUT redirected for each process separately - INFO: On failure, logs for each job will be available in /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601/mpmd.proc_num.out - INFO: The proc_num corresponds to the line in '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601/mpmd_cmdfile' -+ run_mpmd.sh[61][[ srun -l --export=ALL --hint=nomultithread =~ ^srun.* ]] -+ run_mpmd.sh[65]nm=0 -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '0 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_1 pgb2bfile_f024_1 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '1 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_2 pgb2bfile_f024_2 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '2 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_3 pgb2bfile_f024_3 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '3 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_4 pgb2bfile_f024_4 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '4 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_5 pgb2bfile_f024_5 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '5 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_6 pgb2bfile_f024_6 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '6 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_7 pgb2bfile_f024_7 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '7 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_8 pgb2bfile_f024_8 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '8 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_9 pgb2bfile_f024_9 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '9 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_10 pgb2bfile_f024_10 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '10 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_11 pgb2bfile_f024_11 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '11 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_12 pgb2bfile_f024_12 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '12 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_13 pgb2bfile_f024_13 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '13 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_14 pgb2bfile_f024_14 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '14 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_15 pgb2bfile_f024_15 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '15 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_16 pgb2bfile_f024_16 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '16 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_17 pgb2bfile_f024_17 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '17 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_18 pgb2bfile_f024_18 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '18 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_19 pgb2bfile_f024_19 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '19 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_20 pgb2bfile_f024_20 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '20 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_21 pgb2bfile_f024_21 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '21 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_22 pgb2bfile_f024_22 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '22 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_23 pgb2bfile_f024_23 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '23 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f024_24 pgb2bfile_f024_24 0p25:0p50:1p00' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[72]set +e -+ run_mpmd.sh[74]srun -l --export=ALL --hint=nomultithread --multi-prog --output=mpmd.%j.%t.out -n 24 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601/mpmd_cmdfile -+ run_mpmd.sh[75]err=0 -+ run_mpmd.sh[76]set_strict -+ preamble.sh[35][[ YES == \Y\E\S ]] -+ preamble.sh[37]set -eu -+ run_mpmd.sh[103][[ 0 -eq 0 ]] -+ run_mpmd.sh[104]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601/mpmd_cmdfile -++ run_mpmd.sh[105]find . -name 'mpmd.*.out' -+ run_mpmd.sh[105]out_files='./mpmd.5951061.17.out -./mpmd.5951061.0.out -./mpmd.5951061.6.out -./mpmd.5951061.15.out -./mpmd.5951061.9.out -./mpmd.5951061.2.out -./mpmd.5951061.3.out -./mpmd.5951061.23.out -./mpmd.5951061.4.out -./mpmd.5951061.22.out -./mpmd.5951061.1.out -./mpmd.5951061.7.out -./mpmd.5951061.21.out -./mpmd.5951061.12.out -./mpmd.5951061.19.out -./mpmd.5951061.8.out -./mpmd.5951061.16.out -./mpmd.5951061.18.out -./mpmd.5951061.13.out -./mpmd.5951061.10.out -./mpmd.5951061.20.out -./mpmd.5951061.14.out -./mpmd.5951061.5.out -./mpmd.5951061.11.out' -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.17.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.17.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.0.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.0.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.6.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.6.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.15.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.15.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.9.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.9.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.2.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.2.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.3.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.3.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.23.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.23.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.4.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.4.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.22.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.22.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.1.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.1.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.7.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.7.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.21.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.21.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.12.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.12.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.19.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.19.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.8.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.8.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.16.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.16.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.18.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.18.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.13.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.13.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.10.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.10.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.20.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.20.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.14.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.14.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.5.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.5.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.11.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.11.out -+ run_mpmd.sh[110]cat mpmd.out -17: + bash[8]'[' -z '' ']' -17: + bash[9]case "$-" in -17: + bash[12]__lmod_vx=x -17: + bash[16]'[' -n x ']' -17: + bash[16]set +x -17: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -17: Shell debugging restarted -17: + bash[224]unset __lmod_vx -17: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_18 -17: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_18 -17: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -17: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -17: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -17: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -17: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -17: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -17: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -17: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -17: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -17: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -17: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -17: + interp_atmos_master.sh[31]IFS=: -17: + interp_atmos_master.sh[31]read -ra grids -17: + interp_atmos_master.sh[33]output_grids= -17: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -17: + interp_atmos_master.sh[35]gridopt=grid0p25 -17: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_18_0p25' -17: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -17: + interp_atmos_master.sh[35]gridopt=grid0p50 -17: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_18_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_18_0p50' -17: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -17: + interp_atmos_master.sh[35]gridopt=grid1p00 -17: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_18_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_18_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_18_1p00' -17: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_18 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_18_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_18_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_18_1p00 -17: 1:0:d=2021032312:TMP:825 mb:24 hour fcst: -17: 2:20819:d=2021032312:RH:825 mb:24 hour fcst: -17: 3:41465:d=2021032312:TCDC:825 mb:24 hour fcst: -17: 4:55582:d=2021032312:VVEL:825 mb:24 hour fcst: -17: 5:81165:d=2021032312:DZDT:825 mb:24 hour fcst: -17: 6:107914:d=2021032312:UGRD:825 mb:24 hour fcst: -17: 7:129072:d=2021032312:VGRD:825 mb:24 hour fcst: -17: 8:150885:d=2021032312:ABSV:825 mb:24 hour fcst: -17: 9:171915:d=2021032312:CLMR:825 mb:24 hour fcst: -17: 10:185300:d=2021032312:ICMR:825 mb:24 hour fcst: -17: 11:195722:d=2021032312:RWMR:825 mb:24 hour fcst: -17: 12:205381:d=2021032312:SNMR:825 mb:24 hour fcst: -17: 13:217677:d=2021032312:GRLE:825 mb:24 hour fcst: -17: 14:220455:d=2021032312:HGT:875 mb:24 hour fcst: -17: + interp_atmos_master.sh[47]export err=0 -17: + interp_atmos_master.sh[47]err=0 -17: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -17: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -17: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_18_0p25 -17: + product_functions.sh[5]local filename=pgb2bfile_f024_18_0p25 -17: + product_functions.sh[6]wgrib2 pgb2bfile_f024_18_0p25 -not_if :RH: -grib pgb2bfile_f024_18_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_18_0p25.new -17: 1:0:d=2021032312:TMP:825 mb:24 hour fcst: -17: 2:488647:d=2021032312:RH:825 mb:24 hour fcst: -17: 3:970808:d=2021032312:TCDC:825 mb:24 hour fcst: -17: 4:1257405:d=2021032312:VVEL:825 mb:24 hour fcst: -17: 5:1939638:d=2021032312:DZDT:825 mb:24 hour fcst: -17: 6:2678254:d=2021032312:UGRD:825 mb:24 hour fcst: -17: 7:3187209:d=2021032312:VGRD:825 mb:24 hour fcst: -17: 8:3726165:d=2021032312:ABSV:825 mb:24 hour fcst: -17: 9:4233182:d=2021032312:CLMR:825 mb:24 hour fcst: -17: 10:4526038:d=2021032312:ICMR:825 mb:24 hour fcst: -17: 11:4770581:d=2021032312:RWMR:825 mb:24 hour fcst: -17: 12:4952227:d=2021032312:SNMR:825 mb:24 hour fcst: -17: 13:5227923:d=2021032312:GRLE:825 mb:24 hour fcst: -17: 14:5281479:d=2021032312:HGT:875 mb:24 hour fcst: -17: + product_functions.sh[10]rc=0 -17: + product_functions.sh[11](( rc == 0 )) -17: + product_functions.sh[11]mv pgb2bfile_f024_18_0p25.new pgb2bfile_f024_18_0p25 -17: + product_functions.sh[12]return 0 -17: + interp_atmos_master.sh[56]export err=0 -17: + interp_atmos_master.sh[56]err=0 -17: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -17: ++ interp_atmos_master.sh[62]wc -l -17: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_18_0p25 -match 'LAND|ICEC' -17: + interp_atmos_master.sh[62]var_count=0 -17: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -17: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -17: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_18_0p50 -17: + product_functions.sh[5]local filename=pgb2bfile_f024_18_0p50 -17: + product_functions.sh[6]wgrib2 pgb2bfile_f024_18_0p50 -not_if :RH: -grib pgb2bfile_f024_18_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_18_0p50.new -17: 1:0:d=2021032312:TMP:825 mb:24 hour fcst: -17: 2:168703:d=2021032312:RH:825 mb:24 hour fcst: -17: 3:335418:d=2021032312:TCDC:825 mb:24 hour fcst: -17: 4:442802:d=2021032312:VVEL:825 mb:24 hour fcst: -17: 5:675023:d=2021032312:DZDT:825 mb:24 hour fcst: -17: 6:917596:d=2021032312:UGRD:825 mb:24 hour fcst: -17: 7:1092897:d=2021032312:VGRD:825 mb:24 hour fcst: -17: 8:1277421:d=2021032312:ABSV:825 mb:24 hour fcst: -17: 9:1452250:d=2021032312:CLMR:825 mb:24 hour fcst: -17: 10:1559954:d=2021032312:ICMR:825 mb:24 hour fcst: -17: 11:1646206:d=2021032312:RWMR:825 mb:24 hour fcst: -17: 12:1714002:d=2021032312:SNMR:825 mb:24 hour fcst: -17: 13:1812095:d=2021032312:GRLE:825 mb:24 hour fcst: -17: 14:1832001:d=2021032312:HGT:875 mb:24 hour fcst: -17: + product_functions.sh[10]rc=0 -17: + product_functions.sh[11](( rc == 0 )) -17: + product_functions.sh[11]mv pgb2bfile_f024_18_0p50.new pgb2bfile_f024_18_0p50 -17: + product_functions.sh[12]return 0 -17: + interp_atmos_master.sh[56]export err=0 -17: + interp_atmos_master.sh[56]err=0 -17: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -17: ++ interp_atmos_master.sh[62]wc -l -17: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_18_0p50 -match 'LAND|ICEC' -17: + interp_atmos_master.sh[62]var_count=0 -17: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -17: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -17: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_18_1p00 -17: + product_functions.sh[5]local filename=pgb2bfile_f024_18_1p00 -17: + product_functions.sh[6]wgrib2 pgb2bfile_f024_18_1p00 -not_if :RH: -grib pgb2bfile_f024_18_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_18_1p00.new -17: 1:0:d=2021032312:TMP:825 mb:24 hour fcst: -17: 2:54780:d=2021032312:RH:825 mb:24 hour fcst: -17: 3:108722:d=2021032312:TCDC:825 mb:24 hour fcst: -17: 4:146005:d=2021032312:VVEL:825 mb:24 hour fcst: -17: 5:217190:d=2021032312:DZDT:825 mb:24 hour fcst: -17: 6:291047:d=2021032312:UGRD:825 mb:24 hour fcst: -17: 7:347252:d=2021032312:VGRD:825 mb:24 hour fcst: -17: 8:405867:d=2021032312:ABSV:825 mb:24 hour fcst: -17: 9:461753:d=2021032312:CLMR:825 mb:24 hour fcst: -17: 10:497370:d=2021032312:ICMR:825 mb:24 hour fcst: -17: 11:525161:d=2021032312:RWMR:825 mb:24 hour fcst: -17: 12:549360:d=2021032312:SNMR:825 mb:24 hour fcst: -17: 13:582020:d=2021032312:GRLE:825 mb:24 hour fcst: -17: 14:589232:d=2021032312:HGT:875 mb:24 hour fcst: -17: + product_functions.sh[10]rc=0 -17: + product_functions.sh[11](( rc == 0 )) -17: + product_functions.sh[11]mv pgb2bfile_f024_18_1p00.new pgb2bfile_f024_18_1p00 -17: + product_functions.sh[12]return 0 -17: + interp_atmos_master.sh[56]export err=0 -17: + interp_atmos_master.sh[56]err=0 -17: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -17: ++ interp_atmos_master.sh[62]wc -l -17: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_18_1p00 -match 'LAND|ICEC' -17: + interp_atmos_master.sh[62]var_count=0 -17: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -17: + interp_atmos_master.sh[73]exit 0 - 0: + bash[8]'[' -z '' ']' - 0: + bash[9]case "$-" in - 0: + bash[12]__lmod_vx=x - 0: + bash[16]'[' -n x ']' - 0: + bash[16]set +x - 0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 0: Shell debugging restarted - 0: + bash[224]unset __lmod_vx - 0: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_1 - 0: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_1 - 0: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 - 0: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 0: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 0: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 0: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 0: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 0: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 0: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 0: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 0: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 0: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 0: + interp_atmos_master.sh[31]IFS=: - 0: + interp_atmos_master.sh[31]read -ra grids - 0: + interp_atmos_master.sh[33]output_grids= - 0: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 0: + interp_atmos_master.sh[35]gridopt=grid0p25 - 0: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_1_0p25' - 0: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 0: + interp_atmos_master.sh[35]gridopt=grid0p50 - 0: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_1_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_1_0p50' - 0: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 0: + interp_atmos_master.sh[35]gridopt=grid1p00 - 0: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_1_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_1_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_1_1p00' - 0: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_1 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_1_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_1_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_1_1p00 - 0: 1:0:d=2021032312:HGT:1 mb:24 hour fcst: - 0: 2:18247:d=2021032312:TMP:1 mb:24 hour fcst: - 0: 3:34112:d=2021032312:RH:1 mb:24 hour fcst: - 0: 4:43938:d=2021032312:UGRD:1 mb:24 hour fcst: - 0: 5:62539:d=2021032312:VGRD:1 mb:24 hour fcst: - 0: 6:80240:d=2021032312:ABSV:1 mb:24 hour fcst: - 0: 7:96594:d=2021032312:O3MR:1 mb:24 hour fcst: - 0: 8:117662:d=2021032312:HGT:2 mb:24 hour fcst: - 0: 9:137974:d=2021032312:TMP:2 mb:24 hour fcst: - 0: 10:154578:d=2021032312:RH:2 mb:24 hour fcst: - 0: 11:163159:d=2021032312:UGRD:2 mb:24 hour fcst: - 0: 12:174744:d=2021032312:VGRD:2 mb:24 hour fcst: - 0: 13:193090:d=2021032312:ABSV:2 mb:24 hour fcst: - 0: 14:210173:d=2021032312:O3MR:2 mb:24 hour fcst: - 0: + interp_atmos_master.sh[47]export err=0 - 0: + interp_atmos_master.sh[47]err=0 - 0: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 0: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 0: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_1_0p25 - 0: + product_functions.sh[5]local filename=pgb2bfile_f024_1_0p25 - 0: + product_functions.sh[6]wgrib2 pgb2bfile_f024_1_0p25 -not_if :RH: -grib pgb2bfile_f024_1_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_1_0p25.new - 0: 1:0:d=2021032312:HGT:1 mb:24 hour fcst: - 0: 2:409041:d=2021032312:TMP:1 mb:24 hour fcst: - 0: 3:757856:d=2021032312:RH:1 mb:24 hour fcst: - 0: 4:1014158:d=2021032312:UGRD:1 mb:24 hour fcst: - 0: 5:1443579:d=2021032312:VGRD:1 mb:24 hour fcst: - 0: 6:1847799:d=2021032312:ABSV:1 mb:24 hour fcst: - 0: 7:2194918:d=2021032312:O3MR:1 mb:24 hour fcst: - 0: 8:2722844:d=2021032312:HGT:2 mb:24 hour fcst: - 0: 9:3210810:d=2021032312:TMP:2 mb:24 hour fcst: - 0: 10:3583188:d=2021032312:RH:2 mb:24 hour fcst: - 0: 11:3831582:d=2021032312:UGRD:2 mb:24 hour fcst: - 0: 12:4290614:d=2021032312:VGRD:2 mb:24 hour fcst: - 0: 13:4714619:d=2021032312:ABSV:2 mb:24 hour fcst: - 0: 14:5084253:d=2021032312:O3MR:2 mb:24 hour fcst: - 0: + product_functions.sh[10]rc=0 - 0: + product_functions.sh[11](( rc == 0 )) - 0: + product_functions.sh[11]mv pgb2bfile_f024_1_0p25.new pgb2bfile_f024_1_0p25 - 0: + product_functions.sh[12]return 0 - 0: + interp_atmos_master.sh[56]export err=0 - 0: + interp_atmos_master.sh[56]err=0 - 0: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 0: ++ interp_atmos_master.sh[62]wc -l - 0: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_1_0p25 -match 'LAND|ICEC' - 0: + interp_atmos_master.sh[62]var_count=0 - 0: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 0: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 0: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_1_0p50 - 0: + product_functions.sh[5]local filename=pgb2bfile_f024_1_0p50 - 0: + product_functions.sh[6]wgrib2 pgb2bfile_f024_1_0p50 -not_if :RH: -grib pgb2bfile_f024_1_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_1_0p50.new - 0: 1:0:d=2021032312:HGT:1 mb:24 hour fcst: - 0: 2:140441:d=2021032312:TMP:1 mb:24 hour fcst: - 0: 3:258729:d=2021032312:RH:1 mb:24 hour fcst: - 0: 4:330812:d=2021032312:UGRD:1 mb:24 hour fcst: - 0: 5:477805:d=2021032312:VGRD:1 mb:24 hour fcst: - 0: 6:616065:d=2021032312:ABSV:1 mb:24 hour fcst: - 0: 7:734378:d=2021032312:O3MR:1 mb:24 hour fcst: - 0: 8:914913:d=2021032312:HGT:2 mb:24 hour fcst: - 0: 9:1080767:d=2021032312:TMP:2 mb:24 hour fcst: - 0: 10:1207181:d=2021032312:RH:2 mb:24 hour fcst: - 0: 11:1274837:d=2021032312:UGRD:2 mb:24 hour fcst: - 0: 12:1433534:d=2021032312:VGRD:2 mb:24 hour fcst: - 0: 13:1579050:d=2021032312:ABSV:2 mb:24 hour fcst: - 0: 14:1706674:d=2021032312:O3MR:2 mb:24 hour fcst: - 0: + product_functions.sh[10]rc=0 - 0: + product_functions.sh[11](( rc == 0 )) - 0: + product_functions.sh[11]mv pgb2bfile_f024_1_0p50.new pgb2bfile_f024_1_0p50 - 0: + product_functions.sh[12]return 0 - 0: + interp_atmos_master.sh[56]export err=0 - 0: + interp_atmos_master.sh[56]err=0 - 0: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 0: ++ interp_atmos_master.sh[62]wc -l - 0: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_1_0p50 -match 'LAND|ICEC' - 0: + interp_atmos_master.sh[62]var_count=0 - 0: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 0: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 0: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_1_1p00 - 0: + product_functions.sh[5]local filename=pgb2bfile_f024_1_1p00 - 0: + product_functions.sh[6]wgrib2 pgb2bfile_f024_1_1p00 -not_if :RH: -grib pgb2bfile_f024_1_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_1_1p00.new - 0: 1:0:d=2021032312:HGT:1 mb:24 hour fcst: - 0: 2:46183:d=2021032312:TMP:1 mb:24 hour fcst: - 0: 3:86076:d=2021032312:RH:1 mb:24 hour fcst: - 0: 4:109273:d=2021032312:UGRD:1 mb:24 hour fcst: - 0: 5:157124:d=2021032312:VGRD:1 mb:24 hour fcst: - 0: 6:202965:d=2021032312:ABSV:1 mb:24 hour fcst: - 0: 7:243148:d=2021032312:O3MR:1 mb:24 hour fcst: - 0: 8:300344:d=2021032312:HGT:2 mb:24 hour fcst: - 0: 9:353271:d=2021032312:TMP:2 mb:24 hour fcst: - 0: 10:395489:d=2021032312:RH:2 mb:24 hour fcst: - 0: 11:415562:d=2021032312:UGRD:2 mb:24 hour fcst: - 0: 12:466757:d=2021032312:VGRD:2 mb:24 hour fcst: - 0: 13:514317:d=2021032312:ABSV:2 mb:24 hour fcst: - 0: 14:556994:d=2021032312:O3MR:2 mb:24 hour fcst: - 0: + product_functions.sh[10]rc=0 - 0: + product_functions.sh[11](( rc == 0 )) - 0: + product_functions.sh[11]mv pgb2bfile_f024_1_1p00.new pgb2bfile_f024_1_1p00 - 0: + product_functions.sh[12]return 0 - 0: + interp_atmos_master.sh[56]export err=0 - 0: + interp_atmos_master.sh[56]err=0 - 0: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 0: ++ interp_atmos_master.sh[62]wc -l - 0: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_1_1p00 -match 'LAND|ICEC' - 0: + interp_atmos_master.sh[62]var_count=0 - 0: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 0: + interp_atmos_master.sh[73]exit 0 - 6: + bash[8]'[' -z '' ']' - 6: + bash[9]case "$-" in - 6: + bash[12]__lmod_vx=x - 6: + bash[16]'[' -n x ']' - 6: + bash[16]set +x - 6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 6: Shell debugging restarted - 6: + bash[224]unset __lmod_vx - 6: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_7 - 6: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_7 - 6: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 - 6: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 6: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 6: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 6: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 6: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 6: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 6: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 6: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 6: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 6: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 6: + interp_atmos_master.sh[31]IFS=: - 6: + interp_atmos_master.sh[31]read -ra grids - 6: + interp_atmos_master.sh[33]output_grids= - 6: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 6: + interp_atmos_master.sh[35]gridopt=grid0p25 - 6: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_7_0p25' - 6: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 6: + interp_atmos_master.sh[35]gridopt=grid0p50 - 6: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_7_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_7_0p50' - 6: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 6: + interp_atmos_master.sh[35]gridopt=grid1p00 - 6: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_7_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_7_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_7_1p00' - 6: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_7 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_7_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_7_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_7_1p00 - 6: 1:0:d=2021032312:TMP:275 mb:24 hour fcst: - 6: 2:18620:d=2021032312:RH:275 mb:24 hour fcst: - 6: 3:37120:d=2021032312:TCDC:275 mb:24 hour fcst: - 6: 4:48944:d=2021032312:VVEL:275 mb:24 hour fcst: - 6: 5:72234:d=2021032312:DZDT:275 mb:24 hour fcst: - 6: 6:99829:d=2021032312:UGRD:275 mb:24 hour fcst: - 6: 7:114157:d=2021032312:VGRD:275 mb:24 hour fcst: - 6: 8:129314:d=2021032312:ABSV:275 mb:24 hour fcst: - 6: 9:151372:d=2021032312:CLMR:275 mb:24 hour fcst: - 6: 10:151551:d=2021032312:ICMR:275 mb:24 hour fcst: - 6: 11:166076:d=2021032312:RWMR:275 mb:24 hour fcst: - 6: 12:166255:d=2021032312:SNMR:275 mb:24 hour fcst: - 6: 13:176251:d=2021032312:GRLE:275 mb:24 hour fcst: - 6: 14:176610:d=2021032312:HGT:325 mb:24 hour fcst: - 6: + interp_atmos_master.sh[47]export err=0 - 6: + interp_atmos_master.sh[47]err=0 - 6: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 6: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 6: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_7_0p25 - 6: + product_functions.sh[5]local filename=pgb2bfile_f024_7_0p25 - 6: + product_functions.sh[6]wgrib2 pgb2bfile_f024_7_0p25 -not_if :RH: -grib pgb2bfile_f024_7_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_7_0p25.new - 6: 1:0:d=2021032312:TMP:275 mb:24 hour fcst: - 6: 2:420973:d=2021032312:RH:275 mb:24 hour fcst: - 6: 3:840504:d=2021032312:TCDC:275 mb:24 hour fcst: - 6: 4:1067923:d=2021032312:VVEL:275 mb:24 hour fcst: - 6: 5:1683992:d=2021032312:DZDT:275 mb:24 hour fcst: - 6: 6:2462759:d=2021032312:UGRD:275 mb:24 hour fcst: - 6: 7:2779598:d=2021032312:VGRD:275 mb:24 hour fcst: - 6: 8:3111578:d=2021032312:ABSV:275 mb:24 hour fcst: - 6: 9:3655298:d=2021032312:CLMR:275 mb:24 hour fcst: - 6: 10:3655477:d=2021032312:ICMR:275 mb:24 hour fcst: - 6: 11:3967439:d=2021032312:RWMR:275 mb:24 hour fcst: - 6: 12:3967618:d=2021032312:SNMR:275 mb:24 hour fcst: - 6: 13:4162733:d=2021032312:GRLE:275 mb:24 hour fcst: - 6: 14:4165900:d=2021032312:HGT:325 mb:24 hour fcst: - 6: + product_functions.sh[10]rc=0 - 6: + product_functions.sh[11](( rc == 0 )) - 6: + product_functions.sh[11]mv pgb2bfile_f024_7_0p25.new pgb2bfile_f024_7_0p25 - 6: + product_functions.sh[12]return 0 - 6: + interp_atmos_master.sh[56]export err=0 - 6: + interp_atmos_master.sh[56]err=0 - 6: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 6: ++ interp_atmos_master.sh[62]wc -l - 6: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_7_0p25 -match 'LAND|ICEC' - 6: + interp_atmos_master.sh[62]var_count=0 - 6: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 6: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 6: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_7_0p50 - 6: + product_functions.sh[5]local filename=pgb2bfile_f024_7_0p50 - 6: + product_functions.sh[6]wgrib2 pgb2bfile_f024_7_0p50 -not_if :RH: -grib pgb2bfile_f024_7_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_7_0p50.new - 6: 1:0:d=2021032312:TMP:275 mb:24 hour fcst: - 6: 2:144664:d=2021032312:RH:275 mb:24 hour fcst: - 6: 3:290011:d=2021032312:TCDC:275 mb:24 hour fcst: - 6: 4:376949:d=2021032312:VVEL:275 mb:24 hour fcst: - 6: 5:583720:d=2021032312:DZDT:275 mb:24 hour fcst: - 6: 6:836754:d=2021032312:UGRD:275 mb:24 hour fcst: - 6: 7:941297:d=2021032312:VGRD:275 mb:24 hour fcst: - 6: 8:1052034:d=2021032312:ABSV:275 mb:24 hour fcst: - 6: 9:1238065:d=2021032312:CLMR:275 mb:24 hour fcst: - 6: 10:1238244:d=2021032312:ICMR:275 mb:24 hour fcst: - 6: 11:1352866:d=2021032312:RWMR:275 mb:24 hour fcst: - 6: 12:1353045:d=2021032312:SNMR:275 mb:24 hour fcst: - 6: 13:1423747:d=2021032312:GRLE:275 mb:24 hour fcst: - 6: 14:1425142:d=2021032312:HGT:325 mb:24 hour fcst: - 6: + product_functions.sh[10]rc=0 - 6: + product_functions.sh[11](( rc == 0 )) - 6: + product_functions.sh[11]mv pgb2bfile_f024_7_0p50.new pgb2bfile_f024_7_0p50 - 6: + product_functions.sh[12]return 0 - 6: + interp_atmos_master.sh[56]export err=0 - 6: + interp_atmos_master.sh[56]err=0 - 6: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 6: ++ interp_atmos_master.sh[62]wc -l - 6: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_7_0p50 -match 'LAND|ICEC' - 6: + interp_atmos_master.sh[62]var_count=0 - 6: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 6: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 6: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_7_1p00 - 6: + product_functions.sh[5]local filename=pgb2bfile_f024_7_1p00 - 6: + product_functions.sh[6]wgrib2 pgb2bfile_f024_7_1p00 -not_if :RH: -grib pgb2bfile_f024_7_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_7_1p00.new - 6: 1:0:d=2021032312:TMP:275 mb:24 hour fcst: - 6: 2:47911:d=2021032312:RH:275 mb:24 hour fcst: - 6: 3:95777:d=2021032312:TCDC:275 mb:24 hour fcst: - 6: 4:126487:d=2021032312:VVEL:275 mb:24 hour fcst: - 6: 5:190314:d=2021032312:DZDT:275 mb:24 hour fcst: - 6: 6:266356:d=2021032312:UGRD:275 mb:24 hour fcst: - 6: 7:301851:d=2021032312:VGRD:275 mb:24 hour fcst: - 6: 8:339634:d=2021032312:ABSV:275 mb:24 hour fcst: - 6: 9:398641:d=2021032312:CLMR:275 mb:24 hour fcst: - 6: 10:398820:d=2021032312:ICMR:275 mb:24 hour fcst: - 6: 11:436744:d=2021032312:RWMR:275 mb:24 hour fcst: - 6: 12:436923:d=2021032312:SNMR:275 mb:24 hour fcst: - 6: 13:461508:d=2021032312:GRLE:275 mb:24 hour fcst: - 6: 14:462168:d=2021032312:HGT:325 mb:24 hour fcst: - 6: + product_functions.sh[10]rc=0 - 6: + product_functions.sh[11](( rc == 0 )) - 6: + product_functions.sh[11]mv pgb2bfile_f024_7_1p00.new pgb2bfile_f024_7_1p00 - 6: + product_functions.sh[12]return 0 - 6: + interp_atmos_master.sh[56]export err=0 - 6: + interp_atmos_master.sh[56]err=0 - 6: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 6: ++ interp_atmos_master.sh[62]wc -l - 6: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_7_1p00 -match 'LAND|ICEC' - 6: + interp_atmos_master.sh[62]var_count=0 - 6: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 6: + interp_atmos_master.sh[73]exit 0 -15: + bash[8]'[' -z '' ']' -15: + bash[9]case "$-" in -15: + bash[12]__lmod_vx=x -15: + bash[16]'[' -n x ']' -15: + bash[16]set +x -15: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -15: Shell debugging restarted -15: + bash[224]unset __lmod_vx -15: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_16 -15: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_16 -15: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -15: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -15: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -15: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -15: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -15: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -15: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -15: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -15: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -15: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -15: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -15: + interp_atmos_master.sh[31]IFS=: -15: + interp_atmos_master.sh[31]read -ra grids -15: + interp_atmos_master.sh[33]output_grids= -15: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -15: + interp_atmos_master.sh[35]gridopt=grid0p25 -15: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_16_0p25' -15: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -15: + interp_atmos_master.sh[35]gridopt=grid0p50 -15: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_16_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_16_0p50' -15: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -15: + interp_atmos_master.sh[35]gridopt=grid1p00 -15: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_16_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_16_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_16_1p00' -15: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_16 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_16_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_16_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_16_1p00 -15: 1:0:d=2021032312:TMP:725 mb:24 hour fcst: -15: 2:19720:d=2021032312:RH:725 mb:24 hour fcst: -15: 3:40062:d=2021032312:TCDC:725 mb:24 hour fcst: -15: 4:51756:d=2021032312:VVEL:725 mb:24 hour fcst: -15: 5:77444:d=2021032312:DZDT:725 mb:24 hour fcst: -15: 6:104709:d=2021032312:UGRD:725 mb:24 hour fcst: -15: 7:125532:d=2021032312:VGRD:725 mb:24 hour fcst: -15: 8:147162:d=2021032312:ABSV:725 mb:24 hour fcst: -15: 9:168047:d=2021032312:CLMR:725 mb:24 hour fcst: -15: 10:175983:d=2021032312:ICMR:725 mb:24 hour fcst: -15: 11:187867:d=2021032312:RWMR:725 mb:24 hour fcst: -15: 12:193599:d=2021032312:SNMR:725 mb:24 hour fcst: -15: 13:207414:d=2021032312:GRLE:725 mb:24 hour fcst: -15: 14:209348:d=2021032312:HGT:775 mb:24 hour fcst: -15: + interp_atmos_master.sh[47]export err=0 -15: + interp_atmos_master.sh[47]err=0 -15: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -15: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -15: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_16_0p25 -15: + product_functions.sh[5]local filename=pgb2bfile_f024_16_0p25 -15: + product_functions.sh[6]wgrib2 pgb2bfile_f024_16_0p25 -not_if :RH: -grib pgb2bfile_f024_16_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_16_0p25.new -15: 1:0:d=2021032312:TMP:725 mb:24 hour fcst: -15: 2:456210:d=2021032312:RH:725 mb:24 hour fcst: -15: 3:931943:d=2021032312:TCDC:725 mb:24 hour fcst: -15: 4:1163511:d=2021032312:VVEL:725 mb:24 hour fcst: -15: 5:1848884:d=2021032312:DZDT:725 mb:24 hour fcst: -15: 6:2602945:d=2021032312:UGRD:725 mb:24 hour fcst: -15: 7:3105186:d=2021032312:VGRD:725 mb:24 hour fcst: -15: 8:3638004:d=2021032312:ABSV:725 mb:24 hour fcst: -15: 9:4137665:d=2021032312:CLMR:725 mb:24 hour fcst: -15: 10:4311054:d=2021032312:ICMR:725 mb:24 hour fcst: -15: 11:4589813:d=2021032312:RWMR:725 mb:24 hour fcst: -15: 12:4698102:d=2021032312:SNMR:725 mb:24 hour fcst: -15: 13:4998117:d=2021032312:GRLE:725 mb:24 hour fcst: -15: 14:5030891:d=2021032312:HGT:775 mb:24 hour fcst: -15: + product_functions.sh[10]rc=0 -15: + product_functions.sh[11](( rc == 0 )) -15: + product_functions.sh[11]mv pgb2bfile_f024_16_0p25.new pgb2bfile_f024_16_0p25 -15: + product_functions.sh[12]return 0 -15: + interp_atmos_master.sh[56]export err=0 -15: + interp_atmos_master.sh[56]err=0 -15: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -15: ++ interp_atmos_master.sh[62]wc -l -15: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_16_0p25 -match 'LAND|ICEC' -15: + interp_atmos_master.sh[62]var_count=0 -15: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -15: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -15: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_16_0p50 -15: + product_functions.sh[5]local filename=pgb2bfile_f024_16_0p50 -15: + product_functions.sh[6]wgrib2 pgb2bfile_f024_16_0p50 -not_if :RH: -grib pgb2bfile_f024_16_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_16_0p50.new -15: 1:0:d=2021032312:TMP:725 mb:24 hour fcst: -15: 2:157479:d=2021032312:RH:725 mb:24 hour fcst: -15: 3:321400:d=2021032312:TCDC:725 mb:24 hour fcst: -15: 4:408666:d=2021032312:VVEL:725 mb:24 hour fcst: -15: 5:642460:d=2021032312:DZDT:725 mb:24 hour fcst: -15: 6:891478:d=2021032312:UGRD:725 mb:24 hour fcst: -15: 7:1062678:d=2021032312:VGRD:725 mb:24 hour fcst: -15: 8:1244906:d=2021032312:ABSV:725 mb:24 hour fcst: -15: 9:1417663:d=2021032312:CLMR:725 mb:24 hour fcst: -15: 10:1482433:d=2021032312:ICMR:725 mb:24 hour fcst: -15: 11:1581861:d=2021032312:RWMR:725 mb:24 hour fcst: -15: 12:1622355:d=2021032312:SNMR:725 mb:24 hour fcst: -15: 13:1730282:d=2021032312:GRLE:725 mb:24 hour fcst: -15: 14:1742947:d=2021032312:HGT:775 mb:24 hour fcst: -15: + product_functions.sh[10]rc=0 -15: + product_functions.sh[11](( rc == 0 )) -15: + product_functions.sh[11]mv pgb2bfile_f024_16_0p50.new pgb2bfile_f024_16_0p50 -15: + product_functions.sh[12]return 0 -15: + interp_atmos_master.sh[56]export err=0 -15: + interp_atmos_master.sh[56]err=0 -15: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -15: ++ interp_atmos_master.sh[62]wc -l -15: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_16_0p50 -match 'LAND|ICEC' -15: + interp_atmos_master.sh[62]var_count=0 -15: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -15: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -15: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_16_1p00 -15: + product_functions.sh[5]local filename=pgb2bfile_f024_16_1p00 -15: + product_functions.sh[6]wgrib2 pgb2bfile_f024_16_1p00 -not_if :RH: -grib pgb2bfile_f024_16_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_16_1p00.new -15: 1:0:d=2021032312:TMP:725 mb:24 hour fcst: -15: 2:51536:d=2021032312:RH:725 mb:24 hour fcst: -15: 3:105070:d=2021032312:TCDC:725 mb:24 hour fcst: -15: 4:136136:d=2021032312:VVEL:725 mb:24 hour fcst: -15: 5:207630:d=2021032312:DZDT:725 mb:24 hour fcst: -15: 6:282513:d=2021032312:UGRD:725 mb:24 hour fcst: -15: 7:337959:d=2021032312:VGRD:725 mb:24 hour fcst: -15: 8:396067:d=2021032312:ABSV:725 mb:24 hour fcst: -15: 9:451316:d=2021032312:CLMR:725 mb:24 hour fcst: -15: 10:473642:d=2021032312:ICMR:725 mb:24 hour fcst: -15: 11:506329:d=2021032312:RWMR:725 mb:24 hour fcst: -15: 12:520979:d=2021032312:SNMR:725 mb:24 hour fcst: -15: 13:557335:d=2021032312:GRLE:725 mb:24 hour fcst: -15: 14:562235:d=2021032312:HGT:775 mb:24 hour fcst: -15: + product_functions.sh[10]rc=0 -15: + product_functions.sh[11](( rc == 0 )) -15: + product_functions.sh[11]mv pgb2bfile_f024_16_1p00.new pgb2bfile_f024_16_1p00 -15: + product_functions.sh[12]return 0 -15: + interp_atmos_master.sh[56]export err=0 -15: + interp_atmos_master.sh[56]err=0 -15: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -15: ++ interp_atmos_master.sh[62]wc -l -15: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_16_1p00 -match 'LAND|ICEC' -15: + interp_atmos_master.sh[62]var_count=0 -15: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -15: + interp_atmos_master.sh[73]exit 0 - 9: + bash[8]'[' -z '' ']' - 9: + bash[9]case "$-" in - 9: + bash[12]__lmod_vx=x - 9: + bash[16]'[' -n x ']' - 9: + bash[16]set +x - 9: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 9: Shell debugging restarted - 9: + bash[224]unset __lmod_vx - 9: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_10 - 9: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_10 - 9: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 - 9: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 9: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 9: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 9: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 9: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 9: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 9: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 9: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 9: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 9: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 9: + interp_atmos_master.sh[31]IFS=: - 9: + interp_atmos_master.sh[31]read -ra grids - 9: + interp_atmos_master.sh[33]output_grids= - 9: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 9: + interp_atmos_master.sh[35]gridopt=grid0p25 - 9: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_10_0p25' - 9: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 9: + interp_atmos_master.sh[35]gridopt=grid0p50 - 9: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_10_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_10_0p50' - 9: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 9: + interp_atmos_master.sh[35]gridopt=grid1p00 - 9: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_10_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_10_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_10_1p00' - 9: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_10 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_10_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_10_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_10_1p00 - 9: 1:0:d=2021032312:TMP:425 mb:24 hour fcst: - 9: 2:18529:d=2021032312:RH:425 mb:24 hour fcst: - 9: 3:37898:d=2021032312:TCDC:425 mb:24 hour fcst: - 9: 4:50664:d=2021032312:VVEL:425 mb:24 hour fcst: - 9: 5:75743:d=2021032312:DZDT:425 mb:24 hour fcst: - 9: 6:103939:d=2021032312:UGRD:425 mb:24 hour fcst: - 9: 7:125772:d=2021032312:VGRD:425 mb:24 hour fcst: - 9: 8:148222:d=2021032312:ABSV:425 mb:24 hour fcst: - 9: 9:169833:d=2021032312:CLMR:425 mb:24 hour fcst: - 9: 10:170489:d=2021032312:ICMR:425 mb:24 hour fcst: - 9: 11:187291:d=2021032312:RWMR:425 mb:24 hour fcst: - 9: 12:187645:d=2021032312:SNMR:425 mb:24 hour fcst: - 9: 13:201997:d=2021032312:GRLE:425 mb:24 hour fcst: - 9: 14:202871:d=2021032312:HGT:475 mb:24 hour fcst: - 9: + interp_atmos_master.sh[47]export err=0 - 9: + interp_atmos_master.sh[47]err=0 - 9: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 9: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 9: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_10_0p25 - 9: + product_functions.sh[5]local filename=pgb2bfile_f024_10_0p25 - 9: + product_functions.sh[6]wgrib2 pgb2bfile_f024_10_0p25 -not_if :RH: -grib pgb2bfile_f024_10_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_10_0p25.new - 9: 1:0:d=2021032312:TMP:425 mb:24 hour fcst: - 9: 2:414598:d=2021032312:RH:425 mb:24 hour fcst: - 9: 3:855248:d=2021032312:TCDC:425 mb:24 hour fcst: - 9: 4:1090379:d=2021032312:VVEL:425 mb:24 hour fcst: - 9: 5:1755340:d=2021032312:DZDT:425 mb:24 hour fcst: - 9: 6:2544652:d=2021032312:UGRD:425 mb:24 hour fcst: - 9: 7:3084960:d=2021032312:VGRD:425 mb:24 hour fcst: - 9: 8:3653948:d=2021032312:ABSV:425 mb:24 hour fcst: - 9: 9:4183680:d=2021032312:CLMR:425 mb:24 hour fcst: - 9: 10:4193970:d=2021032312:ICMR:425 mb:24 hour fcst: - 9: 11:4567300:d=2021032312:RWMR:425 mb:24 hour fcst: - 9: 12:4569891:d=2021032312:SNMR:425 mb:24 hour fcst: - 9: 13:4878056:d=2021032312:GRLE:425 mb:24 hour fcst: - 9: 14:4887454:d=2021032312:HGT:475 mb:24 hour fcst: - 9: + product_functions.sh[10]rc=0 - 9: + product_functions.sh[11](( rc == 0 )) - 9: + product_functions.sh[11]mv pgb2bfile_f024_10_0p25.new pgb2bfile_f024_10_0p25 - 9: + product_functions.sh[12]return 0 - 9: + interp_atmos_master.sh[56]export err=0 - 9: + interp_atmos_master.sh[56]err=0 - 9: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 9: ++ interp_atmos_master.sh[62]wc -l - 9: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_10_0p25 -match 'LAND|ICEC' - 9: + interp_atmos_master.sh[62]var_count=0 - 9: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 9: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 9: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_10_0p50 - 9: + product_functions.sh[5]local filename=pgb2bfile_f024_10_0p50 - 9: + product_functions.sh[6]wgrib2 pgb2bfile_f024_10_0p50 -not_if :RH: -grib pgb2bfile_f024_10_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_10_0p50.new - 9: 1:0:d=2021032312:TMP:425 mb:24 hour fcst: - 9: 2:143788:d=2021032312:RH:425 mb:24 hour fcst: - 9: 3:297790:d=2021032312:TCDC:425 mb:24 hour fcst: - 9: 4:388438:d=2021032312:VVEL:425 mb:24 hour fcst: - 9: 5:615095:d=2021032312:DZDT:425 mb:24 hour fcst: - 9: 6:872678:d=2021032312:UGRD:425 mb:24 hour fcst: - 9: 7:1058038:d=2021032312:VGRD:425 mb:24 hour fcst: - 9: 8:1251615:d=2021032312:ABSV:425 mb:24 hour fcst: - 9: 9:1432546:d=2021032312:CLMR:425 mb:24 hour fcst: - 9: 10:1436470:d=2021032312:ICMR:425 mb:24 hour fcst: - 9: 11:1571859:d=2021032312:RWMR:425 mb:24 hour fcst: - 9: 12:1572972:d=2021032312:SNMR:425 mb:24 hour fcst: - 9: 13:1683972:d=2021032312:GRLE:425 mb:24 hour fcst: - 9: 14:1687778:d=2021032312:HGT:475 mb:24 hour fcst: - 9: + product_functions.sh[10]rc=0 - 9: + product_functions.sh[11](( rc == 0 )) - 9: + product_functions.sh[11]mv pgb2bfile_f024_10_0p50.new pgb2bfile_f024_10_0p50 - 9: + product_functions.sh[12]return 0 - 9: + interp_atmos_master.sh[56]export err=0 - 9: + interp_atmos_master.sh[56]err=0 - 9: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 9: ++ interp_atmos_master.sh[62]wc -l - 9: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_10_0p50 -match 'LAND|ICEC' - 9: + interp_atmos_master.sh[62]var_count=0 - 9: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 9: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 9: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_10_1p00 - 9: + product_functions.sh[5]local filename=pgb2bfile_f024_10_1p00 - 9: + product_functions.sh[6]wgrib2 pgb2bfile_f024_10_1p00 -not_if :RH: -grib pgb2bfile_f024_10_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_10_1p00.new - 9: 1:0:d=2021032312:TMP:425 mb:24 hour fcst: - 9: 2:47410:d=2021032312:RH:425 mb:24 hour fcst: - 9: 3:97857:d=2021032312:TCDC:425 mb:24 hour fcst: - 9: 4:130178:d=2021032312:VVEL:425 mb:24 hour fcst: - 9: 5:199550:d=2021032312:DZDT:425 mb:24 hour fcst: - 9: 6:277086:d=2021032312:UGRD:425 mb:24 hour fcst: - 9: 7:335777:d=2021032312:VGRD:425 mb:24 hour fcst: - 9: 8:397036:d=2021032312:ABSV:425 mb:24 hour fcst: - 9: 9:455119:d=2021032312:CLMR:425 mb:24 hour fcst: - 9: 10:456682:d=2021032312:ICMR:425 mb:24 hour fcst: - 9: 11:501444:d=2021032312:RWMR:425 mb:24 hour fcst: - 9: 12:501997:d=2021032312:SNMR:425 mb:24 hour fcst: - 9: 13:539268:d=2021032312:GRLE:425 mb:24 hour fcst: - 9: 14:541085:d=2021032312:HGT:475 mb:24 hour fcst: - 9: + product_functions.sh[10]rc=0 - 9: + product_functions.sh[11](( rc == 0 )) - 9: + product_functions.sh[11]mv pgb2bfile_f024_10_1p00.new pgb2bfile_f024_10_1p00 - 9: + product_functions.sh[12]return 0 - 9: + interp_atmos_master.sh[56]export err=0 - 9: + interp_atmos_master.sh[56]err=0 - 9: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 9: ++ interp_atmos_master.sh[62]wc -l - 9: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_10_1p00 -match 'LAND|ICEC' - 9: + interp_atmos_master.sh[62]var_count=0 - 9: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 9: + interp_atmos_master.sh[73]exit 0 - 2: + bash[8]'[' -z '' ']' - 2: + bash[9]case "$-" in - 2: + bash[12]__lmod_vx=x - 2: + bash[16]'[' -n x ']' - 2: + bash[16]set +x - 2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 2: Shell debugging restarted - 2: + bash[224]unset __lmod_vx - 2: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_3 - 2: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_3 - 2: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 - 2: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 2: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 2: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 2: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 2: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 2: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 2: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 2: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 2: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 2: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 2: + interp_atmos_master.sh[31]IFS=: - 2: + interp_atmos_master.sh[31]read -ra grids - 2: + interp_atmos_master.sh[33]output_grids= - 2: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 2: + interp_atmos_master.sh[35]gridopt=grid0p25 - 2: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_3_0p25' - 2: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 2: + interp_atmos_master.sh[35]gridopt=grid0p50 - 2: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_3_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_3_0p50' - 2: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 2: + interp_atmos_master.sh[35]gridopt=grid1p00 - 2: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_3_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_3_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_3_1p00' - 2: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_3 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_3_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_3_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_3_1p00 - 2: 1:0:d=2021032312:HGT:7 mb:24 hour fcst: - 2: 2:17759:d=2021032312:TMP:7 mb:24 hour fcst: - 2: 3:33948:d=2021032312:RH:7 mb:24 hour fcst: - 2: 4:40067:d=2021032312:UGRD:7 mb:24 hour fcst: - 2: 5:51481:d=2021032312:VGRD:7 mb:24 hour fcst: - 2: 6:70173:d=2021032312:ABSV:7 mb:24 hour fcst: - 2: 7:87982:d=2021032312:O3MR:7 mb:24 hour fcst: - 2: 8:111756:d=2021032312:TCDC:70 mb:24 hour fcst: - 2: 9:111935:d=2021032312:CLMR:70 mb:24 hour fcst: - 2: 10:112114:d=2021032312:ICMR:70 mb:24 hour fcst: - 2: 11:112293:d=2021032312:RWMR:70 mb:24 hour fcst: - 2: 12:112472:d=2021032312:SNMR:70 mb:24 hour fcst: - 2: 13:112651:d=2021032312:GRLE:70 mb:24 hour fcst: - 2: 14:112830:d=2021032312:HGT:125 mb:24 hour fcst: - 2: + interp_atmos_master.sh[47]export err=0 - 2: + interp_atmos_master.sh[47]err=0 - 2: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 2: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 2: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_3_0p25 - 2: + product_functions.sh[5]local filename=pgb2bfile_f024_3_0p25 - 2: + product_functions.sh[6]wgrib2 pgb2bfile_f024_3_0p25 -not_if :RH: -grib pgb2bfile_f024_3_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_3_0p25.new - 2: 1:0:d=2021032312:HGT:7 mb:24 hour fcst: - 2: 2:396451:d=2021032312:TMP:7 mb:24 hour fcst: - 2: 3:748607:d=2021032312:RH:7 mb:24 hour fcst: - 2: 4:923356:d=2021032312:UGRD:7 mb:24 hour fcst: - 2: 5:1370527:d=2021032312:VGRD:7 mb:24 hour fcst: - 2: 6:1802527:d=2021032312:ABSV:7 mb:24 hour fcst: - 2: 7:2193050:d=2021032312:O3MR:7 mb:24 hour fcst: - 2: 8:2798627:d=2021032312:TCDC:70 mb:24 hour fcst: - 2: 9:2798806:d=2021032312:CLMR:70 mb:24 hour fcst: - 2: 10:2798985:d=2021032312:ICMR:70 mb:24 hour fcst: - 2: 11:2799164:d=2021032312:RWMR:70 mb:24 hour fcst: - 2: 12:2799343:d=2021032312:SNMR:70 mb:24 hour fcst: - 2: 13:2799522:d=2021032312:GRLE:70 mb:24 hour fcst: - 2: 14:2799701:d=2021032312:HGT:125 mb:24 hour fcst: - 2: + product_functions.sh[10]rc=0 - 2: + product_functions.sh[11](( rc == 0 )) - 2: + product_functions.sh[11]mv pgb2bfile_f024_3_0p25.new pgb2bfile_f024_3_0p25 - 2: + product_functions.sh[12]return 0 - 2: + interp_atmos_master.sh[56]export err=0 - 2: + interp_atmos_master.sh[56]err=0 - 2: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 2: ++ interp_atmos_master.sh[62]wc -l - 2: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_3_0p25 -match 'LAND|ICEC' - 2: + interp_atmos_master.sh[62]var_count=0 - 2: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 2: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 2: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_3_0p50 - 2: + product_functions.sh[5]local filename=pgb2bfile_f024_3_0p50 - 2: + product_functions.sh[6]wgrib2 pgb2bfile_f024_3_0p50 -not_if :RH: -grib pgb2bfile_f024_3_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_3_0p50.new - 2: 1:0:d=2021032312:HGT:7 mb:24 hour fcst: - 2: 2:135687:d=2021032312:TMP:7 mb:24 hour fcst: - 2: 3:257412:d=2021032312:RH:7 mb:24 hour fcst: - 2: 4:313761:d=2021032312:UGRD:7 mb:24 hour fcst: - 2: 5:468958:d=2021032312:VGRD:7 mb:24 hour fcst: - 2: 6:619133:d=2021032312:ABSV:7 mb:24 hour fcst: - 2: 7:755023:d=2021032312:O3MR:7 mb:24 hour fcst: - 2: 8:961593:d=2021032312:TCDC:70 mb:24 hour fcst: - 2: 9:961772:d=2021032312:CLMR:70 mb:24 hour fcst: - 2: 10:961951:d=2021032312:ICMR:70 mb:24 hour fcst: - 2: 11:962130:d=2021032312:RWMR:70 mb:24 hour fcst: - 2: 12:962309:d=2021032312:SNMR:70 mb:24 hour fcst: - 2: 13:962488:d=2021032312:GRLE:70 mb:24 hour fcst: - 2: 14:962667:d=2021032312:HGT:125 mb:24 hour fcst: - 2: + product_functions.sh[10]rc=0 - 2: + product_functions.sh[11](( rc == 0 )) - 2: + product_functions.sh[11]mv pgb2bfile_f024_3_0p50.new pgb2bfile_f024_3_0p50 - 2: + product_functions.sh[12]return 0 - 2: + interp_atmos_master.sh[56]export err=0 - 2: + interp_atmos_master.sh[56]err=0 - 2: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 2: ++ interp_atmos_master.sh[62]wc -l - 2: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_3_0p50 -match 'LAND|ICEC' - 2: + interp_atmos_master.sh[62]var_count=0 - 2: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 2: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 2: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_3_1p00 - 2: + product_functions.sh[5]local filename=pgb2bfile_f024_3_1p00 - 2: + product_functions.sh[6]wgrib2 pgb2bfile_f024_3_1p00 -not_if :RH: -grib pgb2bfile_f024_3_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_3_1p00.new - 2: 1:0:d=2021032312:HGT:7 mb:24 hour fcst: - 2: 2:44441:d=2021032312:TMP:7 mb:24 hour fcst: - 2: 3:85135:d=2021032312:RH:7 mb:24 hour fcst: - 2: 4:101754:d=2021032312:UGRD:7 mb:24 hour fcst: - 2: 5:152468:d=2021032312:VGRD:7 mb:24 hour fcst: - 2: 6:201482:d=2021032312:ABSV:7 mb:24 hour fcst: - 2: 7:246669:d=2021032312:O3MR:7 mb:24 hour fcst: - 2: 8:311068:d=2021032312:TCDC:70 mb:24 hour fcst: - 2: 9:311247:d=2021032312:CLMR:70 mb:24 hour fcst: - 2: 10:311426:d=2021032312:ICMR:70 mb:24 hour fcst: - 2: 11:311605:d=2021032312:RWMR:70 mb:24 hour fcst: - 2: 12:311784:d=2021032312:SNMR:70 mb:24 hour fcst: - 2: 13:311963:d=2021032312:GRLE:70 mb:24 hour fcst: - 2: 14:312142:d=2021032312:HGT:125 mb:24 hour fcst: - 2: + product_functions.sh[10]rc=0 - 2: + product_functions.sh[11](( rc == 0 )) - 2: + product_functions.sh[11]mv pgb2bfile_f024_3_1p00.new pgb2bfile_f024_3_1p00 - 2: + product_functions.sh[12]return 0 - 2: + interp_atmos_master.sh[56]export err=0 - 2: + interp_atmos_master.sh[56]err=0 - 2: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 2: ++ interp_atmos_master.sh[62]wc -l - 2: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_3_1p00 -match 'LAND|ICEC' - 2: + interp_atmos_master.sh[62]var_count=0 - 2: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 2: + interp_atmos_master.sh[73]exit 0 - 3: + bash[8]'[' -z '' ']' - 3: + bash[9]case "$-" in - 3: + bash[12]__lmod_vx=x - 3: + bash[16]'[' -n x ']' - 3: + bash[16]set +x - 3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 3: Shell debugging restarted - 3: + bash[224]unset __lmod_vx - 3: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_4 - 3: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_4 - 3: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 - 3: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 3: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 3: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 3: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 3: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 3: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 3: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 3: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 3: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 3: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 3: + interp_atmos_master.sh[31]IFS=: - 3: + interp_atmos_master.sh[31]read -ra grids - 3: + interp_atmos_master.sh[33]output_grids= - 3: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 3: + interp_atmos_master.sh[35]gridopt=grid0p25 - 3: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_4_0p25' - 3: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 3: + interp_atmos_master.sh[35]gridopt=grid0p50 - 3: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_4_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_4_0p50' - 3: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 3: + interp_atmos_master.sh[35]gridopt=grid1p00 - 3: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_4_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_4_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_4_1p00' - 3: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_4 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_4_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_4_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_4_1p00 - 3: 1:0:d=2021032312:TMP:125 mb:24 hour fcst: - 3: 2:17041:d=2021032312:RH:125 mb:24 hour fcst: - 3: 3:27686:d=2021032312:TCDC:125 mb:24 hour fcst: - 3: 4:30050:d=2021032312:VVEL:125 mb:24 hour fcst: - 3: 5:56494:d=2021032312:DZDT:125 mb:24 hour fcst: - 3: 6:81648:d=2021032312:UGRD:125 mb:24 hour fcst: - 3: 7:101183:d=2021032312:VGRD:125 mb:24 hour fcst: - 3: 8:120835:d=2021032312:ABSV:125 mb:24 hour fcst: - 3: 9:139162:d=2021032312:CLMR:125 mb:24 hour fcst: - 3: 10:139341:d=2021032312:ICMR:125 mb:24 hour fcst: - 3: 11:143735:d=2021032312:RWMR:125 mb:24 hour fcst: - 3: 12:143914:d=2021032312:SNMR:125 mb:24 hour fcst: - 3: 13:147395:d=2021032312:GRLE:125 mb:24 hour fcst: - 3: 14:147880:d=2021032312:HGT:175 mb:24 hour fcst: - 3: + interp_atmos_master.sh[47]export err=0 - 3: + interp_atmos_master.sh[47]err=0 - 3: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 3: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 3: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_4_0p25 - 3: + product_functions.sh[5]local filename=pgb2bfile_f024_4_0p25 - 3: + product_functions.sh[6]wgrib2 pgb2bfile_f024_4_0p25 -not_if :RH: -grib pgb2bfile_f024_4_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_4_0p25.new - 3: 1:0:d=2021032312:TMP:125 mb:24 hour fcst: - 3: 2:376596:d=2021032312:RH:125 mb:24 hour fcst: - 3: 3:623567:d=2021032312:TCDC:125 mb:24 hour fcst: - 3: 4:671911:d=2021032312:VVEL:125 mb:24 hour fcst: - 3: 5:1392100:d=2021032312:DZDT:125 mb:24 hour fcst: - 3: 6:2077685:d=2021032312:UGRD:125 mb:24 hour fcst: - 3: 7:2554041:d=2021032312:VGRD:125 mb:24 hour fcst: - 3: 8:3034742:d=2021032312:ABSV:125 mb:24 hour fcst: - 3: 9:3455151:d=2021032312:CLMR:125 mb:24 hour fcst: - 3: 10:3455330:d=2021032312:ICMR:125 mb:24 hour fcst: - 3: 11:3557304:d=2021032312:RWMR:125 mb:24 hour fcst: - 3: 12:3557483:d=2021032312:SNMR:125 mb:24 hour fcst: - 3: 13:3614406:d=2021032312:GRLE:125 mb:24 hour fcst: - 3: 14:3620662:d=2021032312:HGT:175 mb:24 hour fcst: - 3: + product_functions.sh[10]rc=0 - 3: + product_functions.sh[11](( rc == 0 )) - 3: + product_functions.sh[11]mv pgb2bfile_f024_4_0p25.new pgb2bfile_f024_4_0p25 - 3: + product_functions.sh[12]return 0 - 3: + interp_atmos_master.sh[56]export err=0 - 3: + interp_atmos_master.sh[56]err=0 - 3: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 3: ++ interp_atmos_master.sh[62]wc -l - 3: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_4_0p25 -match 'LAND|ICEC' - 3: + interp_atmos_master.sh[62]var_count=0 - 3: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 3: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 3: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_4_0p50 - 3: + product_functions.sh[5]local filename=pgb2bfile_f024_4_0p50 - 3: + product_functions.sh[6]wgrib2 pgb2bfile_f024_4_0p50 -not_if :RH: -grib pgb2bfile_f024_4_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_4_0p50.new - 3: 1:0:d=2021032312:TMP:125 mb:24 hour fcst: - 3: 2:129353:d=2021032312:RH:125 mb:24 hour fcst: - 3: 3:210322:d=2021032312:TCDC:125 mb:24 hour fcst: - 3: 4:227624:d=2021032312:VVEL:125 mb:24 hour fcst: - 3: 5:472927:d=2021032312:DZDT:125 mb:24 hour fcst: - 3: 6:696446:d=2021032312:UGRD:125 mb:24 hour fcst: - 3: 7:857987:d=2021032312:VGRD:125 mb:24 hour fcst: - 3: 8:1020700:d=2021032312:ABSV:125 mb:24 hour fcst: - 3: 9:1165218:d=2021032312:CLMR:125 mb:24 hour fcst: - 3: 10:1165397:d=2021032312:ICMR:125 mb:24 hour fcst: - 3: 11:1201314:d=2021032312:RWMR:125 mb:24 hour fcst: - 3: 12:1201493:d=2021032312:SNMR:125 mb:24 hour fcst: - 3: 13:1222446:d=2021032312:GRLE:125 mb:24 hour fcst: - 3: 14:1224981:d=2021032312:HGT:175 mb:24 hour fcst: - 3: + product_functions.sh[10]rc=0 - 3: + product_functions.sh[11](( rc == 0 )) - 3: + product_functions.sh[11]mv pgb2bfile_f024_4_0p50.new pgb2bfile_f024_4_0p50 - 3: + product_functions.sh[12]return 0 - 3: + interp_atmos_master.sh[56]export err=0 - 3: + interp_atmos_master.sh[56]err=0 - 3: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 3: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_4_0p50 -match 'LAND|ICEC' - 3: ++ interp_atmos_master.sh[62]wc -l - 3: + interp_atmos_master.sh[62]var_count=0 - 3: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 3: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 3: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_4_1p00 - 3: + product_functions.sh[5]local filename=pgb2bfile_f024_4_1p00 - 3: + product_functions.sh[6]wgrib2 pgb2bfile_f024_4_1p00 -not_if :RH: -grib pgb2bfile_f024_4_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_4_1p00.new - 3: 1:0:d=2021032312:TMP:125 mb:24 hour fcst: - 3: 2:42636:d=2021032312:RH:125 mb:24 hour fcst: - 3: 3:68991:d=2021032312:TCDC:125 mb:24 hour fcst: - 3: 4:75116:d=2021032312:VVEL:125 mb:24 hour fcst: - 3: 5:148859:d=2021032312:DZDT:125 mb:24 hour fcst: - 3: 6:217072:d=2021032312:UGRD:125 mb:24 hour fcst: - 3: 7:269020:d=2021032312:VGRD:125 mb:24 hour fcst: - 3: 8:321191:d=2021032312:ABSV:125 mb:24 hour fcst: - 3: 9:368447:d=2021032312:CLMR:125 mb:24 hour fcst: - 3: 10:368626:d=2021032312:ICMR:125 mb:24 hour fcst: - 3: 11:380280:d=2021032312:RWMR:125 mb:24 hour fcst: - 3: 12:380459:d=2021032312:SNMR:125 mb:24 hour fcst: - 3: 13:388274:d=2021032312:GRLE:125 mb:24 hour fcst: - 3: 14:389279:d=2021032312:HGT:175 mb:24 hour fcst: - 3: + product_functions.sh[10]rc=0 - 3: + product_functions.sh[11](( rc == 0 )) - 3: + product_functions.sh[11]mv pgb2bfile_f024_4_1p00.new pgb2bfile_f024_4_1p00 - 3: + product_functions.sh[12]return 0 - 3: + interp_atmos_master.sh[56]export err=0 - 3: + interp_atmos_master.sh[56]err=0 - 3: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 3: ++ interp_atmos_master.sh[62]wc -l - 3: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_4_1p00 -match 'LAND|ICEC' - 3: + interp_atmos_master.sh[62]var_count=0 - 3: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 3: + interp_atmos_master.sh[73]exit 0 -23: + bash[8]'[' -z '' ']' -23: + bash[9]case "$-" in -23: + bash[12]__lmod_vx=x -23: + bash[16]'[' -n x ']' -23: + bash[16]set +x -23: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -23: Shell debugging restarted -23: + bash[224]unset __lmod_vx -23: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_24 -23: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_24 -23: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -23: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -23: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -23: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -23: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -23: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -23: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -23: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -23: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -23: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -23: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -23: + interp_atmos_master.sh[31]IFS=: -23: + interp_atmos_master.sh[31]read -ra grids -23: + interp_atmos_master.sh[33]output_grids= -23: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -23: + interp_atmos_master.sh[35]gridopt=grid0p25 -23: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_24_0p25' -23: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -23: + interp_atmos_master.sh[35]gridopt=grid0p50 -23: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_24_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_24_0p50' -23: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -23: + interp_atmos_master.sh[35]gridopt=grid1p00 -23: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_24_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_24_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_24_1p00' -23: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_24 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_24_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_24_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_24_1p00 -23: 1:0:d=2021032312:HGT:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -23: 2:17704:d=2021032312:PRES:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -23: 3:35394:d=2021032312:VWSH:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -23: 4:46760:d=2021032312:UGRD:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 5:58104:d=2021032312:VGRD:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 6:69418:d=2021032312:TMP:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 7:85317:d=2021032312:HGT:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 8:104053:d=2021032312:PRES:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 9:122811:d=2021032312:VWSH:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 10:135100:d=2021032312:UGRD:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 11:145257:d=2021032312:VGRD:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 12:156548:d=2021032312:TMP:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 13:172342:d=2021032312:HGT:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 14:190868:d=2021032312:PRES:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 15:209321:d=2021032312:VWSH:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 16:221629:d=2021032312:UGRD:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 17:232733:d=2021032312:VGRD:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 18:243696:d=2021032312:TMP:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 19:258663:d=2021032312:HGT:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 20:276519:d=2021032312:PRES:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 21:294333:d=2021032312:VWSH:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 22:306473:d=2021032312:UGRD:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 23:317538:d=2021032312:VGRD:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 24:328566:d=2021032312:TMP:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 25:343623:d=2021032312:HGT:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 26:361457:d=2021032312:PRES:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 27:379133:d=2021032312:VWSH:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: + interp_atmos_master.sh[47]export err=0 -23: + interp_atmos_master.sh[47]err=0 -23: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -23: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -23: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_24_0p25 -23: + product_functions.sh[5]local filename=pgb2bfile_f024_24_0p25 -23: + product_functions.sh[6]wgrib2 pgb2bfile_f024_24_0p25 -not_if :RH: -grib pgb2bfile_f024_24_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_24_0p25.new -23: 1:0:d=2021032312:HGT:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -23: 2:583593:d=2021032312:PRES:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -23: 3:1165612:d=2021032312:VWSH:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -23: 4:1509231:d=2021032312:UGRD:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 5:1838221:d=2021032312:VGRD:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 6:2164205:d=2021032312:TMP:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 7:2640590:d=2021032312:HGT:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 8:3237612:d=2021032312:PRES:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 9:3830607:d=2021032312:VWSH:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 10:4192473:d=2021032312:UGRD:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 11:4520402:d=2021032312:VGRD:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 12:4843183:d=2021032312:TMP:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 13:5323179:d=2021032312:HGT:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 14:5911377:d=2021032312:PRES:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 15:6499702:d=2021032312:VWSH:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 16:6858672:d=2021032312:UGRD:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 17:7180414:d=2021032312:VGRD:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 18:7500079:d=2021032312:TMP:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 19:7949794:d=2021032312:HGT:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 20:8517117:d=2021032312:PRES:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 21:9060648:d=2021032312:VWSH:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 22:9417806:d=2021032312:UGRD:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 23:9741767:d=2021032312:VGRD:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 24:10059778:d=2021032312:TMP:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 25:10513120:d=2021032312:HGT:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 26:11082956:d=2021032312:PRES:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 27:11626103:d=2021032312:VWSH:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: + product_functions.sh[10]rc=0 -23: + product_functions.sh[11](( rc == 0 )) -23: + product_functions.sh[11]mv pgb2bfile_f024_24_0p25.new pgb2bfile_f024_24_0p25 -23: + product_functions.sh[12]return 0 -23: + interp_atmos_master.sh[56]export err=0 -23: + interp_atmos_master.sh[56]err=0 -23: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -23: ++ interp_atmos_master.sh[62]wc -l -23: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_24_0p25 -match 'LAND|ICEC' -23: + interp_atmos_master.sh[62]var_count=0 -23: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -23: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -23: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_24_0p50 -23: + product_functions.sh[5]local filename=pgb2bfile_f024_24_0p50 -23: + product_functions.sh[6]wgrib2 pgb2bfile_f024_24_0p50 -not_if :RH: -grib pgb2bfile_f024_24_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_24_0p50.new -23: 1:0:d=2021032312:HGT:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -23: 2:183258:d=2021032312:PRES:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -23: 3:366726:d=2021032312:VWSH:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -23: 4:472835:d=2021032312:UGRD:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 5:573954:d=2021032312:VGRD:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 6:673487:d=2021032312:TMP:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 7:824933:d=2021032312:HGT:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 8:1013008:d=2021032312:PRES:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 9:1201229:d=2021032312:VWSH:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 10:1312223:d=2021032312:UGRD:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 11:1412586:d=2021032312:VGRD:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 12:1511231:d=2021032312:TMP:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 13:1663141:d=2021032312:HGT:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 14:1849682:d=2021032312:PRES:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 15:2034417:d=2021032312:VWSH:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 16:2145826:d=2021032312:UGRD:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 17:2243191:d=2021032312:VGRD:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 18:2339959:d=2021032312:TMP:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 19:2481449:d=2021032312:HGT:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 20:2659771:d=2021032312:PRES:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 21:2834750:d=2021032312:VWSH:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 22:2944371:d=2021032312:UGRD:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 23:3042385:d=2021032312:VGRD:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 24:3139338:d=2021032312:TMP:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 25:3281912:d=2021032312:HGT:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 26:3461113:d=2021032312:PRES:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 27:3634428:d=2021032312:VWSH:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: + product_functions.sh[10]rc=0 -23: + product_functions.sh[11](( rc == 0 )) -23: + product_functions.sh[11]mv pgb2bfile_f024_24_0p50.new pgb2bfile_f024_24_0p50 -23: + product_functions.sh[12]return 0 -23: + interp_atmos_master.sh[56]export err=0 -23: + interp_atmos_master.sh[56]err=0 -23: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -23: ++ interp_atmos_master.sh[62]wc -l -23: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_24_0p50 -match 'LAND|ICEC' -23: + interp_atmos_master.sh[62]var_count=0 -23: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -23: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -23: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_24_1p00 -23: + product_functions.sh[5]local filename=pgb2bfile_f024_24_1p00 -23: + product_functions.sh[6]wgrib2 pgb2bfile_f024_24_1p00 -not_if :RH: -grib pgb2bfile_f024_24_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_24_1p00.new -23: 1:0:d=2021032312:HGT:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -23: 2:53150:d=2021032312:PRES:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -23: 3:106112:d=2021032312:VWSH:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -23: 4:137900:d=2021032312:UGRD:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 5:168960:d=2021032312:VGRD:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 6:199700:d=2021032312:TMP:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 7:245115:d=2021032312:HGT:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 8:300016:d=2021032312:PRES:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 9:354975:d=2021032312:VWSH:PV=1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 10:388643:d=2021032312:UGRD:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 11:419524:d=2021032312:VGRD:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 12:450161:d=2021032312:TMP:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 13:495594:d=2021032312:HGT:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 14:549848:d=2021032312:PRES:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 15:603690:d=2021032312:VWSH:PV=-1e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 16:637333:d=2021032312:UGRD:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 17:667327:d=2021032312:VGRD:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 18:697091:d=2021032312:TMP:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 19:739753:d=2021032312:HGT:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 20:792025:d=2021032312:PRES:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 21:843570:d=2021032312:VWSH:PV=1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 22:876761:d=2021032312:UGRD:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 23:906898:d=2021032312:VGRD:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 24:936702:d=2021032312:TMP:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 25:979342:d=2021032312:HGT:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 26:1031544:d=2021032312:PRES:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: 27:1082645:d=2021032312:VWSH:PV=-1.5e-06 (Km^2/kg/s) surface:24 hour fcst: -23: + product_functions.sh[10]rc=0 -23: + product_functions.sh[11](( rc == 0 )) -23: + product_functions.sh[11]mv pgb2bfile_f024_24_1p00.new pgb2bfile_f024_24_1p00 -23: + product_functions.sh[12]return 0 -23: + interp_atmos_master.sh[56]export err=0 -23: + interp_atmos_master.sh[56]err=0 -23: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -23: ++ interp_atmos_master.sh[62]wc -l -23: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_24_1p00 -match 'LAND|ICEC' -23: + interp_atmos_master.sh[62]var_count=0 -23: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -23: + interp_atmos_master.sh[73]exit 0 - 4: + bash[8]'[' -z '' ']' - 4: + bash[9]case "$-" in - 4: + bash[12]__lmod_vx=x - 4: + bash[16]'[' -n x ']' - 4: + bash[16]set +x - 4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 4: Shell debugging restarted - 4: + bash[224]unset __lmod_vx - 4: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_5 - 4: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_5 - 4: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 - 4: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 4: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 4: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 4: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 4: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 4: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 4: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 4: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 4: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 4: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 4: + interp_atmos_master.sh[31]IFS=: - 4: + interp_atmos_master.sh[31]read -ra grids - 4: + interp_atmos_master.sh[33]output_grids= - 4: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 4: + interp_atmos_master.sh[35]gridopt=grid0p25 - 4: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_5_0p25' - 4: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 4: + interp_atmos_master.sh[35]gridopt=grid0p50 - 4: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_5_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_5_0p50' - 4: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 4: + interp_atmos_master.sh[35]gridopt=grid1p00 - 4: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_5_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_5_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_5_1p00' - 4: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_5 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_5_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_5_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_5_1p00 - 4: 1:0:d=2021032312:TMP:175 mb:24 hour fcst: - 4: 2:17528:d=2021032312:RH:175 mb:24 hour fcst: - 4: 3:30683:d=2021032312:TCDC:175 mb:24 hour fcst: - 4: 4:35300:d=2021032312:VVEL:175 mb:24 hour fcst: - 4: 5:63785:d=2021032312:DZDT:175 mb:24 hour fcst: - 4: 6:89919:d=2021032312:UGRD:175 mb:24 hour fcst: - 4: 7:102880:d=2021032312:VGRD:175 mb:24 hour fcst: - 4: 8:123650:d=2021032312:ABSV:175 mb:24 hour fcst: - 4: 9:143049:d=2021032312:CLMR:175 mb:24 hour fcst: - 4: 10:143228:d=2021032312:ICMR:175 mb:24 hour fcst: - 4: 11:149746:d=2021032312:RWMR:175 mb:24 hour fcst: - 4: 12:149925:d=2021032312:SNMR:175 mb:24 hour fcst: - 4: 13:153634:d=2021032312:GRLE:175 mb:24 hour fcst: - 4: 14:154131:d=2021032312:HGT:225 mb:24 hour fcst: - 4: + interp_atmos_master.sh[47]export err=0 - 4: + interp_atmos_master.sh[47]err=0 - 4: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 4: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 4: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_5_0p25 - 4: + product_functions.sh[5]local filename=pgb2bfile_f024_5_0p25 - 4: + product_functions.sh[6]wgrib2 pgb2bfile_f024_5_0p25 -not_if :RH: -grib pgb2bfile_f024_5_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_5_0p25.new - 4: 1:0:d=2021032312:TMP:175 mb:24 hour fcst: - 4: 2:392602:d=2021032312:RH:175 mb:24 hour fcst: - 4: 3:694079:d=2021032312:TCDC:175 mb:24 hour fcst: - 4: 4:786505:d=2021032312:VVEL:175 mb:24 hour fcst: - 4: 5:1565720:d=2021032312:DZDT:175 mb:24 hour fcst: - 4: 6:2280552:d=2021032312:UGRD:175 mb:24 hour fcst: - 4: 7:2796556:d=2021032312:VGRD:175 mb:24 hour fcst: - 4: 8:3315165:d=2021032312:ABSV:175 mb:24 hour fcst: - 4: 9:3763270:d=2021032312:CLMR:175 mb:24 hour fcst: - 4: 10:3763449:d=2021032312:ICMR:175 mb:24 hour fcst: - 4: 11:3909327:d=2021032312:RWMR:175 mb:24 hour fcst: - 4: 12:3909506:d=2021032312:SNMR:175 mb:24 hour fcst: - 4: 13:3973319:d=2021032312:GRLE:175 mb:24 hour fcst: - 4: 14:3980065:d=2021032312:HGT:225 mb:24 hour fcst: - 4: + product_functions.sh[10]rc=0 - 4: + product_functions.sh[11](( rc == 0 )) - 4: + product_functions.sh[11]mv pgb2bfile_f024_5_0p25.new pgb2bfile_f024_5_0p25 - 4: + product_functions.sh[12]return 0 - 4: + interp_atmos_master.sh[56]export err=0 - 4: + interp_atmos_master.sh[56]err=0 - 4: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 4: ++ interp_atmos_master.sh[62]wc -l - 4: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_5_0p25 -match 'LAND|ICEC' - 4: + interp_atmos_master.sh[62]var_count=0 - 4: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 4: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 4: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_5_0p50 - 4: + product_functions.sh[5]local filename=pgb2bfile_f024_5_0p50 - 4: + product_functions.sh[6]wgrib2 pgb2bfile_f024_5_0p50 -not_if :RH: -grib pgb2bfile_f024_5_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_5_0p50.new - 4: 1:0:d=2021032312:TMP:175 mb:24 hour fcst: - 4: 2:135085:d=2021032312:RH:175 mb:24 hour fcst: - 4: 3:237118:d=2021032312:TCDC:175 mb:24 hour fcst: - 4: 4:271606:d=2021032312:VVEL:175 mb:24 hour fcst: - 4: 5:540034:d=2021032312:DZDT:175 mb:24 hour fcst: - 4: 6:774826:d=2021032312:UGRD:175 mb:24 hour fcst: - 4: 7:952075:d=2021032312:VGRD:175 mb:24 hour fcst: - 4: 8:1126321:d=2021032312:ABSV:175 mb:24 hour fcst: - 4: 9:1282340:d=2021032312:CLMR:175 mb:24 hour fcst: - 4: 10:1282519:d=2021032312:ICMR:175 mb:24 hour fcst: - 4: 11:1333938:d=2021032312:RWMR:175 mb:24 hour fcst: - 4: 12:1334117:d=2021032312:SNMR:175 mb:24 hour fcst: - 4: 13:1357417:d=2021032312:GRLE:175 mb:24 hour fcst: - 4: 14:1360151:d=2021032312:HGT:225 mb:24 hour fcst: - 4: + product_functions.sh[10]rc=0 - 4: + product_functions.sh[11](( rc == 0 )) - 4: + product_functions.sh[11]mv pgb2bfile_f024_5_0p50.new pgb2bfile_f024_5_0p50 - 4: + product_functions.sh[12]return 0 - 4: + interp_atmos_master.sh[56]export err=0 - 4: + interp_atmos_master.sh[56]err=0 - 4: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 4: ++ interp_atmos_master.sh[62]wc -l - 4: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_5_0p50 -match 'LAND|ICEC' - 4: + interp_atmos_master.sh[62]var_count=0 - 4: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 4: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 4: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_5_1p00 - 4: + product_functions.sh[5]local filename=pgb2bfile_f024_5_1p00 - 4: + product_functions.sh[6]wgrib2 pgb2bfile_f024_5_1p00 -not_if :RH: -grib pgb2bfile_f024_5_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_5_1p00.new - 4: 1:0:d=2021032312:TMP:175 mb:24 hour fcst: - 4: 2:44667:d=2021032312:RH:175 mb:24 hour fcst: - 4: 3:77911:d=2021032312:TCDC:175 mb:24 hour fcst: - 4: 4:89869:d=2021032312:VVEL:175 mb:24 hour fcst: - 4: 5:169686:d=2021032312:DZDT:175 mb:24 hour fcst: - 4: 6:240277:d=2021032312:UGRD:175 mb:24 hour fcst: - 4: 7:296681:d=2021032312:VGRD:175 mb:24 hour fcst: - 4: 8:352111:d=2021032312:ABSV:175 mb:24 hour fcst: - 4: 9:402410:d=2021032312:CLMR:175 mb:24 hour fcst: - 4: 10:402589:d=2021032312:ICMR:175 mb:24 hour fcst: - 4: 11:419625:d=2021032312:RWMR:175 mb:24 hour fcst: - 4: 12:419804:d=2021032312:SNMR:175 mb:24 hour fcst: - 4: 13:428401:d=2021032312:GRLE:175 mb:24 hour fcst: - 4: 14:429482:d=2021032312:HGT:225 mb:24 hour fcst: - 4: + product_functions.sh[10]rc=0 - 4: + product_functions.sh[11](( rc == 0 )) - 4: + product_functions.sh[11]mv pgb2bfile_f024_5_1p00.new pgb2bfile_f024_5_1p00 - 4: + product_functions.sh[12]return 0 - 4: + interp_atmos_master.sh[56]export err=0 - 4: + interp_atmos_master.sh[56]err=0 - 4: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 4: ++ interp_atmos_master.sh[62]wc -l - 4: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_5_1p00 -match 'LAND|ICEC' - 4: + interp_atmos_master.sh[62]var_count=0 - 4: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 4: + interp_atmos_master.sh[73]exit 0 -22: + bash[8]'[' -z '' ']' -22: + bash[9]case "$-" in -22: + bash[12]__lmod_vx=x -22: + bash[16]'[' -n x ']' -22: + bash[16]set +x -22: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -22: Shell debugging restarted -22: + bash[224]unset __lmod_vx -22: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_23 -22: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_23 -22: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -22: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -22: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -22: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -22: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -22: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -22: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -22: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -22: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -22: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -22: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -22: + interp_atmos_master.sh[31]IFS=: -22: + interp_atmos_master.sh[31]read -ra grids -22: + interp_atmos_master.sh[33]output_grids= -22: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -22: + interp_atmos_master.sh[35]gridopt=grid0p25 -22: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_23_0p25' -22: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -22: + interp_atmos_master.sh[35]gridopt=grid0p50 -22: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_23_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_23_0p50' -22: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -22: + interp_atmos_master.sh[35]gridopt=grid1p00 -22: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_23_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_23_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_23_1p00' -22: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_23 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_23_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_23_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_23_1p00 -22: 1:0:d=2021032312:TMP:180-150 mb above ground:24 hour fcst: -22: 2:21112:d=2021032312:RH:180-150 mb above ground:24 hour fcst: -22: 3:41126:d=2021032312:SPFH:180-150 mb above ground:24 hour fcst: -22: 4:70360:d=2021032312:UGRD:180-150 mb above ground:24 hour fcst: -22: 5:91426:d=2021032312:VGRD:180-150 mb above ground:24 hour fcst: -22: 6:113098:d=2021032312:UGRD:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 7:123780:d=2021032312:VGRD:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 8:134322:d=2021032312:TMP:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 9:148898:d=2021032312:HGT:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 10:165614:d=2021032312:PRES:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 11:182456:d=2021032312:VWSH:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 12:193359:d=2021032312:UGRD:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 13:204423:d=2021032312:VGRD:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 14:218922:d=2021032312:TMP:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: + interp_atmos_master.sh[47]export err=0 -22: + interp_atmos_master.sh[47]err=0 -22: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -22: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -22: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_23_0p25 -22: + product_functions.sh[5]local filename=pgb2bfile_f024_23_0p25 -22: + product_functions.sh[6]wgrib2 pgb2bfile_f024_23_0p25 -not_if :RH: -grib pgb2bfile_f024_23_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_23_0p25.new -22: 1:0:d=2021032312:TMP:180-150 mb above ground:24 hour fcst: -22: 2:507148:d=2021032312:RH:180-150 mb above ground:24 hour fcst: -22: 3:964049:d=2021032312:SPFH:180-150 mb above ground:24 hour fcst: -22: 4:1759086:d=2021032312:UGRD:180-150 mb above ground:24 hour fcst: -22: 5:2261859:d=2021032312:VGRD:180-150 mb above ground:24 hour fcst: -22: 6:2793238:d=2021032312:UGRD:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 7:3107486:d=2021032312:VGRD:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 8:3414856:d=2021032312:TMP:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 9:3880549:d=2021032312:HGT:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 10:4426539:d=2021032312:PRES:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 11:4981293:d=2021032312:VWSH:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 12:5316132:d=2021032312:UGRD:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 13:5785439:d=2021032312:VGRD:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 14:6243105:d=2021032312:TMP:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: + product_functions.sh[10]rc=0 -22: + product_functions.sh[11](( rc == 0 )) -22: + product_functions.sh[11]mv pgb2bfile_f024_23_0p25.new pgb2bfile_f024_23_0p25 -22: + product_functions.sh[12]return 0 -22: + interp_atmos_master.sh[56]export err=0 -22: + interp_atmos_master.sh[56]err=0 -22: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -22: ++ interp_atmos_master.sh[62]wc -l -22: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_23_0p25 -match 'LAND|ICEC' -22: + interp_atmos_master.sh[62]var_count=0 -22: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -22: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -22: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_23_0p50 -22: + product_functions.sh[5]local filename=pgb2bfile_f024_23_0p50 -22: + product_functions.sh[6]wgrib2 pgb2bfile_f024_23_0p50 -not_if :RH: -grib pgb2bfile_f024_23_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_23_0p50.new -22: 1:0:d=2021032312:TMP:180-150 mb above ground:24 hour fcst: -22: 2:175206:d=2021032312:RH:180-150 mb above ground:24 hour fcst: -22: 3:336688:d=2021032312:SPFH:180-150 mb above ground:24 hour fcst: -22: 4:614635:d=2021032312:UGRD:180-150 mb above ground:24 hour fcst: -22: 5:788214:d=2021032312:VGRD:180-150 mb above ground:24 hour fcst: -22: 6:970195:d=2021032312:UGRD:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 7:1067692:d=2021032312:VGRD:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 8:1163575:d=2021032312:TMP:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 9:1309816:d=2021032312:HGT:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 10:1483093:d=2021032312:PRES:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 11:1657418:d=2021032312:VWSH:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 12:1760668:d=2021032312:UGRD:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 13:1906793:d=2021032312:VGRD:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 14:2050016:d=2021032312:TMP:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: + product_functions.sh[10]rc=0 -22: + product_functions.sh[11](( rc == 0 )) -22: + product_functions.sh[11]mv pgb2bfile_f024_23_0p50.new pgb2bfile_f024_23_0p50 -22: + product_functions.sh[12]return 0 -22: + interp_atmos_master.sh[56]export err=0 -22: + interp_atmos_master.sh[56]err=0 -22: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -22: ++ interp_atmos_master.sh[62]wc -l -22: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_23_0p50 -match 'LAND|ICEC' -22: + interp_atmos_master.sh[62]var_count=0 -22: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -22: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -22: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_23_1p00 -22: + product_functions.sh[5]local filename=pgb2bfile_f024_23_1p00 -22: + product_functions.sh[6]wgrib2 pgb2bfile_f024_23_1p00 -not_if :RH: -grib pgb2bfile_f024_23_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_23_1p00.new -22: 1:0:d=2021032312:TMP:180-150 mb above ground:24 hour fcst: -22: 2:55911:d=2021032312:RH:180-150 mb above ground:24 hour fcst: -22: 3:108610:d=2021032312:SPFH:180-150 mb above ground:24 hour fcst: -22: 4:190923:d=2021032312:UGRD:180-150 mb above ground:24 hour fcst: -22: 5:246695:d=2021032312:VGRD:180-150 mb above ground:24 hour fcst: -22: 6:304551:d=2021032312:UGRD:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 7:334478:d=2021032312:VGRD:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 8:363968:d=2021032312:TMP:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 9:406886:d=2021032312:HGT:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 10:456994:d=2021032312:PRES:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 11:507227:d=2021032312:VWSH:PV=5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 12:538043:d=2021032312:UGRD:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 13:581115:d=2021032312:VGRD:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: 14:623580:d=2021032312:TMP:PV=-5e-07 (Km^2/kg/s) surface:24 hour fcst: -22: + product_functions.sh[10]rc=0 -22: + product_functions.sh[11](( rc == 0 )) -22: + product_functions.sh[11]mv pgb2bfile_f024_23_1p00.new pgb2bfile_f024_23_1p00 -22: + product_functions.sh[12]return 0 -22: + interp_atmos_master.sh[56]export err=0 -22: + interp_atmos_master.sh[56]err=0 -22: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -22: ++ interp_atmos_master.sh[62]wc -l -22: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_23_1p00 -match 'LAND|ICEC' -22: + interp_atmos_master.sh[62]var_count=0 -22: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -22: + interp_atmos_master.sh[73]exit 0 - 1: + bash[8]'[' -z '' ']' - 1: + bash[9]case "$-" in - 1: + bash[12]__lmod_vx=x - 1: + bash[16]'[' -n x ']' - 1: + bash[16]set +x - 1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 1: Shell debugging restarted - 1: + bash[224]unset __lmod_vx - 1: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_2 - 1: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_2 - 1: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 - 1: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 1: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 1: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 1: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 1: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 1: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 1: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 1: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 1: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 1: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 1: + interp_atmos_master.sh[31]IFS=: - 1: + interp_atmos_master.sh[31]read -ra grids - 1: + interp_atmos_master.sh[33]output_grids= - 1: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 1: + interp_atmos_master.sh[35]gridopt=grid0p25 - 1: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_2_0p25' - 1: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 1: + interp_atmos_master.sh[35]gridopt=grid0p50 - 1: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_2_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_2_0p50' - 1: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 1: + interp_atmos_master.sh[35]gridopt=grid1p00 - 1: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_2_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_2_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_2_1p00' - 1: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_2 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_2_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_2_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_2_1p00 - 1: 1:0:d=2021032312:HGT:3 mb:24 hour fcst: - 1: 2:17836:d=2021032312:TMP:3 mb:24 hour fcst: - 1: 3:34309:d=2021032312:RH:3 mb:24 hour fcst: - 1: 4:40339:d=2021032312:UGRD:3 mb:24 hour fcst: - 1: 5:51722:d=2021032312:VGRD:3 mb:24 hour fcst: - 1: 6:69859:d=2021032312:ABSV:3 mb:24 hour fcst: - 1: 7:86873:d=2021032312:O3MR:3 mb:24 hour fcst: - 1: 8:110246:d=2021032312:HGT:5 mb:24 hour fcst: - 1: 9:127921:d=2021032312:TMP:5 mb:24 hour fcst: - 1: 10:144550:d=2021032312:RH:5 mb:24 hour fcst: - 1: 11:153592:d=2021032312:UGRD:5 mb:24 hour fcst: - 1: 12:165184:d=2021032312:VGRD:5 mb:24 hour fcst: - 1: 13:183754:d=2021032312:ABSV:5 mb:24 hour fcst: - 1: 14:201375:d=2021032312:O3MR:5 mb:24 hour fcst: - 1: + interp_atmos_master.sh[47]export err=0 - 1: + interp_atmos_master.sh[47]err=0 - 1: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 1: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 1: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_2_0p25 - 1: + product_functions.sh[5]local filename=pgb2bfile_f024_2_0p25 - 1: + product_functions.sh[6]wgrib2 pgb2bfile_f024_2_0p25 -not_if :RH: -grib pgb2bfile_f024_2_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_2_0p25.new - 1: 1:0:d=2021032312:HGT:3 mb:24 hour fcst: - 1: 2:401725:d=2021032312:TMP:3 mb:24 hour fcst: - 1: 3:765990:d=2021032312:RH:3 mb:24 hour fcst: - 1: 4:936438:d=2021032312:UGRD:3 mb:24 hour fcst: - 1: 5:1381348:d=2021032312:VGRD:3 mb:24 hour fcst: - 1: 6:1796037:d=2021032312:ABSV:3 mb:24 hour fcst: - 1: 7:2159740:d=2021032312:O3MR:3 mb:24 hour fcst: - 1: 8:2764021:d=2021032312:HGT:5 mb:24 hour fcst: - 1: 9:3163953:d=2021032312:TMP:5 mb:24 hour fcst: - 1: 10:3529628:d=2021032312:RH:5 mb:24 hour fcst: - 1: 11:3780958:d=2021032312:UGRD:5 mb:24 hour fcst: - 1: 12:4236857:d=2021032312:VGRD:5 mb:24 hour fcst: - 1: 13:4664622:d=2021032312:ABSV:5 mb:24 hour fcst: - 1: 14:5051673:d=2021032312:O3MR:5 mb:24 hour fcst: - 1: + product_functions.sh[10]rc=0 - 1: + product_functions.sh[11](( rc == 0 )) - 1: + product_functions.sh[11]mv pgb2bfile_f024_2_0p25.new pgb2bfile_f024_2_0p25 - 1: + product_functions.sh[12]return 0 - 1: + interp_atmos_master.sh[56]export err=0 - 1: + interp_atmos_master.sh[56]err=0 - 1: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 1: ++ interp_atmos_master.sh[62]wc -l - 1: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_2_0p25 -match 'LAND|ICEC' - 1: + interp_atmos_master.sh[62]var_count=0 - 1: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 1: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 1: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_2_0p50 - 1: + product_functions.sh[5]local filename=pgb2bfile_f024_2_0p50 - 1: + product_functions.sh[6]wgrib2 pgb2bfile_f024_2_0p50 -not_if :RH: -grib pgb2bfile_f024_2_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_2_0p50.new - 1: 1:0:d=2021032312:HGT:3 mb:24 hour fcst: - 1: 2:137394:d=2021032312:TMP:3 mb:24 hour fcst: - 1: 3:261879:d=2021032312:RH:3 mb:24 hour fcst: - 1: 4:318472:d=2021032312:UGRD:3 mb:24 hour fcst: - 1: 5:473051:d=2021032312:VGRD:3 mb:24 hour fcst: - 1: 6:616286:d=2021032312:ABSV:3 mb:24 hour fcst: - 1: 7:742264:d=2021032312:O3MR:3 mb:24 hour fcst: - 1: 8:946570:d=2021032312:HGT:5 mb:24 hour fcst: - 1: 9:1082925:d=2021032312:TMP:5 mb:24 hour fcst: - 1: 10:1208637:d=2021032312:RH:5 mb:24 hour fcst: - 1: 11:1277431:d=2021032312:UGRD:5 mb:24 hour fcst: - 1: 12:1435562:d=2021032312:VGRD:5 mb:24 hour fcst: - 1: 13:1583917:d=2021032312:ABSV:5 mb:24 hour fcst: - 1: 14:1716733:d=2021032312:O3MR:5 mb:24 hour fcst: - 1: + product_functions.sh[10]rc=0 - 1: + product_functions.sh[11](( rc == 0 )) - 1: + product_functions.sh[11]mv pgb2bfile_f024_2_0p50.new pgb2bfile_f024_2_0p50 - 1: + product_functions.sh[12]return 0 - 1: + interp_atmos_master.sh[56]export err=0 - 1: + interp_atmos_master.sh[56]err=0 - 1: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 1: ++ interp_atmos_master.sh[62]wc -l - 1: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_2_0p50 -match 'LAND|ICEC' - 1: + interp_atmos_master.sh[62]var_count=0 - 1: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 1: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 1: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_2_1p00 - 1: + product_functions.sh[5]local filename=pgb2bfile_f024_2_1p00 - 1: + product_functions.sh[6]wgrib2 pgb2bfile_f024_2_1p00 -not_if :RH: -grib pgb2bfile_f024_2_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_2_1p00.new - 1: 1:0:d=2021032312:HGT:3 mb:24 hour fcst: - 1: 2:44979:d=2021032312:TMP:3 mb:24 hour fcst: - 1: 3:86333:d=2021032312:RH:3 mb:24 hour fcst: - 1: 4:102969:d=2021032312:UGRD:3 mb:24 hour fcst: - 1: 5:153359:d=2021032312:VGRD:3 mb:24 hour fcst: - 1: 6:200366:d=2021032312:ABSV:3 mb:24 hour fcst: - 1: 7:242805:d=2021032312:O3MR:3 mb:24 hour fcst: - 1: 8:306494:d=2021032312:HGT:5 mb:24 hour fcst: - 1: 9:350949:d=2021032312:TMP:5 mb:24 hour fcst: - 1: 10:392735:d=2021032312:RH:5 mb:24 hour fcst: - 1: 11:413494:d=2021032312:UGRD:5 mb:24 hour fcst: - 1: 12:464574:d=2021032312:VGRD:5 mb:24 hour fcst: - 1: 13:513482:d=2021032312:ABSV:5 mb:24 hour fcst: - 1: 14:557909:d=2021032312:O3MR:5 mb:24 hour fcst: - 1: + product_functions.sh[10]rc=0 - 1: + product_functions.sh[11](( rc == 0 )) - 1: + product_functions.sh[11]mv pgb2bfile_f024_2_1p00.new pgb2bfile_f024_2_1p00 - 1: + product_functions.sh[12]return 0 - 1: + interp_atmos_master.sh[56]export err=0 - 1: + interp_atmos_master.sh[56]err=0 - 1: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 1: ++ interp_atmos_master.sh[62]wc -l - 1: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_2_1p00 -match 'LAND|ICEC' - 1: + interp_atmos_master.sh[62]var_count=0 - 1: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 1: + interp_atmos_master.sh[73]exit 0 - 7: + bash[8]'[' -z '' ']' - 7: + bash[9]case "$-" in - 7: + bash[12]__lmod_vx=x - 7: + bash[16]'[' -n x ']' - 7: + bash[16]set +x - 7: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 7: Shell debugging restarted - 7: + bash[224]unset __lmod_vx - 7: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_8 - 7: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_8 - 7: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 - 7: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 7: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 7: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 7: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 7: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 7: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 7: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 7: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 7: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 7: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 7: + interp_atmos_master.sh[31]IFS=: - 7: + interp_atmos_master.sh[31]read -ra grids - 7: + interp_atmos_master.sh[33]output_grids= - 7: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 7: + interp_atmos_master.sh[35]gridopt=grid0p25 - 7: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_8_0p25' - 7: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 7: + interp_atmos_master.sh[35]gridopt=grid0p50 - 7: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_8_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_8_0p50' - 7: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 7: + interp_atmos_master.sh[35]gridopt=grid1p00 - 7: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_8_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_8_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_8_1p00' - 7: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_8 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_8_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_8_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_8_1p00 - 7: 1:0:d=2021032312:TMP:325 mb:24 hour fcst: - 7: 2:18366:d=2021032312:RH:325 mb:24 hour fcst: - 7: 3:37389:d=2021032312:TCDC:325 mb:24 hour fcst: - 7: 4:50444:d=2021032312:VVEL:325 mb:24 hour fcst: - 7: 5:74702:d=2021032312:DZDT:325 mb:24 hour fcst: - 7: 6:102946:d=2021032312:UGRD:325 mb:24 hour fcst: - 7: 7:117585:d=2021032312:VGRD:325 mb:24 hour fcst: - 7: 8:132873:d=2021032312:ABSV:325 mb:24 hour fcst: - 7: 9:155195:d=2021032312:CLMR:325 mb:24 hour fcst: - 7: 10:155374:d=2021032312:ICMR:325 mb:24 hour fcst: - 7: 11:171877:d=2021032312:RWMR:325 mb:24 hour fcst: - 7: 12:172056:d=2021032312:SNMR:325 mb:24 hour fcst: - 7: 13:183843:d=2021032312:GRLE:325 mb:24 hour fcst: - 7: 14:184279:d=2021032312:HGT:375 mb:24 hour fcst: - 7: + interp_atmos_master.sh[47]export err=0 - 7: + interp_atmos_master.sh[47]err=0 - 7: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 7: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 7: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_8_0p25 - 7: + product_functions.sh[5]local filename=pgb2bfile_f024_8_0p25 - 7: + product_functions.sh[6]wgrib2 pgb2bfile_f024_8_0p25 -not_if :RH: -grib pgb2bfile_f024_8_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_8_0p25.new - 7: 1:0:d=2021032312:TMP:325 mb:24 hour fcst: - 7: 2:406669:d=2021032312:RH:325 mb:24 hour fcst: - 7: 3:833353:d=2021032312:TCDC:325 mb:24 hour fcst: - 7: 4:1073529:d=2021032312:VVEL:325 mb:24 hour fcst: - 7: 5:1719317:d=2021032312:DZDT:325 mb:24 hour fcst: - 7: 6:2516861:d=2021032312:UGRD:325 mb:24 hour fcst: - 7: 7:2831782:d=2021032312:VGRD:325 mb:24 hour fcst: - 7: 8:3165625:d=2021032312:ABSV:325 mb:24 hour fcst: - 7: 9:3717664:d=2021032312:CLMR:325 mb:24 hour fcst: - 7: 10:3717843:d=2021032312:ICMR:325 mb:24 hour fcst: - 7: 11:4082998:d=2021032312:RWMR:325 mb:24 hour fcst: - 7: 12:4083177:d=2021032312:SNMR:325 mb:24 hour fcst: - 7: 13:4321643:d=2021032312:GRLE:325 mb:24 hour fcst: - 7: 14:4327445:d=2021032312:HGT:375 mb:24 hour fcst: - 7: + product_functions.sh[10]rc=0 - 7: + product_functions.sh[11](( rc == 0 )) - 7: + product_functions.sh[11]mv pgb2bfile_f024_8_0p25.new pgb2bfile_f024_8_0p25 - 7: + product_functions.sh[12]return 0 - 7: + interp_atmos_master.sh[56]export err=0 - 7: + interp_atmos_master.sh[56]err=0 - 7: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 7: ++ interp_atmos_master.sh[62]wc -l - 7: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_8_0p25 -match 'LAND|ICEC' - 7: + interp_atmos_master.sh[62]var_count=0 - 7: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 7: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 7: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_8_0p50 - 7: + product_functions.sh[5]local filename=pgb2bfile_f024_8_0p50 - 7: + product_functions.sh[6]wgrib2 pgb2bfile_f024_8_0p50 -not_if :RH: -grib pgb2bfile_f024_8_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_8_0p50.new - 7: 1:0:d=2021032312:TMP:325 mb:24 hour fcst: - 7: 2:142114:d=2021032312:RH:325 mb:24 hour fcst: - 7: 3:291252:d=2021032312:TCDC:325 mb:24 hour fcst: - 7: 4:385738:d=2021032312:VVEL:325 mb:24 hour fcst: - 7: 5:605107:d=2021032312:DZDT:325 mb:24 hour fcst: - 7: 6:863224:d=2021032312:UGRD:325 mb:24 hour fcst: - 7: 7:967369:d=2021032312:VGRD:325 mb:24 hour fcst: - 7: 8:1078930:d=2021032312:ABSV:325 mb:24 hour fcst: - 7: 9:1267907:d=2021032312:CLMR:325 mb:24 hour fcst: - 7: 10:1268086:d=2021032312:ICMR:325 mb:24 hour fcst: - 7: 11:1401254:d=2021032312:RWMR:325 mb:24 hour fcst: - 7: 12:1401433:d=2021032312:SNMR:325 mb:24 hour fcst: - 7: 13:1487730:d=2021032312:GRLE:325 mb:24 hour fcst: - 7: 14:1490044:d=2021032312:HGT:375 mb:24 hour fcst: - 7: + product_functions.sh[10]rc=0 - 7: + product_functions.sh[11](( rc == 0 )) - 7: + product_functions.sh[11]mv pgb2bfile_f024_8_0p50.new pgb2bfile_f024_8_0p50 - 7: + product_functions.sh[12]return 0 - 7: + interp_atmos_master.sh[56]export err=0 - 7: + interp_atmos_master.sh[56]err=0 - 7: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 7: ++ interp_atmos_master.sh[62]wc -l - 7: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_8_0p50 -match 'LAND|ICEC' - 7: + interp_atmos_master.sh[62]var_count=0 - 7: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 7: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 7: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_8_1p00 - 7: + product_functions.sh[5]local filename=pgb2bfile_f024_8_1p00 - 7: + product_functions.sh[6]wgrib2 pgb2bfile_f024_8_1p00 -not_if :RH: -grib pgb2bfile_f024_8_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_8_1p00.new - 7: 1:0:d=2021032312:TMP:325 mb:24 hour fcst: - 7: 2:46976:d=2021032312:RH:325 mb:24 hour fcst: - 7: 3:95687:d=2021032312:TCDC:325 mb:24 hour fcst: - 7: 4:128870:d=2021032312:VVEL:325 mb:24 hour fcst: - 7: 5:196140:d=2021032312:DZDT:325 mb:24 hour fcst: - 7: 6:273538:d=2021032312:UGRD:325 mb:24 hour fcst: - 7: 7:309326:d=2021032312:VGRD:325 mb:24 hour fcst: - 7: 8:347487:d=2021032312:ABSV:325 mb:24 hour fcst: - 7: 9:407355:d=2021032312:CLMR:325 mb:24 hour fcst: - 7: 10:407534:d=2021032312:ICMR:325 mb:24 hour fcst: - 7: 11:451440:d=2021032312:RWMR:325 mb:24 hour fcst: - 7: 12:451619:d=2021032312:SNMR:325 mb:24 hour fcst: - 7: 13:481479:d=2021032312:GRLE:325 mb:24 hour fcst: - 7: 14:482438:d=2021032312:HGT:375 mb:24 hour fcst: - 7: + product_functions.sh[10]rc=0 - 7: + product_functions.sh[11](( rc == 0 )) - 7: + product_functions.sh[11]mv pgb2bfile_f024_8_1p00.new pgb2bfile_f024_8_1p00 - 7: + product_functions.sh[12]return 0 - 7: + interp_atmos_master.sh[56]export err=0 - 7: + interp_atmos_master.sh[56]err=0 - 7: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 7: ++ interp_atmos_master.sh[62]wc -l - 7: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_8_1p00 -match 'LAND|ICEC' - 7: + interp_atmos_master.sh[62]var_count=0 - 7: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 7: + interp_atmos_master.sh[73]exit 0 -21: + bash[8]'[' -z '' ']' -21: + bash[9]case "$-" in -21: + bash[12]__lmod_vx=x -21: + bash[16]'[' -n x ']' -21: + bash[16]set +x -21: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -21: Shell debugging restarted -21: + bash[224]unset __lmod_vx -21: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_22 -21: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_22 -21: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -21: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -21: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -21: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -21: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -21: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -21: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -21: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -21: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -21: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -21: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -21: + interp_atmos_master.sh[31]IFS=: -21: + interp_atmos_master.sh[31]read -ra grids -21: + interp_atmos_master.sh[33]output_grids= -21: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -21: + interp_atmos_master.sh[35]gridopt=grid0p25 -21: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_22_0p25' -21: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -21: + interp_atmos_master.sh[35]gridopt=grid0p50 -21: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_22_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_22_0p50' -21: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -21: + interp_atmos_master.sh[35]gridopt=grid1p00 -21: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_22_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_22_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_22_1p00' -21: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_22 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_22_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_22_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_22_1p00 -21: 1:0:d=2021032312:RH:90-60 mb above ground:24 hour fcst: -21: 2:19113:d=2021032312:SPFH:90-60 mb above ground:24 hour fcst: -21: 3:48028:d=2021032312:UGRD:90-60 mb above ground:24 hour fcst: -21: 4:69414:d=2021032312:VGRD:90-60 mb above ground:24 hour fcst: -21: 5:91425:d=2021032312:TMP:120-90 mb above ground:24 hour fcst: -21: 6:112369:d=2021032312:RH:120-90 mb above ground:24 hour fcst: -21: 7:131766:d=2021032312:SPFH:120-90 mb above ground:24 hour fcst: -21: 8:160664:d=2021032312:UGRD:120-90 mb above ground:24 hour fcst: -21: 9:181793:d=2021032312:VGRD:120-90 mb above ground:24 hour fcst: -21: 10:203606:d=2021032312:TMP:150-120 mb above ground:24 hour fcst: -21: 11:224546:d=2021032312:RH:150-120 mb above ground:24 hour fcst: -21: 12:244240:d=2021032312:SPFH:150-120 mb above ground:24 hour fcst: -21: 13:273374:d=2021032312:UGRD:150-120 mb above ground:24 hour fcst: -21: 14:294409:d=2021032312:VGRD:150-120 mb above ground:24 hour fcst: -21: + interp_atmos_master.sh[47]export err=0 -21: + interp_atmos_master.sh[47]err=0 -21: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -21: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -21: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_22_0p25 -21: + product_functions.sh[5]local filename=pgb2bfile_f024_22_0p25 -21: + product_functions.sh[6]wgrib2 pgb2bfile_f024_22_0p25 -not_if :RH: -grib pgb2bfile_f024_22_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_22_0p25.new -21: 1:0:d=2021032312:RH:90-60 mb above ground:24 hour fcst: -21: 2:426220:d=2021032312:SPFH:90-60 mb above ground:24 hour fcst: -21: 3:1216459:d=2021032312:UGRD:90-60 mb above ground:24 hour fcst: -21: 4:1731196:d=2021032312:VGRD:90-60 mb above ground:24 hour fcst: -21: 5:2277890:d=2021032312:TMP:120-90 mb above ground:24 hour fcst: -21: 6:2776186:d=2021032312:RH:120-90 mb above ground:24 hour fcst: -21: 7:3209014:d=2021032312:SPFH:120-90 mb above ground:24 hour fcst: -21: 8:3997554:d=2021032312:UGRD:120-90 mb above ground:24 hour fcst: -21: 9:4504983:d=2021032312:VGRD:120-90 mb above ground:24 hour fcst: -21: 10:5044681:d=2021032312:TMP:150-120 mb above ground:24 hour fcst: -21: 11:5546594:d=2021032312:RH:150-120 mb above ground:24 hour fcst: -21: 12:5994695:d=2021032312:SPFH:150-120 mb above ground:24 hour fcst: -21: 13:6787079:d=2021032312:UGRD:150-120 mb above ground:24 hour fcst: -21: 14:7290518:d=2021032312:VGRD:150-120 mb above ground:24 hour fcst: -21: + product_functions.sh[10]rc=0 -21: + product_functions.sh[11](( rc == 0 )) -21: + product_functions.sh[11]mv pgb2bfile_f024_22_0p25.new pgb2bfile_f024_22_0p25 -21: + product_functions.sh[12]return 0 -21: + interp_atmos_master.sh[56]export err=0 -21: + interp_atmos_master.sh[56]err=0 -21: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -21: ++ interp_atmos_master.sh[62]wc -l -21: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_22_0p25 -match 'LAND|ICEC' -21: + interp_atmos_master.sh[62]var_count=0 -21: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -21: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -21: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_22_0p50 -21: + product_functions.sh[5]local filename=pgb2bfile_f024_22_0p50 -21: + product_functions.sh[6]wgrib2 pgb2bfile_f024_22_0p50 -not_if :RH: -grib pgb2bfile_f024_22_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_22_0p50.new -21: 1:0:d=2021032312:RH:90-60 mb above ground:24 hour fcst: -21: 2:150439:d=2021032312:SPFH:90-60 mb above ground:24 hour fcst: -21: 3:423302:d=2021032312:UGRD:90-60 mb above ground:24 hour fcst: -21: 4:600557:d=2021032312:VGRD:90-60 mb above ground:24 hour fcst: -21: 5:785761:d=2021032312:TMP:120-90 mb above ground:24 hour fcst: -21: 6:957960:d=2021032312:RH:120-90 mb above ground:24 hour fcst: -21: 7:1111037:d=2021032312:SPFH:120-90 mb above ground:24 hour fcst: -21: 8:1383475:d=2021032312:UGRD:120-90 mb above ground:24 hour fcst: -21: 9:1558300:d=2021032312:VGRD:120-90 mb above ground:24 hour fcst: -21: 10:1741325:d=2021032312:TMP:150-120 mb above ground:24 hour fcst: -21: 11:1914799:d=2021032312:RH:150-120 mb above ground:24 hour fcst: -21: 12:2072671:d=2021032312:SPFH:150-120 mb above ground:24 hour fcst: -21: 13:2347658:d=2021032312:UGRD:150-120 mb above ground:24 hour fcst: -21: 14:2521227:d=2021032312:VGRD:150-120 mb above ground:24 hour fcst: -21: + product_functions.sh[10]rc=0 -21: + product_functions.sh[11](( rc == 0 )) -21: + product_functions.sh[11]mv pgb2bfile_f024_22_0p50.new pgb2bfile_f024_22_0p50 -21: + product_functions.sh[12]return 0 -21: + interp_atmos_master.sh[56]export err=0 -21: + interp_atmos_master.sh[56]err=0 -21: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -21: ++ interp_atmos_master.sh[62]wc -l -21: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_22_0p50 -match 'LAND|ICEC' -21: + interp_atmos_master.sh[62]var_count=0 -21: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -21: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -21: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_22_1p00 -21: + product_functions.sh[5]local filename=pgb2bfile_f024_22_1p00 -21: + product_functions.sh[6]wgrib2 pgb2bfile_f024_22_1p00 -not_if :RH: -grib pgb2bfile_f024_22_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_22_1p00.new -21: 1:0:d=2021032312:RH:90-60 mb above ground:24 hour fcst: -21: 2:49350:d=2021032312:SPFH:90-60 mb above ground:24 hour fcst: -21: 3:130558:d=2021032312:UGRD:90-60 mb above ground:24 hour fcst: -21: 4:187557:d=2021032312:VGRD:90-60 mb above ground:24 hour fcst: -21: 5:246756:d=2021032312:TMP:120-90 mb above ground:24 hour fcst: -21: 6:301988:d=2021032312:RH:120-90 mb above ground:24 hour fcst: -21: 7:352352:d=2021032312:SPFH:120-90 mb above ground:24 hour fcst: -21: 8:433204:d=2021032312:UGRD:120-90 mb above ground:24 hour fcst: -21: 9:489410:d=2021032312:VGRD:120-90 mb above ground:24 hour fcst: -21: 10:548036:d=2021032312:TMP:150-120 mb above ground:24 hour fcst: -21: 11:603458:d=2021032312:RH:150-120 mb above ground:24 hour fcst: -21: 12:655144:d=2021032312:SPFH:150-120 mb above ground:24 hour fcst: -21: 13:736835:d=2021032312:UGRD:150-120 mb above ground:24 hour fcst: -21: 14:792672:d=2021032312:VGRD:150-120 mb above ground:24 hour fcst: -21: + product_functions.sh[10]rc=0 -21: + product_functions.sh[11](( rc == 0 )) -21: + product_functions.sh[11]mv pgb2bfile_f024_22_1p00.new pgb2bfile_f024_22_1p00 -21: + product_functions.sh[12]return 0 -21: + interp_atmos_master.sh[56]export err=0 -21: + interp_atmos_master.sh[56]err=0 -21: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -21: ++ interp_atmos_master.sh[62]wc -l -21: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_22_1p00 -match 'LAND|ICEC' -21: + interp_atmos_master.sh[62]var_count=0 -21: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -21: + interp_atmos_master.sh[73]exit 0 -12: + bash[8]'[' -z '' ']' -12: + bash[9]case "$-" in -12: + bash[12]__lmod_vx=x -12: + bash[16]'[' -n x ']' -12: + bash[16]set +x -12: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -12: Shell debugging restarted -12: + bash[224]unset __lmod_vx -12: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_13 -12: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_13 -12: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -12: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -12: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -12: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -12: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -12: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -12: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -12: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -12: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -12: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -12: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -12: + interp_atmos_master.sh[31]IFS=: -12: + interp_atmos_master.sh[31]read -ra grids -12: + interp_atmos_master.sh[33]output_grids= -12: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -12: + interp_atmos_master.sh[35]gridopt=grid0p25 -12: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_13_0p25' -12: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -12: + interp_atmos_master.sh[35]gridopt=grid0p50 -12: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_13_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_13_0p50' -12: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -12: + interp_atmos_master.sh[35]gridopt=grid1p00 -12: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_13_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_13_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_13_1p00' -12: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_13 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_13_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_13_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_13_1p00 -12: 1:0:d=2021032312:TMP:575 mb:24 hour fcst: -12: 2:18888:d=2021032312:RH:575 mb:24 hour fcst: -12: 3:38823:d=2021032312:TCDC:575 mb:24 hour fcst: -12: 4:51135:d=2021032312:VVEL:575 mb:24 hour fcst: -12: 5:76264:d=2021032312:DZDT:575 mb:24 hour fcst: -12: 6:103674:d=2021032312:UGRD:575 mb:24 hour fcst: -12: 7:124602:d=2021032312:VGRD:575 mb:24 hour fcst: -12: 8:146251:d=2021032312:ABSV:575 mb:24 hour fcst: -12: 9:167122:d=2021032312:CLMR:575 mb:24 hour fcst: -12: 10:171258:d=2021032312:ICMR:575 mb:24 hour fcst: -12: 11:187195:d=2021032312:RWMR:575 mb:24 hour fcst: -12: 12:189653:d=2021032312:SNMR:575 mb:24 hour fcst: -12: 13:204195:d=2021032312:GRLE:575 mb:24 hour fcst: -12: 14:206406:d=2021032312:HGT:625 mb:24 hour fcst: -12: + interp_atmos_master.sh[47]export err=0 -12: + interp_atmos_master.sh[47]err=0 -12: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -12: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -12: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_13_0p25 -12: + product_functions.sh[5]local filename=pgb2bfile_f024_13_0p25 -12: + product_functions.sh[6]wgrib2 pgb2bfile_f024_13_0p25 -not_if :RH: -grib pgb2bfile_f024_13_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_13_0p25.new -12: 1:0:d=2021032312:TMP:575 mb:24 hour fcst: -12: 2:423626:d=2021032312:RH:575 mb:24 hour fcst: -12: 3:879685:d=2021032312:TCDC:575 mb:24 hour fcst: -12: 4:1107981:d=2021032312:VVEL:575 mb:24 hour fcst: -12: 5:1779404:d=2021032312:DZDT:575 mb:24 hour fcst: -12: 6:2546742:d=2021032312:UGRD:575 mb:24 hour fcst: -12: 7:3056521:d=2021032312:VGRD:575 mb:24 hour fcst: -12: 8:3594673:d=2021032312:ABSV:575 mb:24 hour fcst: -12: 9:4081684:d=2021032312:CLMR:575 mb:24 hour fcst: -12: 10:4170086:d=2021032312:ICMR:575 mb:24 hour fcst: -12: 11:4522290:d=2021032312:RWMR:575 mb:24 hour fcst: -12: 12:4569894:d=2021032312:SNMR:575 mb:24 hour fcst: -12: 13:4884574:d=2021032312:GRLE:575 mb:24 hour fcst: -12: 14:4924557:d=2021032312:HGT:625 mb:24 hour fcst: -12: + product_functions.sh[10]rc=0 -12: + product_functions.sh[11](( rc == 0 )) -12: + product_functions.sh[11]mv pgb2bfile_f024_13_0p25.new pgb2bfile_f024_13_0p25 -12: + product_functions.sh[12]return 0 -12: + interp_atmos_master.sh[56]export err=0 -12: + interp_atmos_master.sh[56]err=0 -12: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -12: ++ interp_atmos_master.sh[62]wc -l -12: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_13_0p25 -match 'LAND|ICEC' -12: + interp_atmos_master.sh[62]var_count=0 -12: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -12: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -12: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_13_0p50 -12: + product_functions.sh[5]local filename=pgb2bfile_f024_13_0p50 -12: + product_functions.sh[6]wgrib2 pgb2bfile_f024_13_0p50 -not_if :RH: -grib pgb2bfile_f024_13_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_13_0p50.new -12: 1:0:d=2021032312:TMP:575 mb:24 hour fcst: -12: 2:148021:d=2021032312:RH:575 mb:24 hour fcst: -12: 3:307734:d=2021032312:TCDC:575 mb:24 hour fcst: -12: 4:394509:d=2021032312:VVEL:575 mb:24 hour fcst: -12: 5:623834:d=2021032312:DZDT:575 mb:24 hour fcst: -12: 6:874975:d=2021032312:UGRD:575 mb:24 hour fcst: -12: 7:1048269:d=2021032312:VGRD:575 mb:24 hour fcst: -12: 8:1231708:d=2021032312:ABSV:575 mb:24 hour fcst: -12: 9:1402812:d=2021032312:CLMR:575 mb:24 hour fcst: -12: 10:1436136:d=2021032312:ICMR:575 mb:24 hour fcst: -12: 11:1563097:d=2021032312:RWMR:575 mb:24 hour fcst: -12: 12:1580322:d=2021032312:SNMR:575 mb:24 hour fcst: -12: 13:1691483:d=2021032312:GRLE:575 mb:24 hour fcst: -12: 14:1706828:d=2021032312:HGT:625 mb:24 hour fcst: -12: + product_functions.sh[10]rc=0 -12: + product_functions.sh[11](( rc == 0 )) -12: + product_functions.sh[11]mv pgb2bfile_f024_13_0p50.new pgb2bfile_f024_13_0p50 -12: + product_functions.sh[12]return 0 -12: + interp_atmos_master.sh[56]export err=0 -12: + interp_atmos_master.sh[56]err=0 -12: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -12: ++ interp_atmos_master.sh[62]wc -l -12: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_13_0p50 -match 'LAND|ICEC' -12: + interp_atmos_master.sh[62]var_count=0 -12: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -12: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -12: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_13_1p00 -12: + product_functions.sh[5]local filename=pgb2bfile_f024_13_1p00 -12: + product_functions.sh[6]wgrib2 pgb2bfile_f024_13_1p00 -not_if :RH: -grib pgb2bfile_f024_13_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_13_1p00.new -12: 1:0:d=2021032312:TMP:575 mb:24 hour fcst: -12: 2:48389:d=2021032312:RH:575 mb:24 hour fcst: -12: 3:100518:d=2021032312:TCDC:575 mb:24 hour fcst: -12: 4:131661:d=2021032312:VVEL:575 mb:24 hour fcst: -12: 5:202004:d=2021032312:DZDT:575 mb:24 hour fcst: -12: 6:277544:d=2021032312:UGRD:575 mb:24 hour fcst: -12: 7:333670:d=2021032312:VGRD:575 mb:24 hour fcst: -12: 8:392085:d=2021032312:ABSV:575 mb:24 hour fcst: -12: 9:447268:d=2021032312:CLMR:575 mb:24 hour fcst: -12: 10:458965:d=2021032312:ICMR:575 mb:24 hour fcst: -12: 11:501195:d=2021032312:RWMR:575 mb:24 hour fcst: -12: 12:507553:d=2021032312:SNMR:575 mb:24 hour fcst: -12: 13:545113:d=2021032312:GRLE:575 mb:24 hour fcst: -12: 14:550726:d=2021032312:HGT:625 mb:24 hour fcst: -12: + product_functions.sh[10]rc=0 -12: + product_functions.sh[11](( rc == 0 )) -12: + product_functions.sh[11]mv pgb2bfile_f024_13_1p00.new pgb2bfile_f024_13_1p00 -12: + product_functions.sh[12]return 0 -12: + interp_atmos_master.sh[56]export err=0 -12: + interp_atmos_master.sh[56]err=0 -12: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -12: ++ interp_atmos_master.sh[62]wc -l -12: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_13_1p00 -match 'LAND|ICEC' -12: + interp_atmos_master.sh[62]var_count=0 -12: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -12: + interp_atmos_master.sh[73]exit 0 -19: + bash[8]'[' -z '' ']' -19: + bash[9]case "$-" in -19: + bash[12]__lmod_vx=x -19: + bash[16]'[' -n x ']' -19: + bash[16]set +x -19: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -19: Shell debugging restarted -19: + bash[224]unset __lmod_vx -19: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_20 -19: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_20 -19: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -19: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -19: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -19: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -19: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -19: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -19: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -19: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -19: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -19: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -19: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -19: + interp_atmos_master.sh[31]IFS=: -19: + interp_atmos_master.sh[31]read -ra grids -19: + interp_atmos_master.sh[33]output_grids= -19: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -19: + interp_atmos_master.sh[35]gridopt=grid0p25 -19: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_20_0p25' -19: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -19: + interp_atmos_master.sh[35]gridopt=grid0p50 -19: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_20_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_20_0p50' -19: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -19: + interp_atmos_master.sh[35]gridopt=grid1p00 -19: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_20_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_20_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_20_1p00' -19: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_20 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_20_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_20_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_20_1p00 -19: 1:0:d=2021032312:SOILL:0.1-0.4 m below ground:24 hour fcst: -19: 2:8894:d=2021032312:SOILL:0.4-1 m below ground:24 hour fcst: -19: 3:17798:d=2021032312:SOILL:1-2 m below ground:24 hour fcst: -19: 4:26874:d=2021032312:CNWAT:surface:24 hour fcst: -19: 5:32459:d=2021032312:ICETK:surface:24 hour fcst: -19: 6:35873:d=2021032312:DUVB:surface:18-24 hour ave fcst: -19: 7:56563:d=2021032312:CDUVB:surface:18-24 hour ave fcst: -19: 8:72411:d=2021032312:TMP:305 m above mean sea level:24 hour fcst: -19: 9:90745:d=2021032312:UGRD:305 m above mean sea level:24 hour fcst: -19: 10:109917:d=2021032312:VGRD:305 m above mean sea level:24 hour fcst: -19: 11:129503:d=2021032312:TMP:457 m above mean sea level:24 hour fcst: -19: 12:149009:d=2021032312:UGRD:457 m above mean sea level:24 hour fcst: -19: 13:169239:d=2021032312:VGRD:457 m above mean sea level:24 hour fcst: -19: 14:189794:d=2021032312:TMP:610 m above mean sea level:24 hour fcst: -19: + interp_atmos_master.sh[47]export err=0 -19: + interp_atmos_master.sh[47]err=0 -19: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -19: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -19: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_20_0p25 -19: + product_functions.sh[5]local filename=pgb2bfile_f024_20_0p25 -19: + product_functions.sh[6]wgrib2 pgb2bfile_f024_20_0p25 -not_if :RH: -grib pgb2bfile_f024_20_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_20_0p25.new -19: 1:0:d=2021032312:SOILL:0.1-0.4 m below ground:24 hour fcst: -19: 2:293984:d=2021032312:SOILL:0.4-1 m below ground:24 hour fcst: -19: 3:589530:d=2021032312:SOILL:1-2 m below ground:24 hour fcst: -19: 4:885778:d=2021032312:CNWAT:surface:24 hour fcst: -19: 5:1072333:d=2021032312:ICETK:surface:24 hour fcst: -19: 6:1125571:d=2021032312:DUVB:surface:18-24 hour ave fcst: -19: 7:1655980:d=2021032312:CDUVB:surface:18-24 hour ave fcst: -19: 8:2027862:d=2021032312:TMP:305 m above mean sea level:24 hour fcst: -19: 9:2539637:d=2021032312:UGRD:305 m above mean sea level:24 hour fcst: -19: 10:3099838:d=2021032312:VGRD:305 m above mean sea level:24 hour fcst: -19: 11:3676388:d=2021032312:TMP:457 m above mean sea level:24 hour fcst: -19: 12:4218993:d=2021032312:UGRD:457 m above mean sea level:24 hour fcst: -19: 13:4805435:d=2021032312:VGRD:457 m above mean sea level:24 hour fcst: -19: 14:5409702:d=2021032312:TMP:610 m above mean sea level:24 hour fcst: -19: + product_functions.sh[10]rc=0 -19: + product_functions.sh[11](( rc == 0 )) -19: + product_functions.sh[11]mv pgb2bfile_f024_20_0p25.new pgb2bfile_f024_20_0p25 -19: + product_functions.sh[12]return 0 -19: + interp_atmos_master.sh[56]export err=0 -19: + interp_atmos_master.sh[56]err=0 -19: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -19: ++ interp_atmos_master.sh[62]wc -l -19: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_20_0p25 -match 'LAND|ICEC' -19: + interp_atmos_master.sh[62]var_count=0 -19: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -19: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -19: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_20_0p50 -19: + product_functions.sh[5]local filename=pgb2bfile_f024_20_0p50 -19: + product_functions.sh[6]wgrib2 pgb2bfile_f024_20_0p50 -not_if :RH: -grib pgb2bfile_f024_20_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_20_0p50.new -19: 1:0:d=2021032312:SOILL:0.1-0.4 m below ground:24 hour fcst: -19: 2:88993:d=2021032312:SOILL:0.4-1 m below ground:24 hour fcst: -19: 3:178288:d=2021032312:SOILL:1-2 m below ground:24 hour fcst: -19: 4:267756:d=2021032312:CNWAT:surface:24 hour fcst: -19: 5:321145:d=2021032312:ICETK:surface:24 hour fcst: -19: 6:341409:d=2021032312:DUVB:surface:18-24 hour ave fcst: -19: 7:522231:d=2021032312:CDUVB:surface:18-24 hour ave fcst: -19: 8:647734:d=2021032312:TMP:305 m above mean sea level:24 hour fcst: -19: 9:811601:d=2021032312:UGRD:305 m above mean sea level:24 hour fcst: -19: 10:990535:d=2021032312:VGRD:305 m above mean sea level:24 hour fcst: -19: 11:1175226:d=2021032312:TMP:457 m above mean sea level:24 hour fcst: -19: 12:1349371:d=2021032312:UGRD:457 m above mean sea level:24 hour fcst: -19: 13:1536750:d=2021032312:VGRD:457 m above mean sea level:24 hour fcst: -19: 14:1729642:d=2021032312:TMP:610 m above mean sea level:24 hour fcst: -19: + product_functions.sh[10]rc=0 -19: + product_functions.sh[11](( rc == 0 )) -19: + product_functions.sh[11]mv pgb2bfile_f024_20_0p50.new pgb2bfile_f024_20_0p50 -19: + product_functions.sh[12]return 0 -19: + interp_atmos_master.sh[56]export err=0 -19: + interp_atmos_master.sh[56]err=0 -19: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -19: ++ interp_atmos_master.sh[62]wc -l -19: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_20_0p50 -match 'LAND|ICEC' -19: + interp_atmos_master.sh[62]var_count=0 -19: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -19: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -19: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_20_1p00 -19: + product_functions.sh[5]local filename=pgb2bfile_f024_20_1p00 -19: + product_functions.sh[6]wgrib2 pgb2bfile_f024_20_1p00 -not_if :RH: -grib pgb2bfile_f024_20_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_20_1p00.new -19: 1:0:d=2021032312:SOILL:0.1-0.4 m below ground:24 hour fcst: -19: 2:25977:d=2021032312:SOILL:0.4-1 m below ground:24 hour fcst: -19: 3:52060:d=2021032312:SOILL:1-2 m below ground:24 hour fcst: -19: 4:78201:d=2021032312:CNWAT:surface:24 hour fcst: -19: 5:93974:d=2021032312:ICETK:surface:24 hour fcst: -19: 6:101822:d=2021032312:DUVB:surface:18-24 hour ave fcst: -19: 7:157727:d=2021032312:CDUVB:surface:18-24 hour ave fcst: -19: 8:198178:d=2021032312:TMP:305 m above mean sea level:24 hour fcst: -19: 9:248933:d=2021032312:UGRD:305 m above mean sea level:24 hour fcst: -19: 10:303477:d=2021032312:VGRD:305 m above mean sea level:24 hour fcst: -19: 11:359693:d=2021032312:TMP:457 m above mean sea level:24 hour fcst: -19: 12:413648:d=2021032312:UGRD:457 m above mean sea level:24 hour fcst: -19: 13:470896:d=2021032312:VGRD:457 m above mean sea level:24 hour fcst: -19: 14:529719:d=2021032312:TMP:610 m above mean sea level:24 hour fcst: -19: + product_functions.sh[10]rc=0 -19: + product_functions.sh[11](( rc == 0 )) -19: + product_functions.sh[11]mv pgb2bfile_f024_20_1p00.new pgb2bfile_f024_20_1p00 -19: + product_functions.sh[12]return 0 -19: + interp_atmos_master.sh[56]export err=0 -19: + interp_atmos_master.sh[56]err=0 -19: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -19: ++ interp_atmos_master.sh[62]wc -l -19: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_20_1p00 -match 'LAND|ICEC' -19: + interp_atmos_master.sh[62]var_count=0 -19: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -19: + interp_atmos_master.sh[73]exit 0 - 8: + bash[8]'[' -z '' ']' - 8: + bash[9]case "$-" in - 8: + bash[12]__lmod_vx=x - 8: + bash[16]'[' -n x ']' - 8: + bash[16]set +x - 8: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 8: Shell debugging restarted - 8: + bash[224]unset __lmod_vx - 8: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_9 - 8: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_9 - 8: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 - 8: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 8: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 8: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 8: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 8: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 8: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 8: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 8: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 8: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 8: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 8: + interp_atmos_master.sh[31]IFS=: - 8: + interp_atmos_master.sh[31]read -ra grids - 8: + interp_atmos_master.sh[33]output_grids= - 8: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 8: + interp_atmos_master.sh[35]gridopt=grid0p25 - 8: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_9_0p25' - 8: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 8: + interp_atmos_master.sh[35]gridopt=grid0p50 - 8: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_9_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_9_0p50' - 8: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 8: + interp_atmos_master.sh[35]gridopt=grid1p00 - 8: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_9_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_9_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_9_1p00' - 8: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_9 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_9_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_9_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_9_1p00 - 8: 1:0:d=2021032312:TMP:375 mb:24 hour fcst: - 8: 2:18300:d=2021032312:RH:375 mb:24 hour fcst: - 8: 3:37464:d=2021032312:TCDC:375 mb:24 hour fcst: - 8: 4:50733:d=2021032312:VVEL:375 mb:24 hour fcst: - 8: 5:75560:d=2021032312:DZDT:375 mb:24 hour fcst: - 8: 6:103703:d=2021032312:UGRD:375 mb:24 hour fcst: - 8: 7:118150:d=2021032312:VGRD:375 mb:24 hour fcst: - 8: 8:141088:d=2021032312:ABSV:375 mb:24 hour fcst: - 8: 9:163304:d=2021032312:CLMR:375 mb:24 hour fcst: - 8: 10:163483:d=2021032312:ICMR:375 mb:24 hour fcst: - 8: 11:180411:d=2021032312:RWMR:375 mb:24 hour fcst: - 8: 12:180590:d=2021032312:SNMR:375 mb:24 hour fcst: - 8: 13:193623:d=2021032312:GRLE:375 mb:24 hour fcst: - 8: 14:194235:d=2021032312:HGT:425 mb:24 hour fcst: - 8: + interp_atmos_master.sh[47]export err=0 - 8: + interp_atmos_master.sh[47]err=0 - 8: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 8: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 8: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_9_0p25 - 8: + product_functions.sh[5]local filename=pgb2bfile_f024_9_0p25 - 8: + product_functions.sh[6]wgrib2 pgb2bfile_f024_9_0p25 -not_if :RH: -grib pgb2bfile_f024_9_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_9_0p25.new - 8: 1:0:d=2021032312:TMP:375 mb:24 hour fcst: - 8: 2:409306:d=2021032312:RH:375 mb:24 hour fcst: - 8: 3:842744:d=2021032312:TCDC:375 mb:24 hour fcst: - 8: 4:1082283:d=2021032312:VVEL:375 mb:24 hour fcst: - 8: 5:1741693:d=2021032312:DZDT:375 mb:24 hour fcst: - 8: 6:2536903:d=2021032312:UGRD:375 mb:24 hour fcst: - 8: 7:3090576:d=2021032312:VGRD:375 mb:24 hour fcst: - 8: 8:3671580:d=2021032312:ABSV:375 mb:24 hour fcst: - 8: 9:4215660:d=2021032312:CLMR:375 mb:24 hour fcst: - 8: 10:4215839:d=2021032312:ICMR:375 mb:24 hour fcst: - 8: 11:4590449:d=2021032312:RWMR:375 mb:24 hour fcst: - 8: 12:4590628:d=2021032312:SNMR:375 mb:24 hour fcst: - 8: 13:4856387:d=2021032312:GRLE:375 mb:24 hour fcst: - 8: 14:4864532:d=2021032312:HGT:425 mb:24 hour fcst: - 8: + product_functions.sh[10]rc=0 - 8: + product_functions.sh[11](( rc == 0 )) - 8: + product_functions.sh[11]mv pgb2bfile_f024_9_0p25.new pgb2bfile_f024_9_0p25 - 8: + product_functions.sh[12]return 0 - 8: + interp_atmos_master.sh[56]export err=0 - 8: + interp_atmos_master.sh[56]err=0 - 8: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 8: ++ interp_atmos_master.sh[62]wc -l - 8: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_9_0p25 -match 'LAND|ICEC' - 8: + interp_atmos_master.sh[62]var_count=0 - 8: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 8: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 8: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_9_0p50 - 8: + product_functions.sh[5]local filename=pgb2bfile_f024_9_0p50 - 8: + product_functions.sh[6]wgrib2 pgb2bfile_f024_9_0p50 -not_if :RH: -grib pgb2bfile_f024_9_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_9_0p50.new - 8: 1:0:d=2021032312:TMP:375 mb:24 hour fcst: - 8: 2:142478:d=2021032312:RH:375 mb:24 hour fcst: - 8: 3:294183:d=2021032312:TCDC:375 mb:24 hour fcst: - 8: 4:387797:d=2021032312:VVEL:375 mb:24 hour fcst: - 8: 5:612223:d=2021032312:DZDT:375 mb:24 hour fcst: - 8: 6:871446:d=2021032312:UGRD:375 mb:24 hour fcst: - 8: 7:1062553:d=2021032312:VGRD:375 mb:24 hour fcst: - 8: 8:1260658:d=2021032312:ABSV:375 mb:24 hour fcst: - 8: 9:1447417:d=2021032312:CLMR:375 mb:24 hour fcst: - 8: 10:1447596:d=2021032312:ICMR:375 mb:24 hour fcst: - 8: 11:1584069:d=2021032312:RWMR:375 mb:24 hour fcst: - 8: 12:1584248:d=2021032312:SNMR:375 mb:24 hour fcst: - 8: 13:1680523:d=2021032312:GRLE:375 mb:24 hour fcst: - 8: 14:1683662:d=2021032312:HGT:425 mb:24 hour fcst: - 8: + product_functions.sh[10]rc=0 - 8: + product_functions.sh[11](( rc == 0 )) - 8: + product_functions.sh[11]mv pgb2bfile_f024_9_0p50.new pgb2bfile_f024_9_0p50 - 8: + product_functions.sh[12]return 0 - 8: + interp_atmos_master.sh[56]export err=0 - 8: + interp_atmos_master.sh[56]err=0 - 8: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 8: ++ interp_atmos_master.sh[62]wc -l - 8: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_9_0p50 -match 'LAND|ICEC' - 8: + interp_atmos_master.sh[62]var_count=0 - 8: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 8: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 8: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_9_1p00 - 8: + product_functions.sh[5]local filename=pgb2bfile_f024_9_1p00 - 8: + product_functions.sh[6]wgrib2 pgb2bfile_f024_9_1p00 -not_if :RH: -grib pgb2bfile_f024_9_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_9_1p00.new - 8: 1:0:d=2021032312:TMP:375 mb:24 hour fcst: - 8: 2:46883:d=2021032312:RH:375 mb:24 hour fcst: - 8: 3:96677:d=2021032312:TCDC:375 mb:24 hour fcst: - 8: 4:130441:d=2021032312:VVEL:375 mb:24 hour fcst: - 8: 5:198908:d=2021032312:DZDT:375 mb:24 hour fcst: - 8: 6:276532:d=2021032312:UGRD:375 mb:24 hour fcst: - 8: 7:337125:d=2021032312:VGRD:375 mb:24 hour fcst: - 8: 8:399983:d=2021032312:ABSV:375 mb:24 hour fcst: - 8: 9:459344:d=2021032312:CLMR:375 mb:24 hour fcst: - 8: 10:459523:d=2021032312:ICMR:375 mb:24 hour fcst: - 8: 11:504679:d=2021032312:RWMR:375 mb:24 hour fcst: - 8: 12:504858:d=2021032312:SNMR:375 mb:24 hour fcst: - 8: 13:537965:d=2021032312:GRLE:375 mb:24 hour fcst: - 8: 14:539387:d=2021032312:HGT:425 mb:24 hour fcst: - 8: + product_functions.sh[10]rc=0 - 8: + product_functions.sh[11](( rc == 0 )) - 8: + product_functions.sh[11]mv pgb2bfile_f024_9_1p00.new pgb2bfile_f024_9_1p00 - 8: + product_functions.sh[12]return 0 - 8: + interp_atmos_master.sh[56]export err=0 - 8: + interp_atmos_master.sh[56]err=0 - 8: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 8: ++ interp_atmos_master.sh[62]wc -l - 8: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_9_1p00 -match 'LAND|ICEC' - 8: + interp_atmos_master.sh[62]var_count=0 - 8: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 8: + interp_atmos_master.sh[73]exit 0 -16: + bash[8]'[' -z '' ']' -16: + bash[9]case "$-" in -16: + bash[12]__lmod_vx=x -16: + bash[16]'[' -n x ']' -16: + bash[16]set +x -16: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -16: Shell debugging restarted -16: + bash[224]unset __lmod_vx -16: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_17 -16: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_17 -16: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -16: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -16: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -16: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -16: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -16: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -16: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -16: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -16: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -16: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -16: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -16: + interp_atmos_master.sh[31]IFS=: -16: + interp_atmos_master.sh[31]read -ra grids -16: + interp_atmos_master.sh[33]output_grids= -16: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -16: + interp_atmos_master.sh[35]gridopt=grid0p25 -16: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_17_0p25' -16: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -16: + interp_atmos_master.sh[35]gridopt=grid0p50 -16: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_17_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_17_0p50' -16: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -16: + interp_atmos_master.sh[35]gridopt=grid1p00 -16: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_17_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_17_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_17_1p00' -16: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_17 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_17_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_17_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_17_1p00 -16: 1:0:d=2021032312:TMP:775 mb:24 hour fcst: -16: 2:20093:d=2021032312:RH:775 mb:24 hour fcst: -16: 3:40626:d=2021032312:TCDC:775 mb:24 hour fcst: -16: 4:53554:d=2021032312:VVEL:775 mb:24 hour fcst: -16: 5:79244:d=2021032312:DZDT:775 mb:24 hour fcst: -16: 6:106186:d=2021032312:UGRD:775 mb:24 hour fcst: -16: 7:127046:d=2021032312:VGRD:775 mb:24 hour fcst: -16: 8:148731:d=2021032312:ABSV:775 mb:24 hour fcst: -16: 9:169690:d=2021032312:CLMR:775 mb:24 hour fcst: -16: 10:180200:d=2021032312:ICMR:775 mb:24 hour fcst: -16: 11:191679:d=2021032312:RWMR:775 mb:24 hour fcst: -16: 12:198587:d=2021032312:SNMR:775 mb:24 hour fcst: -16: 13:211731:d=2021032312:GRLE:775 mb:24 hour fcst: -16: 14:214063:d=2021032312:HGT:825 mb:24 hour fcst: -16: + interp_atmos_master.sh[47]export err=0 -16: + interp_atmos_master.sh[47]err=0 -16: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -16: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -16: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_17_0p25 -16: + product_functions.sh[5]local filename=pgb2bfile_f024_17_0p25 -16: + product_functions.sh[6]wgrib2 pgb2bfile_f024_17_0p25 -not_if :RH: -grib pgb2bfile_f024_17_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_17_0p25.new -16: 1:0:d=2021032312:TMP:775 mb:24 hour fcst: -16: 2:469002:d=2021032312:RH:775 mb:24 hour fcst: -16: 3:951048:d=2021032312:TCDC:775 mb:24 hour fcst: -16: 4:1208066:d=2021032312:VVEL:775 mb:24 hour fcst: -16: 5:1893847:d=2021032312:DZDT:775 mb:24 hour fcst: -16: 6:2640240:d=2021032312:UGRD:775 mb:24 hour fcst: -16: 7:3145560:d=2021032312:VGRD:775 mb:24 hour fcst: -16: 8:3681683:d=2021032312:ABSV:775 mb:24 hour fcst: -16: 9:4185854:d=2021032312:CLMR:775 mb:24 hour fcst: -16: 10:4411057:d=2021032312:ICMR:775 mb:24 hour fcst: -16: 11:4680962:d=2021032312:RWMR:775 mb:24 hour fcst: -16: 12:4807667:d=2021032312:SNMR:775 mb:24 hour fcst: -16: 13:5103182:d=2021032312:GRLE:775 mb:24 hour fcst: -16: 14:5145306:d=2021032312:HGT:825 mb:24 hour fcst: -16: + product_functions.sh[10]rc=0 -16: + product_functions.sh[11](( rc == 0 )) -16: + product_functions.sh[11]mv pgb2bfile_f024_17_0p25.new pgb2bfile_f024_17_0p25 -16: + product_functions.sh[12]return 0 -16: + interp_atmos_master.sh[56]export err=0 -16: + interp_atmos_master.sh[56]err=0 -16: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -16: ++ interp_atmos_master.sh[62]wc -l -16: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_17_0p25 -match 'LAND|ICEC' -16: + interp_atmos_master.sh[62]var_count=0 -16: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -16: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -16: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_17_0p50 -16: + product_functions.sh[5]local filename=pgb2bfile_f024_17_0p50 -16: + product_functions.sh[6]wgrib2 pgb2bfile_f024_17_0p50 -not_if :RH: -grib pgb2bfile_f024_17_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_17_0p50.new -16: 1:0:d=2021032312:TMP:775 mb:24 hour fcst: -16: 2:162464:d=2021032312:RH:775 mb:24 hour fcst: -16: 3:328758:d=2021032312:TCDC:775 mb:24 hour fcst: -16: 4:424947:d=2021032312:VVEL:775 mb:24 hour fcst: -16: 5:658633:d=2021032312:DZDT:775 mb:24 hour fcst: -16: 6:903986:d=2021032312:UGRD:775 mb:24 hour fcst: -16: 7:1077795:d=2021032312:VGRD:775 mb:24 hour fcst: -16: 8:1261396:d=2021032312:ABSV:775 mb:24 hour fcst: -16: 9:1435527:d=2021032312:CLMR:775 mb:24 hour fcst: -16: 10:1519789:d=2021032312:ICMR:775 mb:24 hour fcst: -16: 11:1614242:d=2021032312:RWMR:775 mb:24 hour fcst: -16: 12:1661331:d=2021032312:SNMR:775 mb:24 hour fcst: -16: 13:1766395:d=2021032312:GRLE:775 mb:24 hour fcst: -16: 14:1782288:d=2021032312:HGT:825 mb:24 hour fcst: -16: + product_functions.sh[10]rc=0 -16: + product_functions.sh[11](( rc == 0 )) -16: + product_functions.sh[11]mv pgb2bfile_f024_17_0p50.new pgb2bfile_f024_17_0p50 -16: + product_functions.sh[12]return 0 -16: + interp_atmos_master.sh[56]export err=0 -16: + interp_atmos_master.sh[56]err=0 -16: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -16: ++ interp_atmos_master.sh[62]wc -l -16: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_17_0p50 -match 'LAND|ICEC' -16: + interp_atmos_master.sh[62]var_count=0 -16: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -16: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -16: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_17_1p00 -16: + product_functions.sh[5]local filename=pgb2bfile_f024_17_1p00 -16: + product_functions.sh[6]wgrib2 pgb2bfile_f024_17_1p00 -not_if :RH: -grib pgb2bfile_f024_17_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_17_1p00.new -16: 1:0:d=2021032312:TMP:775 mb:24 hour fcst: -16: 2:52851:d=2021032312:RH:775 mb:24 hour fcst: -16: 3:106990:d=2021032312:TCDC:775 mb:24 hour fcst: -16: 4:140646:d=2021032312:VVEL:775 mb:24 hour fcst: -16: 5:212239:d=2021032312:DZDT:775 mb:24 hour fcst: -16: 6:286751:d=2021032312:UGRD:775 mb:24 hour fcst: -16: 7:342637:d=2021032312:VGRD:775 mb:24 hour fcst: -16: 8:401057:d=2021032312:ABSV:775 mb:24 hour fcst: -16: 9:456588:d=2021032312:CLMR:775 mb:24 hour fcst: -16: 10:485022:d=2021032312:ICMR:775 mb:24 hour fcst: -16: 11:515614:d=2021032312:RWMR:775 mb:24 hour fcst: -16: 12:532382:d=2021032312:SNMR:775 mb:24 hour fcst: -16: 13:567590:d=2021032312:GRLE:775 mb:24 hour fcst: -16: 14:573596:d=2021032312:HGT:825 mb:24 hour fcst: -16: + product_functions.sh[10]rc=0 -16: + product_functions.sh[11](( rc == 0 )) -16: + product_functions.sh[11]mv pgb2bfile_f024_17_1p00.new pgb2bfile_f024_17_1p00 -16: + product_functions.sh[12]return 0 -16: + interp_atmos_master.sh[56]export err=0 -16: + interp_atmos_master.sh[56]err=0 -16: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -16: ++ interp_atmos_master.sh[62]wc -l -16: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_17_1p00 -match 'LAND|ICEC' -16: + interp_atmos_master.sh[62]var_count=0 -16: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -16: + interp_atmos_master.sh[73]exit 0 -18: + bash[8]'[' -z '' ']' -18: + bash[9]case "$-" in -18: + bash[12]__lmod_vx=x -18: + bash[16]'[' -n x ']' -18: + bash[16]set +x -18: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -18: Shell debugging restarted -18: + bash[224]unset __lmod_vx -18: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_19 -18: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_19 -18: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -18: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -18: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -18: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -18: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -18: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -18: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -18: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -18: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -18: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -18: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -18: + interp_atmos_master.sh[31]IFS=: -18: + interp_atmos_master.sh[31]read -ra grids -18: + interp_atmos_master.sh[33]output_grids= -18: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -18: + interp_atmos_master.sh[35]gridopt=grid0p25 -18: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_19_0p25' -18: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -18: + interp_atmos_master.sh[35]gridopt=grid0p50 -18: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_19_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_19_0p50' -18: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -18: + interp_atmos_master.sh[35]gridopt=grid1p00 -18: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_19_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_19_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_19_1p00' -18: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_19 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_19_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_19_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_19_1p00 -18: 1:0:d=2021032312:TMP:875 mb:24 hour fcst: -18: 2:20765:d=2021032312:RH:875 mb:24 hour fcst: -18: 3:40766:d=2021032312:TCDC:875 mb:24 hour fcst: -18: 4:54849:d=2021032312:VVEL:875 mb:24 hour fcst: -18: 5:79970:d=2021032312:DZDT:875 mb:24 hour fcst: -18: 6:106363:d=2021032312:UGRD:875 mb:24 hour fcst: -18: 7:127591:d=2021032312:VGRD:875 mb:24 hour fcst: -18: 8:149391:d=2021032312:ABSV:875 mb:24 hour fcst: -18: 9:170600:d=2021032312:CLMR:875 mb:24 hour fcst: -18: 10:183203:d=2021032312:ICMR:875 mb:24 hour fcst: -18: 11:193102:d=2021032312:RWMR:875 mb:24 hour fcst: -18: 12:204862:d=2021032312:SNMR:875 mb:24 hour fcst: -18: 13:215886:d=2021032312:GRLE:875 mb:24 hour fcst: -18: 14:218519:d=2021032312:SOILL:0-0.1 m below ground:24 hour fcst: -18: + interp_atmos_master.sh[47]export err=0 -18: + interp_atmos_master.sh[47]err=0 -18: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -18: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -18: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_19_0p25 -18: + product_functions.sh[5]local filename=pgb2bfile_f024_19_0p25 -18: + product_functions.sh[6]wgrib2 pgb2bfile_f024_19_0p25 -not_if :RH: -grib pgb2bfile_f024_19_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_19_0p25.new -18: 1:0:d=2021032312:TMP:875 mb:24 hour fcst: -18: 2:492421:d=2021032312:RH:875 mb:24 hour fcst: -18: 3:960157:d=2021032312:TCDC:875 mb:24 hour fcst: -18: 4:1247255:d=2021032312:VVEL:875 mb:24 hour fcst: -18: 5:1920780:d=2021032312:DZDT:875 mb:24 hour fcst: -18: 6:2644458:d=2021032312:UGRD:875 mb:24 hour fcst: -18: 7:3156862:d=2021032312:VGRD:875 mb:24 hour fcst: -18: 8:3700344:d=2021032312:ABSV:875 mb:24 hour fcst: -18: 9:4205504:d=2021032312:CLMR:875 mb:24 hour fcst: -18: 10:4480888:d=2021032312:ICMR:875 mb:24 hour fcst: -18: 11:4720793:d=2021032312:RWMR:875 mb:24 hour fcst: -18: 12:4948601:d=2021032312:SNMR:875 mb:24 hour fcst: -18: 13:5199198:d=2021032312:GRLE:875 mb:24 hour fcst: -18: 14:5247841:d=2021032312:SOILL:0-0.1 m below ground:24 hour fcst: -18: + product_functions.sh[10]rc=0 -18: + product_functions.sh[11](( rc == 0 )) -18: + product_functions.sh[11]mv pgb2bfile_f024_19_0p25.new pgb2bfile_f024_19_0p25 -18: + product_functions.sh[12]return 0 -18: + interp_atmos_master.sh[56]export err=0 -18: + interp_atmos_master.sh[56]err=0 -18: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -18: ++ interp_atmos_master.sh[62]wc -l -18: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_19_0p25 -match 'LAND|ICEC' -18: + interp_atmos_master.sh[62]var_count=0 -18: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -18: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -18: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_19_0p50 -18: + product_functions.sh[5]local filename=pgb2bfile_f024_19_0p50 -18: + product_functions.sh[6]wgrib2 pgb2bfile_f024_19_0p50 -not_if :RH: -grib pgb2bfile_f024_19_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_19_0p50.new -18: 1:0:d=2021032312:TMP:875 mb:24 hour fcst: -18: 2:170762:d=2021032312:RH:875 mb:24 hour fcst: -18: 3:332365:d=2021032312:TCDC:875 mb:24 hour fcst: -18: 4:439562:d=2021032312:VVEL:875 mb:24 hour fcst: -18: 5:668450:d=2021032312:DZDT:875 mb:24 hour fcst: -18: 6:906247:d=2021032312:UGRD:875 mb:24 hour fcst: -18: 7:1082554:d=2021032312:VGRD:875 mb:24 hour fcst: -18: 8:1268200:d=2021032312:ABSV:875 mb:24 hour fcst: -18: 9:1442606:d=2021032312:CLMR:875 mb:24 hour fcst: -18: 10:1545792:d=2021032312:ICMR:875 mb:24 hour fcst: -18: 11:1630514:d=2021032312:RWMR:875 mb:24 hour fcst: -18: 12:1715030:d=2021032312:SNMR:875 mb:24 hour fcst: -18: 13:1803184:d=2021032312:GRLE:875 mb:24 hour fcst: -18: 14:1821010:d=2021032312:SOILL:0-0.1 m below ground:24 hour fcst: -18: + product_functions.sh[10]rc=0 -18: + product_functions.sh[11](( rc == 0 )) -18: + product_functions.sh[11]mv pgb2bfile_f024_19_0p50.new pgb2bfile_f024_19_0p50 -18: + product_functions.sh[12]return 0 -18: + interp_atmos_master.sh[56]export err=0 -18: + interp_atmos_master.sh[56]err=0 -18: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -18: ++ interp_atmos_master.sh[62]wc -l -18: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_19_0p50 -match 'LAND|ICEC' -18: + interp_atmos_master.sh[62]var_count=0 -18: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -18: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -18: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_19_1p00 -18: + product_functions.sh[5]local filename=pgb2bfile_f024_19_1p00 -18: + product_functions.sh[6]wgrib2 pgb2bfile_f024_19_1p00 -not_if :RH: -grib pgb2bfile_f024_19_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_19_1p00.new -18: 1:0:d=2021032312:TMP:875 mb:24 hour fcst: -18: 2:54942:d=2021032312:RH:875 mb:24 hour fcst: -18: 3:107486:d=2021032312:TCDC:875 mb:24 hour fcst: -18: 4:144680:d=2021032312:VVEL:875 mb:24 hour fcst: -18: 5:214621:d=2021032312:DZDT:875 mb:24 hour fcst: -18: 6:286568:d=2021032312:UGRD:875 mb:24 hour fcst: -18: 7:343071:d=2021032312:VGRD:875 mb:24 hour fcst: -18: 8:401764:d=2021032312:ABSV:875 mb:24 hour fcst: -18: 9:457784:d=2021032312:CLMR:875 mb:24 hour fcst: -18: 10:492106:d=2021032312:ICMR:875 mb:24 hour fcst: -18: 11:519544:d=2021032312:RWMR:875 mb:24 hour fcst: -18: 12:549341:d=2021032312:SNMR:875 mb:24 hour fcst: -18: 13:578314:d=2021032312:GRLE:875 mb:24 hour fcst: -18: 14:584828:d=2021032312:SOILL:0-0.1 m below ground:24 hour fcst: -18: + product_functions.sh[10]rc=0 -18: + product_functions.sh[11](( rc == 0 )) -18: + product_functions.sh[11]mv pgb2bfile_f024_19_1p00.new pgb2bfile_f024_19_1p00 -18: + product_functions.sh[12]return 0 -18: + interp_atmos_master.sh[56]export err=0 -18: + interp_atmos_master.sh[56]err=0 -18: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -18: ++ interp_atmos_master.sh[62]wc -l -18: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_19_1p00 -match 'LAND|ICEC' -18: + interp_atmos_master.sh[62]var_count=0 -18: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -18: + interp_atmos_master.sh[73]exit 0 -13: + bash[8]'[' -z '' ']' -13: + bash[9]case "$-" in -13: + bash[12]__lmod_vx=x -13: + bash[16]'[' -n x ']' -13: + bash[16]set +x -13: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -13: Shell debugging restarted -13: + bash[224]unset __lmod_vx -13: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_14 -13: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_14 -13: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -13: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -13: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -13: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -13: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -13: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -13: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -13: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -13: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -13: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -13: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -13: + interp_atmos_master.sh[31]IFS=: -13: + interp_atmos_master.sh[31]read -ra grids -13: + interp_atmos_master.sh[33]output_grids= -13: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -13: + interp_atmos_master.sh[35]gridopt=grid0p25 -13: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_14_0p25' -13: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -13: + interp_atmos_master.sh[35]gridopt=grid0p50 -13: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_14_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_14_0p50' -13: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -13: + interp_atmos_master.sh[35]gridopt=grid1p00 -13: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_14_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_14_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_14_1p00' -13: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_14 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_14_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_14_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_14_1p00 -13: 1:0:d=2021032312:TMP:625 mb:24 hour fcst: -13: 2:19130:d=2021032312:RH:625 mb:24 hour fcst: -13: 3:39305:d=2021032312:TCDC:625 mb:24 hour fcst: -13: 4:51145:d=2021032312:VVEL:625 mb:24 hour fcst: -13: 5:76631:d=2021032312:DZDT:625 mb:24 hour fcst: -13: 6:104047:d=2021032312:UGRD:625 mb:24 hour fcst: -13: 7:125076:d=2021032312:VGRD:625 mb:24 hour fcst: -13: 8:146524:d=2021032312:ABSV:625 mb:24 hour fcst: -13: 9:167334:d=2021032312:CLMR:625 mb:24 hour fcst: -13: 10:172392:d=2021032312:ICMR:625 mb:24 hour fcst: -13: 11:185593:d=2021032312:RWMR:625 mb:24 hour fcst: -13: 12:189847:d=2021032312:SNMR:625 mb:24 hour fcst: -13: 13:205323:d=2021032312:GRLE:625 mb:24 hour fcst: -13: 14:207822:d=2021032312:HGT:675 mb:24 hour fcst: -13: + interp_atmos_master.sh[47]export err=0 -13: + interp_atmos_master.sh[47]err=0 -13: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -13: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -13: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_14_0p25 -13: + product_functions.sh[5]local filename=pgb2bfile_f024_14_0p25 -13: + product_functions.sh[6]wgrib2 pgb2bfile_f024_14_0p25 -not_if :RH: -grib pgb2bfile_f024_14_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_14_0p25.new -13: 1:0:d=2021032312:TMP:625 mb:24 hour fcst: -13: 2:427901:d=2021032312:RH:625 mb:24 hour fcst: -13: 3:897547:d=2021032312:TCDC:625 mb:24 hour fcst: -13: 4:1117278:d=2021032312:VVEL:625 mb:24 hour fcst: -13: 5:1793706:d=2021032312:DZDT:625 mb:24 hour fcst: -13: 6:2551031:d=2021032312:UGRD:625 mb:24 hour fcst: -13: 7:3057590:d=2021032312:VGRD:625 mb:24 hour fcst: -13: 8:3593524:d=2021032312:ABSV:625 mb:24 hour fcst: -13: 9:4084565:d=2021032312:CLMR:625 mb:24 hour fcst: -13: 10:4186130:d=2021032312:ICMR:625 mb:24 hour fcst: -13: 11:4478195:d=2021032312:RWMR:625 mb:24 hour fcst: -13: 12:4567440:d=2021032312:SNMR:625 mb:24 hour fcst: -13: 13:4904210:d=2021032312:GRLE:625 mb:24 hour fcst: -13: 14:4949448:d=2021032312:HGT:675 mb:24 hour fcst: -13: + product_functions.sh[10]rc=0 -13: + product_functions.sh[11](( rc == 0 )) -13: + product_functions.sh[11]mv pgb2bfile_f024_14_0p25.new pgb2bfile_f024_14_0p25 -13: + product_functions.sh[12]return 0 -13: + interp_atmos_master.sh[56]export err=0 -13: + interp_atmos_master.sh[56]err=0 -13: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -13: ++ interp_atmos_master.sh[62]wc -l -13: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_14_0p25 -match 'LAND|ICEC' -13: + interp_atmos_master.sh[62]var_count=0 -13: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -13: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -13: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_14_0p50 -13: + product_functions.sh[5]local filename=pgb2bfile_f024_14_0p50 -13: + product_functions.sh[6]wgrib2 pgb2bfile_f024_14_0p50 -not_if :RH: -grib pgb2bfile_f024_14_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_14_0p50.new -13: 1:0:d=2021032312:TMP:625 mb:24 hour fcst: -13: 2:149937:d=2021032312:RH:625 mb:24 hour fcst: -13: 3:312268:d=2021032312:TCDC:625 mb:24 hour fcst: -13: 4:395975:d=2021032312:VVEL:625 mb:24 hour fcst: -13: 5:626826:d=2021032312:DZDT:625 mb:24 hour fcst: -13: 6:877198:d=2021032312:UGRD:625 mb:24 hour fcst: -13: 7:1051336:d=2021032312:VGRD:625 mb:24 hour fcst: -13: 8:1234047:d=2021032312:ABSV:625 mb:24 hour fcst: -13: 9:1404422:d=2021032312:CLMR:625 mb:24 hour fcst: -13: 10:1442696:d=2021032312:ICMR:625 mb:24 hour fcst: -13: 11:1546605:d=2021032312:RWMR:625 mb:24 hour fcst: -13: 12:1580255:d=2021032312:SNMR:625 mb:24 hour fcst: -13: 13:1700651:d=2021032312:GRLE:625 mb:24 hour fcst: -13: 14:1717865:d=2021032312:HGT:675 mb:24 hour fcst: -13: + product_functions.sh[10]rc=0 -13: + product_functions.sh[11](( rc == 0 )) -13: + product_functions.sh[11]mv pgb2bfile_f024_14_0p50.new pgb2bfile_f024_14_0p50 -13: + product_functions.sh[12]return 0 -13: + interp_atmos_master.sh[56]export err=0 -13: + interp_atmos_master.sh[56]err=0 -13: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -13: ++ interp_atmos_master.sh[62]wc -l -13: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_14_0p50 -match 'LAND|ICEC' -13: + interp_atmos_master.sh[62]var_count=0 -13: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -13: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -13: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_14_1p00 -13: + product_functions.sh[5]local filename=pgb2bfile_f024_14_1p00 -13: + product_functions.sh[6]wgrib2 pgb2bfile_f024_14_1p00 -not_if :RH: -grib pgb2bfile_f024_14_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_14_1p00.new -13: 1:0:d=2021032312:TMP:625 mb:24 hour fcst: -13: 2:49429:d=2021032312:RH:625 mb:24 hour fcst: -13: 3:102417:d=2021032312:TCDC:625 mb:24 hour fcst: -13: 4:132784:d=2021032312:VVEL:625 mb:24 hour fcst: -13: 5:203612:d=2021032312:DZDT:625 mb:24 hour fcst: -13: 6:279172:d=2021032312:UGRD:625 mb:24 hour fcst: -13: 7:335070:d=2021032312:VGRD:625 mb:24 hour fcst: -13: 8:393235:d=2021032312:ABSV:625 mb:24 hour fcst: -13: 9:448098:d=2021032312:CLMR:625 mb:24 hour fcst: -13: 10:461729:d=2021032312:ICMR:625 mb:24 hour fcst: -13: 11:496147:d=2021032312:RWMR:625 mb:24 hour fcst: -13: 12:507531:d=2021032312:SNMR:625 mb:24 hour fcst: -13: 13:547910:d=2021032312:GRLE:625 mb:24 hour fcst: -13: 14:554371:d=2021032312:HGT:675 mb:24 hour fcst: -13: + product_functions.sh[10]rc=0 -13: + product_functions.sh[11](( rc == 0 )) -13: + product_functions.sh[11]mv pgb2bfile_f024_14_1p00.new pgb2bfile_f024_14_1p00 -13: + product_functions.sh[12]return 0 -13: + interp_atmos_master.sh[56]export err=0 -13: + interp_atmos_master.sh[56]err=0 -13: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -13: ++ interp_atmos_master.sh[62]wc -l -13: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_14_1p00 -match 'LAND|ICEC' -13: + interp_atmos_master.sh[62]var_count=0 -13: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -13: + interp_atmos_master.sh[73]exit 0 -10: + bash[8]'[' -z '' ']' -10: + bash[9]case "$-" in -10: + bash[12]__lmod_vx=x -10: + bash[16]'[' -n x ']' -10: + bash[16]set +x -10: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -10: Shell debugging restarted -10: + bash[224]unset __lmod_vx -10: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_11 -10: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_11 -10: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -10: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -10: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -10: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -10: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -10: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -10: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -10: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -10: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -10: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -10: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -10: + interp_atmos_master.sh[31]IFS=: -10: + interp_atmos_master.sh[31]read -ra grids -10: + interp_atmos_master.sh[33]output_grids= -10: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -10: + interp_atmos_master.sh[35]gridopt=grid0p25 -10: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_11_0p25' -10: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -10: + interp_atmos_master.sh[35]gridopt=grid0p50 -10: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_11_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_11_0p50' -10: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -10: + interp_atmos_master.sh[35]gridopt=grid1p00 -10: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_11_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_11_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_11_1p00' -10: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_11 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_11_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_11_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_11_1p00 -10: 1:0:d=2021032312:TMP:475 mb:24 hour fcst: -10: 2:18474:d=2021032312:RH:475 mb:24 hour fcst: -10: 3:37972:d=2021032312:TCDC:475 mb:24 hour fcst: -10: 4:50346:d=2021032312:VVEL:475 mb:24 hour fcst: -10: 5:75325:d=2021032312:DZDT:475 mb:24 hour fcst: -10: 6:103330:d=2021032312:UGRD:475 mb:24 hour fcst: -10: 7:124932:d=2021032312:VGRD:475 mb:24 hour fcst: -10: 8:147088:d=2021032312:ABSV:475 mb:24 hour fcst: -10: 9:168398:d=2021032312:CLMR:475 mb:24 hour fcst: -10: 10:169772:d=2021032312:ICMR:475 mb:24 hour fcst: -10: 11:185956:d=2021032312:RWMR:475 mb:24 hour fcst: -10: 12:186653:d=2021032312:SNMR:475 mb:24 hour fcst: -10: 13:200608:d=2021032312:GRLE:475 mb:24 hour fcst: -10: 14:202179:d=2021032312:HGT:525 mb:24 hour fcst: -10: + interp_atmos_master.sh[47]export err=0 -10: + interp_atmos_master.sh[47]err=0 -10: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -10: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -10: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_11_0p25 -10: + product_functions.sh[5]local filename=pgb2bfile_f024_11_0p25 -10: + product_functions.sh[6]wgrib2 pgb2bfile_f024_11_0p25 -not_if :RH: -grib pgb2bfile_f024_11_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_11_0p25.new -10: 1:0:d=2021032312:TMP:475 mb:24 hour fcst: -10: 2:422175:d=2021032312:RH:475 mb:24 hour fcst: -10: 3:866414:d=2021032312:TCDC:475 mb:24 hour fcst: -10: 4:1099847:d=2021032312:VVEL:475 mb:24 hour fcst: -10: 5:1768336:d=2021032312:DZDT:475 mb:24 hour fcst: -10: 6:2551159:d=2021032312:UGRD:475 mb:24 hour fcst: -10: 7:3079980:d=2021032312:VGRD:475 mb:24 hour fcst: -10: 8:3638259:d=2021032312:ABSV:475 mb:24 hour fcst: -10: 9:4156817:d=2021032312:CLMR:475 mb:24 hour fcst: -10: 10:4183598:d=2021032312:ICMR:475 mb:24 hour fcst: -10: 11:4545238:d=2021032312:RWMR:475 mb:24 hour fcst: -10: 12:4554842:d=2021032312:SNMR:475 mb:24 hour fcst: -10: 13:4848361:d=2021032312:GRLE:475 mb:24 hour fcst: -10: 14:4870501:d=2021032312:HGT:525 mb:24 hour fcst: -10: + product_functions.sh[10]rc=0 -10: + product_functions.sh[11](( rc == 0 )) -10: + product_functions.sh[11]mv pgb2bfile_f024_11_0p25.new pgb2bfile_f024_11_0p25 -10: + product_functions.sh[12]return 0 -10: + interp_atmos_master.sh[56]export err=0 -10: + interp_atmos_master.sh[56]err=0 -10: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -10: ++ interp_atmos_master.sh[62]wc -l -10: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_11_0p25 -match 'LAND|ICEC' -10: + interp_atmos_master.sh[62]var_count=0 -10: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -10: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -10: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_11_0p50 -10: + product_functions.sh[5]local filename=pgb2bfile_f024_11_0p50 -10: + product_functions.sh[6]wgrib2 pgb2bfile_f024_11_0p50 -not_if :RH: -grib pgb2bfile_f024_11_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_11_0p50.new -10: 1:0:d=2021032312:TMP:475 mb:24 hour fcst: -10: 2:144951:d=2021032312:RH:475 mb:24 hour fcst: -10: 3:300286:d=2021032312:TCDC:475 mb:24 hour fcst: -10: 4:389510:d=2021032312:VVEL:475 mb:24 hour fcst: -10: 5:617331:d=2021032312:DZDT:475 mb:24 hour fcst: -10: 6:872858:d=2021032312:UGRD:475 mb:24 hour fcst: -10: 7:1054169:d=2021032312:VGRD:475 mb:24 hour fcst: -10: 8:1244405:d=2021032312:ABSV:475 mb:24 hour fcst: -10: 9:1422849:d=2021032312:CLMR:475 mb:24 hour fcst: -10: 10:1433121:d=2021032312:ICMR:475 mb:24 hour fcst: -10: 11:1563916:d=2021032312:RWMR:475 mb:24 hour fcst: -10: 12:1567742:d=2021032312:SNMR:475 mb:24 hour fcst: -10: 13:1673328:d=2021032312:GRLE:475 mb:24 hour fcst: -10: 14:1682088:d=2021032312:HGT:525 mb:24 hour fcst: -10: + product_functions.sh[10]rc=0 -10: + product_functions.sh[11](( rc == 0 )) -10: + product_functions.sh[11]mv pgb2bfile_f024_11_0p50.new pgb2bfile_f024_11_0p50 -10: + product_functions.sh[12]return 0 -10: + interp_atmos_master.sh[56]export err=0 -10: + interp_atmos_master.sh[56]err=0 -10: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -10: ++ interp_atmos_master.sh[62]wc -l -10: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_11_0p50 -match 'LAND|ICEC' -10: + interp_atmos_master.sh[62]var_count=0 -10: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -10: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -10: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_11_1p00 -10: + product_functions.sh[5]local filename=pgb2bfile_f024_11_1p00 -10: + product_functions.sh[6]wgrib2 pgb2bfile_f024_11_1p00 -not_if :RH: -grib pgb2bfile_f024_11_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_11_1p00.new -10: 1:0:d=2021032312:TMP:475 mb:24 hour fcst: -10: 2:47432:d=2021032312:RH:475 mb:24 hour fcst: -10: 3:98433:d=2021032312:TCDC:475 mb:24 hour fcst: -10: 4:130165:d=2021032312:VVEL:475 mb:24 hour fcst: -10: 5:199582:d=2021032312:DZDT:475 mb:24 hour fcst: -10: 6:276312:d=2021032312:UGRD:475 mb:24 hour fcst: -10: 7:334076:d=2021032312:VGRD:475 mb:24 hour fcst: -10: 8:394306:d=2021032312:ABSV:475 mb:24 hour fcst: -10: 9:451008:d=2021032312:CLMR:475 mb:24 hour fcst: -10: 10:454891:d=2021032312:ICMR:475 mb:24 hour fcst: -10: 11:498124:d=2021032312:RWMR:475 mb:24 hour fcst: -10: 12:499680:d=2021032312:SNMR:475 mb:24 hour fcst: -10: 13:535186:d=2021032312:GRLE:475 mb:24 hour fcst: -10: 14:538829:d=2021032312:HGT:525 mb:24 hour fcst: -10: + product_functions.sh[10]rc=0 -10: + product_functions.sh[11](( rc == 0 )) -10: + product_functions.sh[11]mv pgb2bfile_f024_11_1p00.new pgb2bfile_f024_11_1p00 -10: + product_functions.sh[12]return 0 -10: + interp_atmos_master.sh[56]export err=0 -10: + interp_atmos_master.sh[56]err=0 -10: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -10: ++ interp_atmos_master.sh[62]wc -l -10: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_11_1p00 -match 'LAND|ICEC' -10: + interp_atmos_master.sh[62]var_count=0 -10: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -10: + interp_atmos_master.sh[73]exit 0 -20: + bash[8]'[' -z '' ']' -20: + bash[9]case "$-" in -20: + bash[12]__lmod_vx=x -20: + bash[16]'[' -n x ']' -20: + bash[16]set +x -20: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -20: Shell debugging restarted -20: + bash[224]unset __lmod_vx -20: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_21 -20: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_21 -20: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -20: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -20: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -20: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -20: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -20: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -20: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -20: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -20: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -20: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -20: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -20: + interp_atmos_master.sh[31]IFS=: -20: + interp_atmos_master.sh[31]read -ra grids -20: + interp_atmos_master.sh[33]output_grids= -20: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -20: + interp_atmos_master.sh[35]gridopt=grid0p25 -20: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_21_0p25' -20: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -20: + interp_atmos_master.sh[35]gridopt=grid0p50 -20: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_21_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_21_0p50' -20: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -20: + interp_atmos_master.sh[35]gridopt=grid1p00 -20: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_21_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_21_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_21_1p00' -20: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_21 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_21_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_21_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_21_1p00 -20: 1:0:d=2021032312:UGRD:610 m above mean sea level:24 hour fcst: -20: 2:20852:d=2021032312:VGRD:610 m above mean sea level:24 hour fcst: -20: 3:42043:d=2021032312:TMP:914 m above mean sea level:24 hour fcst: -20: 4:62933:d=2021032312:UGRD:914 m above mean sea level:24 hour fcst: -20: 5:84306:d=2021032312:VGRD:914 m above mean sea level:24 hour fcst: -20: 6:106158:d=2021032312:TMP:4572 m above mean sea level:24 hour fcst: -20: 7:127332:d=2021032312:UGRD:4572 m above mean sea level:24 hour fcst: -20: 8:150726:d=2021032312:VGRD:4572 m above mean sea level:24 hour fcst: -20: 9:174696:d=2021032312:TMP:60-30 mb above ground:24 hour fcst: -20: 10:195517:d=2021032312:RH:60-30 mb above ground:24 hour fcst: -20: 11:214215:d=2021032312:SPFH:60-30 mb above ground:24 hour fcst: -20: 12:242962:d=2021032312:UGRD:60-30 mb above ground:24 hour fcst: -20: 13:264589:d=2021032312:VGRD:60-30 mb above ground:24 hour fcst: -20: 14:286653:d=2021032312:TMP:90-60 mb above ground:24 hour fcst: -20: + interp_atmos_master.sh[47]export err=0 -20: + interp_atmos_master.sh[47]err=0 -20: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -20: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -20: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_21_0p25 -20: + product_functions.sh[5]local filename=pgb2bfile_f024_21_0p25 -20: + product_functions.sh[6]wgrib2 pgb2bfile_f024_21_0p25 -not_if :RH: -grib pgb2bfile_f024_21_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_21_0p25.new -20: 1:0:d=2021032312:UGRD:610 m above mean sea level:24 hour fcst: -20: 2:592097:d=2021032312:VGRD:610 m above mean sea level:24 hour fcst: -20: 3:1209032:d=2021032312:TMP:914 m above mean sea level:24 hour fcst: -20: 4:1794108:d=2021032312:UGRD:914 m above mean sea level:24 hour fcst: -20: 5:2396693:d=2021032312:VGRD:914 m above mean sea level:24 hour fcst: -20: 6:3026886:d=2021032312:TMP:4572 m above mean sea level:24 hour fcst: -20: 7:3600354:d=2021032312:UGRD:4572 m above mean sea level:24 hour fcst: -20: 8:4246341:d=2021032312:VGRD:4572 m above mean sea level:24 hour fcst: -20: 9:4927512:d=2021032312:TMP:60-30 mb above ground:24 hour fcst: -20: 10:5413290:d=2021032312:RH:60-30 mb above ground:24 hour fcst: -20: 11:5824811:d=2021032312:SPFH:60-30 mb above ground:24 hour fcst: -20: 12:6611094:d=2021032312:UGRD:60-30 mb above ground:24 hour fcst: -20: 13:7133757:d=2021032312:VGRD:60-30 mb above ground:24 hour fcst: -20: 14:7684491:d=2021032312:TMP:90-60 mb above ground:24 hour fcst: -20: + product_functions.sh[10]rc=0 -20: + product_functions.sh[11](( rc == 0 )) -20: + product_functions.sh[11]mv pgb2bfile_f024_21_0p25.new pgb2bfile_f024_21_0p25 -20: + product_functions.sh[12]return 0 -20: + interp_atmos_master.sh[56]export err=0 -20: + interp_atmos_master.sh[56]err=0 -20: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -20: ++ interp_atmos_master.sh[62]wc -l -20: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_21_0p25 -match 'LAND|ICEC' -20: + interp_atmos_master.sh[62]var_count=0 -20: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -20: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -20: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_21_0p50 -20: + product_functions.sh[5]local filename=pgb2bfile_f024_21_0p50 -20: + product_functions.sh[6]wgrib2 pgb2bfile_f024_21_0p50 -not_if :RH: -grib pgb2bfile_f024_21_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_21_0p50.new -20: 1:0:d=2021032312:UGRD:610 m above mean sea level:24 hour fcst: -20: 2:191321:d=2021032312:VGRD:610 m above mean sea level:24 hour fcst: -20: 3:388355:d=2021032312:TMP:914 m above mean sea level:24 hour fcst: -20: 4:575753:d=2021032312:UGRD:914 m above mean sea level:24 hour fcst: -20: 5:769175:d=2021032312:VGRD:914 m above mean sea level:24 hour fcst: -20: 6:970580:d=2021032312:TMP:4572 m above mean sea level:24 hour fcst: -20: 7:1153523:d=2021032312:UGRD:4572 m above mean sea level:24 hour fcst: -20: 8:1360943:d=2021032312:VGRD:4572 m above mean sea level:24 hour fcst: -20: 9:1577060:d=2021032312:TMP:60-30 mb above ground:24 hour fcst: -20: 10:1745628:d=2021032312:RH:60-30 mb above ground:24 hour fcst: -20: 11:1891650:d=2021032312:SPFH:60-30 mb above ground:24 hour fcst: -20: 12:2162382:d=2021032312:UGRD:60-30 mb above ground:24 hour fcst: -20: 13:2342538:d=2021032312:VGRD:60-30 mb above ground:24 hour fcst: -20: 14:2529700:d=2021032312:TMP:90-60 mb above ground:24 hour fcst: -20: + product_functions.sh[10]rc=0 -20: + product_functions.sh[11](( rc == 0 )) -20: + product_functions.sh[11]mv pgb2bfile_f024_21_0p50.new pgb2bfile_f024_21_0p50 -20: + product_functions.sh[12]return 0 -20: + interp_atmos_master.sh[56]export err=0 -20: + interp_atmos_master.sh[56]err=0 -20: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -20: ++ interp_atmos_master.sh[62]wc -l -20: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_21_0p50 -match 'LAND|ICEC' -20: + interp_atmos_master.sh[62]var_count=0 -20: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -20: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -20: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_21_1p00 -20: + product_functions.sh[5]local filename=pgb2bfile_f024_21_1p00 -20: + product_functions.sh[6]wgrib2 pgb2bfile_f024_21_1p00 -not_if :RH: -grib pgb2bfile_f024_21_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_21_1p00.new -20: 1:0:d=2021032312:UGRD:610 m above mean sea level:24 hour fcst: -20: 2:58608:d=2021032312:VGRD:610 m above mean sea level:24 hour fcst: -20: 3:118627:d=2021032312:TMP:914 m above mean sea level:24 hour fcst: -20: 4:176454:d=2021032312:UGRD:914 m above mean sea level:24 hour fcst: -20: 5:236117:d=2021032312:VGRD:914 m above mean sea level:24 hour fcst: -20: 6:297669:d=2021032312:TMP:4572 m above mean sea level:24 hour fcst: -20: 7:354739:d=2021032312:UGRD:4572 m above mean sea level:24 hour fcst: -20: 8:418805:d=2021032312:VGRD:4572 m above mean sea level:24 hour fcst: -20: 9:485438:d=2021032312:TMP:60-30 mb above ground:24 hour fcst: -20: 10:540285:d=2021032312:RH:60-30 mb above ground:24 hour fcst: -20: 11:588673:d=2021032312:SPFH:60-30 mb above ground:24 hour fcst: -20: 12:669430:d=2021032312:UGRD:60-30 mb above ground:24 hour fcst: -20: 13:727084:d=2021032312:VGRD:60-30 mb above ground:24 hour fcst: -20: 14:786826:d=2021032312:TMP:90-60 mb above ground:24 hour fcst: -20: + product_functions.sh[10]rc=0 -20: + product_functions.sh[11](( rc == 0 )) -20: + product_functions.sh[11]mv pgb2bfile_f024_21_1p00.new pgb2bfile_f024_21_1p00 -20: + product_functions.sh[12]return 0 -20: + interp_atmos_master.sh[56]export err=0 -20: + interp_atmos_master.sh[56]err=0 -20: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -20: ++ interp_atmos_master.sh[62]wc -l -20: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_21_1p00 -match 'LAND|ICEC' -20: + interp_atmos_master.sh[62]var_count=0 -20: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -20: + interp_atmos_master.sh[73]exit 0 -14: + bash[8]'[' -z '' ']' -14: + bash[9]case "$-" in -14: + bash[12]__lmod_vx=x -14: + bash[16]'[' -n x ']' -14: + bash[16]set +x -14: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -14: Shell debugging restarted -14: + bash[224]unset __lmod_vx -14: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_15 -14: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_15 -14: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -14: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -14: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -14: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -14: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -14: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -14: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -14: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -14: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -14: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -14: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -14: + interp_atmos_master.sh[31]IFS=: -14: + interp_atmos_master.sh[31]read -ra grids -14: + interp_atmos_master.sh[33]output_grids= -14: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -14: + interp_atmos_master.sh[35]gridopt=grid0p25 -14: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_15_0p25' -14: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -14: + interp_atmos_master.sh[35]gridopt=grid0p50 -14: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_15_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_15_0p50' -14: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -14: + interp_atmos_master.sh[35]gridopt=grid1p00 -14: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_15_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_15_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_15_1p00' -14: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_15 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_15_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_15_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_15_1p00 -14: 1:0:d=2021032312:TMP:675 mb:24 hour fcst: -14: 2:19448:d=2021032312:RH:675 mb:24 hour fcst: -14: 3:39740:d=2021032312:TCDC:675 mb:24 hour fcst: -14: 4:51454:d=2021032312:VVEL:675 mb:24 hour fcst: -14: 5:77096:d=2021032312:DZDT:675 mb:24 hour fcst: -14: 6:104508:d=2021032312:UGRD:675 mb:24 hour fcst: -14: 7:125596:d=2021032312:VGRD:675 mb:24 hour fcst: -14: 8:147275:d=2021032312:ABSV:675 mb:24 hour fcst: -14: 9:168315:d=2021032312:CLMR:675 mb:24 hour fcst: -14: 10:174007:d=2021032312:ICMR:675 mb:24 hour fcst: -14: 11:185902:d=2021032312:RWMR:675 mb:24 hour fcst: -14: 12:190832:d=2021032312:SNMR:675 mb:24 hour fcst: -14: 13:205456:d=2021032312:GRLE:675 mb:24 hour fcst: -14: 14:207579:d=2021032312:HGT:725 mb:24 hour fcst: -14: + interp_atmos_master.sh[47]export err=0 -14: + interp_atmos_master.sh[47]err=0 -14: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -14: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -14: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_15_0p25 -14: + product_functions.sh[5]local filename=pgb2bfile_f024_15_0p25 -14: + product_functions.sh[6]wgrib2 pgb2bfile_f024_15_0p25 -not_if :RH: -grib pgb2bfile_f024_15_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_15_0p25.new -14: 1:0:d=2021032312:TMP:675 mb:24 hour fcst: -14: 2:439440:d=2021032312:RH:675 mb:24 hour fcst: -14: 3:912992:d=2021032312:TCDC:675 mb:24 hour fcst: -14: 4:1139355:d=2021032312:VVEL:675 mb:24 hour fcst: -14: 5:1820895:d=2021032312:DZDT:675 mb:24 hour fcst: -14: 6:2577156:d=2021032312:UGRD:675 mb:24 hour fcst: -14: 7:3080805:d=2021032312:VGRD:675 mb:24 hour fcst: -14: 8:3615394:d=2021032312:ABSV:675 mb:24 hour fcst: -14: 9:4115051:d=2021032312:CLMR:675 mb:24 hour fcst: -14: 10:4235335:d=2021032312:ICMR:675 mb:24 hour fcst: -14: 11:4504601:d=2021032312:RWMR:675 mb:24 hour fcst: -14: 12:4602109:d=2021032312:SNMR:675 mb:24 hour fcst: -14: 13:4920566:d=2021032312:GRLE:675 mb:24 hour fcst: -14: 14:4956889:d=2021032312:HGT:725 mb:24 hour fcst: -14: + product_functions.sh[10]rc=0 -14: + product_functions.sh[11](( rc == 0 )) -14: + product_functions.sh[11]mv pgb2bfile_f024_15_0p25.new pgb2bfile_f024_15_0p25 -14: + product_functions.sh[12]return 0 -14: + interp_atmos_master.sh[56]export err=0 -14: + interp_atmos_master.sh[56]err=0 -14: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -14: ++ interp_atmos_master.sh[62]wc -l -14: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_15_0p25 -match 'LAND|ICEC' -14: + interp_atmos_master.sh[62]var_count=0 -14: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -14: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -14: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_15_0p50 -14: + product_functions.sh[5]local filename=pgb2bfile_f024_15_0p50 -14: + product_functions.sh[6]wgrib2 pgb2bfile_f024_15_0p50 -not_if :RH: -grib pgb2bfile_f024_15_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_15_0p50.new -14: 1:0:d=2021032312:TMP:675 mb:24 hour fcst: -14: 2:152582:d=2021032312:RH:675 mb:24 hour fcst: -14: 3:315880:d=2021032312:TCDC:675 mb:24 hour fcst: -14: 4:401777:d=2021032312:VVEL:675 mb:24 hour fcst: -14: 5:634476:d=2021032312:DZDT:675 mb:24 hour fcst: -14: 6:884404:d=2021032312:UGRD:675 mb:24 hour fcst: -14: 7:1057555:d=2021032312:VGRD:675 mb:24 hour fcst: -14: 8:1240240:d=2021032312:ABSV:675 mb:24 hour fcst: -14: 9:1413057:d=2021032312:CLMR:675 mb:24 hour fcst: -14: 10:1458523:d=2021032312:ICMR:675 mb:24 hour fcst: -14: 11:1552869:d=2021032312:RWMR:675 mb:24 hour fcst: -14: 12:1589198:d=2021032312:SNMR:675 mb:24 hour fcst: -14: 13:1703733:d=2021032312:GRLE:675 mb:24 hour fcst: -14: 14:1717457:d=2021032312:HGT:725 mb:24 hour fcst: -14: + product_functions.sh[10]rc=0 -14: + product_functions.sh[11](( rc == 0 )) -14: + product_functions.sh[11]mv pgb2bfile_f024_15_0p50.new pgb2bfile_f024_15_0p50 -14: + product_functions.sh[12]return 0 -14: + interp_atmos_master.sh[56]export err=0 -14: + interp_atmos_master.sh[56]err=0 -14: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -14: ++ interp_atmos_master.sh[62]wc -l -14: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_15_0p50 -match 'LAND|ICEC' -14: + interp_atmos_master.sh[62]var_count=0 -14: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -14: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -14: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_15_1p00 -14: + product_functions.sh[5]local filename=pgb2bfile_f024_15_1p00 -14: + product_functions.sh[6]wgrib2 pgb2bfile_f024_15_1p00 -not_if :RH: -grib pgb2bfile_f024_15_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_15_1p00.new -14: 1:0:d=2021032312:TMP:675 mb:24 hour fcst: -14: 2:50329:d=2021032312:RH:675 mb:24 hour fcst: -14: 3:103610:d=2021032312:TCDC:675 mb:24 hour fcst: -14: 4:134516:d=2021032312:VVEL:675 mb:24 hour fcst: -14: 5:205823:d=2021032312:DZDT:675 mb:24 hour fcst: -14: 6:280958:d=2021032312:UGRD:675 mb:24 hour fcst: -14: 7:336806:d=2021032312:VGRD:675 mb:24 hour fcst: -14: 8:395057:d=2021032312:ABSV:675 mb:24 hour fcst: -14: 9:450556:d=2021032312:CLMR:675 mb:24 hour fcst: -14: 10:466593:d=2021032312:ICMR:675 mb:24 hour fcst: -14: 11:497899:d=2021032312:RWMR:675 mb:24 hour fcst: -14: 12:510549:d=2021032312:SNMR:675 mb:24 hour fcst: -14: 13:549072:d=2021032312:GRLE:675 mb:24 hour fcst: -14: 14:554531:d=2021032312:HGT:725 mb:24 hour fcst: -14: + product_functions.sh[10]rc=0 -14: + product_functions.sh[11](( rc == 0 )) -14: + product_functions.sh[11]mv pgb2bfile_f024_15_1p00.new pgb2bfile_f024_15_1p00 -14: + product_functions.sh[12]return 0 -14: + interp_atmos_master.sh[56]export err=0 -14: + interp_atmos_master.sh[56]err=0 -14: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -14: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_15_1p00 -match 'LAND|ICEC' -14: ++ interp_atmos_master.sh[62]wc -l -14: + interp_atmos_master.sh[62]var_count=0 -14: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -14: + interp_atmos_master.sh[73]exit 0 - 5: + bash[8]'[' -z '' ']' - 5: + bash[9]case "$-" in - 5: + bash[12]__lmod_vx=x - 5: + bash[16]'[' -n x ']' - 5: + bash[16]set +x - 5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 5: Shell debugging restarted - 5: + bash[224]unset __lmod_vx - 5: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_6 - 5: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_6 - 5: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 - 5: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 5: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 5: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 5: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 5: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 5: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 5: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 5: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 5: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 5: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 5: + interp_atmos_master.sh[31]IFS=: - 5: + interp_atmos_master.sh[31]read -ra grids - 5: + interp_atmos_master.sh[33]output_grids= - 5: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 5: + interp_atmos_master.sh[35]gridopt=grid0p25 - 5: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_6_0p25' - 5: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 5: + interp_atmos_master.sh[35]gridopt=grid0p50 - 5: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_6_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_6_0p50' - 5: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 5: + interp_atmos_master.sh[35]gridopt=grid1p00 - 5: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_6_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_6_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_6_1p00' - 5: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_6 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_6_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_6_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_6_1p00 - 5: 1:0:d=2021032312:TMP:225 mb:24 hour fcst: - 5: 2:18439:d=2021032312:RH:225 mb:24 hour fcst: - 5: 3:35067:d=2021032312:TCDC:225 mb:24 hour fcst: - 5: 4:43732:d=2021032312:VVEL:225 mb:24 hour fcst: - 5: 5:71297:d=2021032312:DZDT:225 mb:24 hour fcst: - 5: 6:98097:d=2021032312:UGRD:225 mb:24 hour fcst: - 5: 7:112077:d=2021032312:VGRD:225 mb:24 hour fcst: - 5: 8:126449:d=2021032312:ABSV:225 mb:24 hour fcst: - 5: 9:147166:d=2021032312:CLMR:225 mb:24 hour fcst: - 5: 10:147345:d=2021032312:ICMR:225 mb:24 hour fcst: - 5: 11:158817:d=2021032312:RWMR:225 mb:24 hour fcst: - 5: 12:158996:d=2021032312:SNMR:225 mb:24 hour fcst: - 5: 13:165921:d=2021032312:GRLE:225 mb:24 hour fcst: - 5: 14:166232:d=2021032312:HGT:275 mb:24 hour fcst: - 5: + interp_atmos_master.sh[47]export err=0 - 5: + interp_atmos_master.sh[47]err=0 - 5: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 5: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 5: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_6_0p25 - 5: + product_functions.sh[5]local filename=pgb2bfile_f024_6_0p25 - 5: + product_functions.sh[6]wgrib2 pgb2bfile_f024_6_0p25 -not_if :RH: -grib pgb2bfile_f024_6_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_6_0p25.new - 5: 1:0:d=2021032312:TMP:225 mb:24 hour fcst: - 5: 2:420782:d=2021032312:RH:225 mb:24 hour fcst: - 5: 3:805483:d=2021032312:TCDC:225 mb:24 hour fcst: - 5: 4:979058:d=2021032312:VVEL:225 mb:24 hour fcst: - 5: 5:1739083:d=2021032312:DZDT:225 mb:24 hour fcst: - 5: 6:2489214:d=2021032312:UGRD:225 mb:24 hour fcst: - 5: 7:2799537:d=2021032312:VGRD:225 mb:24 hour fcst: - 5: 8:3120229:d=2021032312:ABSV:225 mb:24 hour fcst: - 5: 9:3623795:d=2021032312:CLMR:225 mb:24 hour fcst: - 5: 10:3623974:d=2021032312:ICMR:225 mb:24 hour fcst: - 5: 11:3870987:d=2021032312:RWMR:225 mb:24 hour fcst: - 5: 12:3871166:d=2021032312:SNMR:225 mb:24 hour fcst: - 5: 13:4001334:d=2021032312:GRLE:225 mb:24 hour fcst: - 5: 14:4003763:d=2021032312:HGT:275 mb:24 hour fcst: - 5: + product_functions.sh[10]rc=0 - 5: + product_functions.sh[11](( rc == 0 )) - 5: + product_functions.sh[11]mv pgb2bfile_f024_6_0p25.new pgb2bfile_f024_6_0p25 - 5: + product_functions.sh[12]return 0 - 5: + interp_atmos_master.sh[56]export err=0 - 5: + interp_atmos_master.sh[56]err=0 - 5: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 5: ++ interp_atmos_master.sh[62]wc -l - 5: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_6_0p25 -match 'LAND|ICEC' - 5: + interp_atmos_master.sh[62]var_count=0 - 5: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 5: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 5: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_6_0p50 - 5: + product_functions.sh[5]local filename=pgb2bfile_f024_6_0p50 - 5: + product_functions.sh[6]wgrib2 pgb2bfile_f024_6_0p50 -not_if :RH: -grib pgb2bfile_f024_6_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_6_0p50.new - 5: 1:0:d=2021032312:TMP:225 mb:24 hour fcst: - 5: 2:144549:d=2021032312:RH:225 mb:24 hour fcst: - 5: 3:275984:d=2021032312:TCDC:225 mb:24 hour fcst: - 5: 4:340774:d=2021032312:VVEL:225 mb:24 hour fcst: - 5: 5:598681:d=2021032312:DZDT:225 mb:24 hour fcst: - 5: 6:842626:d=2021032312:UGRD:225 mb:24 hour fcst: - 5: 7:943217:d=2021032312:VGRD:225 mb:24 hour fcst: - 5: 8:1048205:d=2021032312:ABSV:225 mb:24 hour fcst: - 5: 9:1219299:d=2021032312:CLMR:225 mb:24 hour fcst: - 5: 10:1219478:d=2021032312:ICMR:225 mb:24 hour fcst: - 5: 11:1308439:d=2021032312:RWMR:225 mb:24 hour fcst: - 5: 12:1308618:d=2021032312:SNMR:225 mb:24 hour fcst: - 5: 13:1355402:d=2021032312:GRLE:225 mb:24 hour fcst: - 5: 14:1356454:d=2021032312:HGT:275 mb:24 hour fcst: - 5: + product_functions.sh[10]rc=0 - 5: + product_functions.sh[11](( rc == 0 )) - 5: + product_functions.sh[11]mv pgb2bfile_f024_6_0p50.new pgb2bfile_f024_6_0p50 - 5: + product_functions.sh[12]return 0 - 5: + interp_atmos_master.sh[56]export err=0 - 5: + interp_atmos_master.sh[56]err=0 - 5: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 5: ++ interp_atmos_master.sh[62]wc -l - 5: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_6_0p50 -match 'LAND|ICEC' - 5: + interp_atmos_master.sh[62]var_count=0 - 5: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 5: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 5: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_6_1p00 - 5: + product_functions.sh[5]local filename=pgb2bfile_f024_6_1p00 - 5: + product_functions.sh[6]wgrib2 pgb2bfile_f024_6_1p00 -not_if :RH: -grib pgb2bfile_f024_6_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_6_1p00.new - 5: 1:0:d=2021032312:TMP:225 mb:24 hour fcst: - 5: 2:47835:d=2021032312:RH:225 mb:24 hour fcst: - 5: 3:90742:d=2021032312:TCDC:225 mb:24 hour fcst: - 5: 4:113306:d=2021032312:VVEL:225 mb:24 hour fcst: - 5: 5:190838:d=2021032312:DZDT:225 mb:24 hour fcst: - 5: 6:263949:d=2021032312:UGRD:225 mb:24 hour fcst: - 5: 7:297707:d=2021032312:VGRD:225 mb:24 hour fcst: - 5: 8:333251:d=2021032312:ABSV:225 mb:24 hour fcst: - 5: 9:388129:d=2021032312:CLMR:225 mb:24 hour fcst: - 5: 10:388308:d=2021032312:ICMR:225 mb:24 hour fcst: - 5: 11:418123:d=2021032312:RWMR:225 mb:24 hour fcst: - 5: 12:418302:d=2021032312:SNMR:225 mb:24 hour fcst: - 5: 13:434953:d=2021032312:GRLE:225 mb:24 hour fcst: - 5: 14:435452:d=2021032312:HGT:275 mb:24 hour fcst: - 5: + product_functions.sh[10]rc=0 - 5: + product_functions.sh[11](( rc == 0 )) - 5: + product_functions.sh[11]mv pgb2bfile_f024_6_1p00.new pgb2bfile_f024_6_1p00 - 5: + product_functions.sh[12]return 0 - 5: + interp_atmos_master.sh[56]export err=0 - 5: + interp_atmos_master.sh[56]err=0 - 5: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 5: ++ interp_atmos_master.sh[62]wc -l - 5: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_6_1p00 -match 'LAND|ICEC' - 5: + interp_atmos_master.sh[62]var_count=0 - 5: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 5: + interp_atmos_master.sh[73]exit 0 -11: + bash[8]'[' -z '' ']' -11: + bash[9]case "$-" in -11: + bash[12]__lmod_vx=x -11: + bash[16]'[' -n x ']' -11: + bash[16]set +x -11: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -11: Shell debugging restarted -11: + bash[224]unset __lmod_vx -11: + interp_atmos_master.sh[7]input_file=tmpfileb_f024_12 -11: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f024_12 -11: + interp_atmos_master.sh[9]grid_string=0p25:0p50:1p00 -11: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -11: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -11: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -11: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -11: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -11: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -11: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -11: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -11: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -11: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -11: + interp_atmos_master.sh[31]IFS=: -11: + interp_atmos_master.sh[31]read -ra grids -11: + interp_atmos_master.sh[33]output_grids= -11: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -11: + interp_atmos_master.sh[35]gridopt=grid0p25 -11: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_12_0p25' -11: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -11: + interp_atmos_master.sh[35]gridopt=grid0p50 -11: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_12_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_12_0p50' -11: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -11: + interp_atmos_master.sh[35]gridopt=grid1p00 -11: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_12_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_12_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_12_1p00' -11: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f024_12 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f024_12_0p25 -new_grid latlon 0:720:0.5 90:361:-0.5 pgb2bfile_f024_12_0p50 -new_grid latlon 0:360:1.0 90:181:-1.0 pgb2bfile_f024_12_1p00 -11: 1:0:d=2021032312:TMP:525 mb:24 hour fcst: -11: 2:18630:d=2021032312:RH:525 mb:24 hour fcst: -11: 3:38345:d=2021032312:TCDC:525 mb:24 hour fcst: -11: 4:50381:d=2021032312:VVEL:525 mb:24 hour fcst: -11: 5:75457:d=2021032312:DZDT:525 mb:24 hour fcst: -11: 6:103079:d=2021032312:UGRD:525 mb:24 hour fcst: -11: 7:124210:d=2021032312:VGRD:525 mb:24 hour fcst: -11: 8:146112:d=2021032312:ABSV:525 mb:24 hour fcst: -11: 9:167143:d=2021032312:CLMR:525 mb:24 hour fcst: -11: 10:169204:d=2021032312:ICMR:525 mb:24 hour fcst: -11: 11:186192:d=2021032312:RWMR:525 mb:24 hour fcst: -11: 12:187357:d=2021032312:SNMR:525 mb:24 hour fcst: -11: 13:201710:d=2021032312:GRLE:525 mb:24 hour fcst: -11: 14:203252:d=2021032312:HGT:575 mb:24 hour fcst: -11: + interp_atmos_master.sh[47]export err=0 -11: + interp_atmos_master.sh[47]err=0 -11: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -11: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -11: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_12_0p25 -11: + product_functions.sh[5]local filename=pgb2bfile_f024_12_0p25 -11: + product_functions.sh[6]wgrib2 pgb2bfile_f024_12_0p25 -not_if :RH: -grib pgb2bfile_f024_12_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_12_0p25.new -11: 1:0:d=2021032312:TMP:525 mb:24 hour fcst: -11: 2:425817:d=2021032312:RH:525 mb:24 hour fcst: -11: 3:876341:d=2021032312:TCDC:525 mb:24 hour fcst: -11: 4:1104601:d=2021032312:VVEL:525 mb:24 hour fcst: -11: 5:1772952:d=2021032312:DZDT:525 mb:24 hour fcst: -11: 6:2546680:d=2021032312:UGRD:525 mb:24 hour fcst: -11: 7:3064804:d=2021032312:VGRD:525 mb:24 hour fcst: -11: 8:3610569:d=2021032312:ABSV:525 mb:24 hour fcst: -11: 9:4110532:d=2021032312:CLMR:525 mb:24 hour fcst: -11: 10:4152971:d=2021032312:ICMR:525 mb:24 hour fcst: -11: 11:4536207:d=2021032312:RWMR:525 mb:24 hour fcst: -11: 12:4555738:d=2021032312:SNMR:525 mb:24 hour fcst: -11: 13:4865477:d=2021032312:GRLE:525 mb:24 hour fcst: -11: 14:4889055:d=2021032312:HGT:575 mb:24 hour fcst: -11: + product_functions.sh[10]rc=0 -11: + product_functions.sh[11](( rc == 0 )) -11: + product_functions.sh[11]mv pgb2bfile_f024_12_0p25.new pgb2bfile_f024_12_0p25 -11: + product_functions.sh[12]return 0 -11: + interp_atmos_master.sh[56]export err=0 -11: + interp_atmos_master.sh[56]err=0 -11: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -11: ++ interp_atmos_master.sh[62]wc -l -11: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_12_0p25 -match 'LAND|ICEC' -11: + interp_atmos_master.sh[62]var_count=0 -11: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -11: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -11: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_12_0p50 -11: + product_functions.sh[5]local filename=pgb2bfile_f024_12_0p50 -11: + product_functions.sh[6]wgrib2 pgb2bfile_f024_12_0p50 -not_if :RH: -grib pgb2bfile_f024_12_0p50.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_12_0p50.new -11: 1:0:d=2021032312:TMP:525 mb:24 hour fcst: -11: 2:146716:d=2021032312:RH:525 mb:24 hour fcst: -11: 3:304091:d=2021032312:TCDC:525 mb:24 hour fcst: -11: 4:390833:d=2021032312:VVEL:525 mb:24 hour fcst: -11: 5:619444:d=2021032312:DZDT:525 mb:24 hour fcst: -11: 6:872463:d=2021032312:UGRD:525 mb:24 hour fcst: -11: 7:1050082:d=2021032312:VGRD:525 mb:24 hour fcst: -11: 8:1236436:d=2021032312:ABSV:525 mb:24 hour fcst: -11: 9:1409665:d=2021032312:CLMR:525 mb:24 hour fcst: -11: 10:1425802:d=2021032312:ICMR:525 mb:24 hour fcst: -11: 11:1562796:d=2021032312:RWMR:525 mb:24 hour fcst: -11: 12:1570454:d=2021032312:SNMR:525 mb:24 hour fcst: -11: 13:1680004:d=2021032312:GRLE:525 mb:24 hour fcst: -11: 14:1689236:d=2021032312:HGT:575 mb:24 hour fcst: -11: + product_functions.sh[10]rc=0 -11: + product_functions.sh[11](( rc == 0 )) -11: + product_functions.sh[11]mv pgb2bfile_f024_12_0p50.new pgb2bfile_f024_12_0p50 -11: + product_functions.sh[12]return 0 -11: + interp_atmos_master.sh[56]export err=0 -11: + interp_atmos_master.sh[56]err=0 -11: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -11: ++ interp_atmos_master.sh[62]wc -l -11: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_12_0p50 -match 'LAND|ICEC' -11: + interp_atmos_master.sh[62]var_count=0 -11: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -11: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -11: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f024_12_1p00 -11: + product_functions.sh[5]local filename=pgb2bfile_f024_12_1p00 -11: + product_functions.sh[6]wgrib2 pgb2bfile_f024_12_1p00 -not_if :RH: -grib pgb2bfile_f024_12_1p00.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f024_12_1p00.new -11: 1:0:d=2021032312:TMP:525 mb:24 hour fcst: -11: 2:48092:d=2021032312:RH:525 mb:24 hour fcst: -11: 3:99646:d=2021032312:TCDC:525 mb:24 hour fcst: -11: 4:130543:d=2021032312:VVEL:525 mb:24 hour fcst: -11: 5:200297:d=2021032312:DZDT:525 mb:24 hour fcst: -11: 6:276528:d=2021032312:UGRD:525 mb:24 hour fcst: -11: 7:333167:d=2021032312:VGRD:525 mb:24 hour fcst: -11: 8:392347:d=2021032312:ABSV:525 mb:24 hour fcst: -11: 9:448150:d=2021032312:CLMR:525 mb:24 hour fcst: -11: 10:453973:d=2021032312:ICMR:525 mb:24 hour fcst: -11: 11:499239:d=2021032312:RWMR:525 mb:24 hour fcst: -11: 12:502255:d=2021032312:SNMR:525 mb:24 hour fcst: -11: 13:539222:d=2021032312:GRLE:525 mb:24 hour fcst: -11: 14:542905:d=2021032312:HGT:575 mb:24 hour fcst: -11: + product_functions.sh[10]rc=0 -11: + product_functions.sh[11](( rc == 0 )) -11: + product_functions.sh[11]mv pgb2bfile_f024_12_1p00.new pgb2bfile_f024_12_1p00 -11: + product_functions.sh[12]return 0 -11: + interp_atmos_master.sh[56]export err=0 -11: + interp_atmos_master.sh[56]err=0 -11: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -11: ++ interp_atmos_master.sh[62]wc -l -11: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f024_12_1p00 -match 'LAND|ICEC' -11: + interp_atmos_master.sh[62]var_count=0 -11: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -11: + interp_atmos_master.sh[73]exit 0 -+ run_mpmd.sh[113]exit 0 -+ run_mpmd.sh[1]postamble run_mpmd.sh 1753755935 0 -+ preamble.sh[62]set +x -End run_mpmd.sh at 02:25:58 with error code 0 (time elapsed: 00:00:23) -+ exglobal_atmos_products.sh[142]true -+ exglobal_atmos_products.sh[143]export err=0 -+ exglobal_atmos_products.sh[143]err=0 -+ exglobal_atmos_products.sh[144][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[149]mv mpmd.out mpmd_2.out -+ exglobal_atmos_products.sh[153]echo 'Concatenating processor-specific grib2 files into a single product file' -Concatenating processor-specific grib2 files into a single product file -+ exglobal_atmos_products.sh[154](( iproc = 1 )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_1_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_1_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_1_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_1_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_1_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_1_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_1 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_2_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_2_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_2_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_2_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_2_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_2_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_2 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_3_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_3_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_3_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_3_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_3_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_3_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_3 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_4_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_4_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_4_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_4_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_4_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_4_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_4 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_5_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_5_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_5_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_5_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_5_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_5_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_5 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_6_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_6_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_6_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_6_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_6_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_6_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_6 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_7_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_7_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_7_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_7_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_7_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_7_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_7 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_8_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_8_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_8_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_8_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_8_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_8_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_8 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_9_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_9_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_9_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_9_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_9_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_9_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_9 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_10_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_10_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_10_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_10_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_10_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_10_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_10 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_11_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_11_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_11_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_11_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_11_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_11_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_11 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_12_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_12_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_12_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_12_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_12_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_12_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_12 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_13_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_13_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_13_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_13_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_13_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_13_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_13 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_14_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_14_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_14_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_14_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_14_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_14_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_14 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_15_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_15_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_15_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_15_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_15_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_15_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_15 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_16_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_16_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_16_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_16_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_16_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_16_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_16 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_17_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_17_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_17_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_17_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_17_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_17_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_17 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_18_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_18_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_18_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_18_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_18_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_18_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_18 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_19_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_19_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_19_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_19_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_19_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_19_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_19 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_20_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_20_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_20_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_20_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_20_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_20_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_20 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_21_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_21_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_21_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_21_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_21_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_21_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_21 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_22_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_22_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_22_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_22_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_22_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_22_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_22 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_23_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_23_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_23_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_23_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_23_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_23_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_23 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_24_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_24_0p25 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_24_0p50 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_24_0p50 -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f024_24_1p00 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f024_24_1p00 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f024_24 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[164]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[165]prod_dir=COMOUT_ATMOS_GRIB_0p25 -+ exglobal_atmos_products.sh[166]cpfs pgb2bfile_f024_0p25 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f024 -+ cpfs[3]'[' 2 -ne 2 ']' -+ cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f024 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f024 = ./ ']' -+ cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f024 ']' -+ cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f024 -+ cpfs[16]cp pgb2bfile_f024_0p25 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f024.cptmp -+ cpfs[18]'[' 0 -ne 0 ']' -+ cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f024.cptmp -+ cpfs[23]'[' 0 -ne 0 ']' -+ cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f024.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f024 -+ cpfs[28]'[' 0 -ne 0 ']' -+ exglobal_atmos_products.sh[167]wgrib2 -s pgb2bfile_f024_0p25 -+ exglobal_atmos_products.sh[164]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[165]prod_dir=COMOUT_ATMOS_GRIB_0p50 -+ exglobal_atmos_products.sh[166]cpfs pgb2bfile_f024_0p50 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50/gfs.t12z.pgrb2b.0p50.f024 -+ cpfs[3]'[' 2 -ne 2 ']' -+ cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50/gfs.t12z.pgrb2b.0p50.f024 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50/gfs.t12z.pgrb2b.0p50.f024 = ./ ']' -+ cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50/gfs.t12z.pgrb2b.0p50.f024 ']' -+ cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50/gfs.t12z.pgrb2b.0p50.f024 -+ cpfs[16]cp pgb2bfile_f024_0p50 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50/gfs.t12z.pgrb2b.0p50.f024.cptmp -+ cpfs[18]'[' 0 -ne 0 ']' -+ cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50/gfs.t12z.pgrb2b.0p50.f024.cptmp -+ cpfs[23]'[' 0 -ne 0 ']' -+ cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50/gfs.t12z.pgrb2b.0p50.f024.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50/gfs.t12z.pgrb2b.0p50.f024 -+ cpfs[28]'[' 0 -ne 0 ']' -+ exglobal_atmos_products.sh[167]wgrib2 -s pgb2bfile_f024_0p50 -+ exglobal_atmos_products.sh[164]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[165]prod_dir=COMOUT_ATMOS_GRIB_1p00 -+ exglobal_atmos_products.sh[166]cpfs pgb2bfile_f024_1p00 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.pgrb2b.1p00.f024 -+ cpfs[3]'[' 2 -ne 2 ']' -+ cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.pgrb2b.1p00.f024 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.pgrb2b.1p00.f024 = ./ ']' -+ cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.pgrb2b.1p00.f024 ']' -+ cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.pgrb2b.1p00.f024 -+ cpfs[16]cp pgb2bfile_f024_1p00 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.pgrb2b.1p00.f024.cptmp -+ cpfs[18]'[' 0 -ne 0 ']' -+ cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.pgrb2b.1p00.f024.cptmp -+ cpfs[23]'[' 0 -ne 0 ']' -+ cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.pgrb2b.1p00.f024.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.pgrb2b.1p00.f024 -+ cpfs[28]'[' 0 -ne 0 ']' -+ exglobal_atmos_products.sh[167]wgrib2 -s pgb2bfile_f024_1p00 -+ exglobal_atmos_products.sh[170]echo 'Finished processing nset = 2' -Finished processing nset = 2 -+ exglobal_atmos_products.sh[154](( nset++ )) -+ exglobal_atmos_products.sh[154](( nset <= downset )) -+ exglobal_atmos_products.sh[177]FLUX_FILE=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.sfluxgrbf024.grib2 -+ exglobal_atmos_products.sh[178][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.sfluxgrbf024.grib2 ]] -+ exglobal_atmos_products.sh[179]wgrib2 -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.sfluxgrbf024.grib2 -+ exglobal_atmos_products.sh[185][[ YES == \Y\E\S ]] -+ exglobal_atmos_products.sh[188]input_file=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.sfluxgrbf024.grib2 -+ exglobal_atmos_products.sh[189]output_file_prefix=sflux_f024 -+ exglobal_atmos_products.sh[190]grid_string=1p00 -+ exglobal_atmos_products.sh[191]/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_sflux.sh /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.sfluxgrbf024.grib2 sflux_f024 1p00 -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ interp_atmos_sflux.sh[6]input_file=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.sfluxgrbf024.grib2 -+ interp_atmos_sflux.sh[7]output_file_prefix=sflux_f024 -+ interp_atmos_sflux.sh[8]grid_string=1p00 -+ interp_atmos_sflux.sh[11]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -+ interp_atmos_sflux.sh[12]interp_winds='-new_grid_winds earth' -+ interp_atmos_sflux.sh[13]interp_bilinear='-new_grid_interpolation bilinear' -+ interp_atmos_sflux.sh[14]interp_neighbor='-if :(LAND|CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -+ interp_atmos_sflux.sh[15]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -+ interp_atmos_sflux.sh[16]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -+ interp_atmos_sflux.sh[20]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -+ interp_atmos_sflux.sh[22]grid0p50='latlon 0:720:0.5 90:361:-0.5' -+ interp_atmos_sflux.sh[24]grid1p00='latlon 0:360:1.0 90:181:-1.0' -+ interp_atmos_sflux.sh[27]IFS=: -+ interp_atmos_sflux.sh[27]read -ra grids -+ interp_atmos_sflux.sh[29]output_grids= -+ interp_atmos_sflux.sh[30]for grid in "${grids[@]}" -+ interp_atmos_sflux.sh[31]gridopt=grid1p00 -+ interp_atmos_sflux.sh[32]output_grids=' -new_grid latlon 0:360:1.0 90:181:-1.0 sflux_f024_1p00' -+ interp_atmos_sflux.sh[36]wgrib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.sfluxgrbf024.grib2 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(LAND|CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:360:1.0 90:181:-1.0 sflux_f024_1p00 -1:0:d=2021032312:HGT:1 hybrid level:24 hour fcst: -2:11781:d=2021032312:TMP:1 hybrid level:24 hour fcst: -3:32598:d=2021032312:SPFH:1 hybrid level:24 hour fcst: -4:61460:d=2021032312:UGRD:1 hybrid level:24 hour fcst: -5:81763:d=2021032312:VGRD:1 hybrid level:24 hour fcst: -6:102665:d=2021032312:PRES:surface:24 hour fcst: -7:128314:d=2021032312:HGT:surface:24 hour fcst: -8:144384:d=2021032312:TMP:surface:24 hour fcst: -9:161014:d=2021032312:TSOIL:0-0.1 m below ground:24 hour fcst: -10:172945:d=2021032312:SOILW:0-0.1 m below ground:24 hour fcst: -11:181856:d=2021032312:SOILL:0-0.1 m below ground:24 hour fcst: -12:190653:d=2021032312:TSOIL:0.1-0.4 m below ground:24 hour fcst: -13:206113:d=2021032312:SOILW:0.1-0.4 m below ground:24 hour fcst: -14:215130:d=2021032312:SOILL:0.1-0.4 m below ground:24 hour fcst: -15:224024:d=2021032312:TSOIL:0.4-1 m below ground:24 hour fcst: -16:238463:d=2021032312:SOILW:0.4-1 m below ground:24 hour fcst: -17:247596:d=2021032312:SOILL:0.4-1 m below ground:24 hour fcst: -18:256500:d=2021032312:TSOIL:1-2 m below ground:24 hour fcst: -19:270762:d=2021032312:SOILW:1-2 m below ground:24 hour fcst: -20:279822:d=2021032312:SOILL:1-2 m below ground:24 hour fcst: -21:288898:d=2021032312:CISOILM:0-2 m below ground:24 hour fcst: -22:299404:d=2021032312:CNWAT:surface:24 hour fcst: -23:304989:d=2021032312:WEASD:surface:24 hour fcst: -24:317786:d=2021032312:SNOWC:surface:18-24 hour ave fcst: -25:322813:d=2021032312:SNOD:surface:24 hour fcst: -26:337159:d=2021032312:PEVPR:surface:24 hour fcst: -27:352809:d=2021032312:ICETK:surface:24 hour fcst: -28:356223:d=2021032312:ACOND:surface:24 hour fcst: -29:373788:d=2021032312:EVCW:surface:18-24 hour ave fcst: -30:384421:d=2021032312:EVBS:surface:18-24 hour ave fcst: -31:398838:d=2021032312:TRANS:surface:18-24 hour ave fcst: -32:408057:d=2021032312:SBSNO:surface:18-24 hour ave fcst: -33:417055:d=2021032312:var discipline=2 center=7 local_table=1 parmcat=0 parm=233:surface:24 hour fcst: -34:425076:d=2021032312:var discipline=2 center=7 local_table=1 parmcat=0 parm=235:surface:24 hour fcst: -35:429121:d=2021032312:var discipline=2 center=7 local_table=1 parmcat=0 parm=236:surface:24 hour fcst: -36:435809:d=2021032312:var discipline=2 center=7 local_table=1 parmcat=0 parm=237:surface:18-24 hour acc fcst: -37:442696:d=2021032312:var discipline=2 center=7 local_table=1 parmcat=0 parm=238:surface:18-24 hour acc fcst: -38:448217:d=2021032312:var discipline=2 center=7 local_table=1 parmcat=0 parm=239:surface:18-24 hour acc fcst: -39:455852:d=2021032312:var discipline=2 center=7 local_table=1 parmcat=0 parm=235:surface:18-24 hour ave fcst: -40:460218:d=2021032312:TMP:2 m above ground:24 hour fcst: -41:480978:d=2021032312:SPFH:2 m above ground:24 hour fcst: -42:509845:d=2021032312:TMAX:2 m above ground:18-24 hour max fcst: -43:530541:d=2021032312:TMIN:2 m above ground:18-24 hour min fcst: -44:551198:d=2021032312:QMAX:2 m above ground:18-24 hour max fcst: -45:568327:d=2021032312:QMIN:2 m above ground:18-24 hour min fcst: -46:585279:d=2021032312:UGRD:10 m above ground:24 hour fcst: -47:605794:d=2021032312:VGRD:10 m above ground:24 hour fcst: -48:626835:d=2021032312:CPOFP:surface:24 hour fcst: -49:646942:d=2021032312:CPRAT:surface:18-24 hour ave fcst: -50:670996:d=2021032312:PRATE:surface:18-24 hour ave fcst: -51:696054:d=2021032312:SSRUN:surface:18-24 hour acc fcst: -52:703742:d=2021032312:WATR:surface:18-24 hour acc fcst: -53:716153:d=2021032312:LHTFL:surface:18-24 hour ave fcst: -54:737797:d=2021032312:SHTFL:surface:18-24 hour ave fcst: -55:758458:d=2021032312:GFLUX:surface:18-24 hour ave fcst: -56:770516:d=2021032312:SNOHF:surface:18-24 hour ave fcst: -57:774176:d=2021032312:UFLX:surface:18-24 hour ave fcst: -58:790419:d=2021032312:VFLX:surface:18-24 hour ave fcst: -59:806782:d=2021032312:SFCR:surface:24 hour fcst: -60:825248:d=2021032312:FRICV:surface:24 hour fcst: -61:846952:d=2021032312:U-GWD:surface:18-24 hour ave fcst: -62:861711:d=2021032312:V-GWD:surface:18-24 hour ave fcst: -63:876494:d=2021032312:SHTFL:surface:24 hour fcst: -64:897345:d=2021032312:LHTFL:surface:24 hour fcst: -65:919096:d=2021032312:SFEXC:surface:24 hour fcst: -66:935602:d=2021032312:VEG:surface:24 hour fcst: -67:943626:d=2021032312:GFLUX:surface:24 hour fcst: -68:955752:d=2021032312:VGTYP:surface:24 hour fcst: -69:964285:d=2021032312:SOTYP:surface:24 hour fcst: -70:972244:d=2021032312:SLTYP:surface:24 hour fcst: -71:982456:d=2021032312:WILT:surface:24 hour fcst: -72:992366:d=2021032312:FLDCP:surface:24 hour fcst: -73:1002572:d=2021032312:SUNSD:surface:24 hour fcst: -74:1019670:d=2021032312:PEVPR:surface:18-24 hour ave fcst: -75:1035363:d=2021032312:PWAT:entire atmosphere (considered as a single layer):24 hour fcst: -76:1062264:d=2021032312:LCDC:low cloud layer:18-24 hour ave fcst: -77:1082218:d=2021032312:MCDC:middle cloud layer:18-24 hour ave fcst: -78:1099486:d=2021032312:HCDC:high cloud layer:18-24 hour ave fcst: -79:1117585:d=2021032312:TCDC:entire atmosphere (considered as a single layer):18-24 hour ave fcst: -80:1137668:d=2021032312:PRES:convective cloud bottom level:24 hour fcst: -81:1151322:d=2021032312:PRES:low cloud bottom level:18-24 hour ave fcst: -82:1174162:d=2021032312:PRES:middle cloud bottom level:18-24 hour ave fcst: -83:1191889:d=2021032312:PRES:high cloud bottom level:18-24 hour ave fcst: -84:1210781:d=2021032312:PRES:convective cloud top level:24 hour fcst: -85:1226116:d=2021032312:PRES:low cloud top level:18-24 hour ave fcst: -86:1249130:d=2021032312:PRES:middle cloud top level:18-24 hour ave fcst: -87:1266841:d=2021032312:PRES:high cloud top level:18-24 hour ave fcst: -88:1286063:d=2021032312:TMP:low cloud top level:18-24 hour ave fcst: -89:1303219:d=2021032312:TMP:middle cloud top level:18-24 hour ave fcst: -90:1317265:d=2021032312:TMP:high cloud top level:18-24 hour ave fcst: -91:1333226:d=2021032312:TCDC:convective cloud layer:24 hour fcst: -92:1354044:d=2021032312:TCDC:boundary layer cloud layer:18-24 hour ave fcst: -93:1371959:d=2021032312:CWORK:entire atmosphere (considered as a single layer):18-24 hour ave fcst: -94:1380866:d=2021032312:DSWRF:surface:18-24 hour ave fcst: -95:1403145:d=2021032312:DUVB:surface:18-24 hour ave fcst: -96:1423835:d=2021032312:CDUVB:surface:18-24 hour ave fcst: -97:1439683:d=2021032312:DLWRF:surface:18-24 hour ave fcst: -98:1460953:d=2021032312:USWRF:surface:18-24 hour ave fcst: -99:1480737:d=2021032312:ULWRF:surface:18-24 hour ave fcst: -100:1499491:d=2021032312:USWRF:top of atmosphere:18-24 hour ave fcst: -101:1522080:d=2021032312:ULWRF:top of atmosphere:18-24 hour ave fcst: -102:1542843:d=2021032312:DSWRF:surface:24 hour fcst: -103:1560102:d=2021032312:DLWRF:surface:24 hour fcst: -104:1581995:d=2021032312:USWRF:surface:24 hour fcst: -105:1598742:d=2021032312:ULWRF:surface:24 hour fcst: -106:1617718:d=2021032312:CSDSF:surface:18-24 hour ave fcst: -107:1634156:d=2021032312:CSUSF:surface:18-24 hour ave fcst: -108:1651569:d=2021032312:CSUSF:top of atmosphere:18-24 hour ave fcst: -109:1669339:d=2021032312:DSWRF:top of atmosphere:18-24 hour ave fcst: -110:1681009:d=2021032312:CSDLF:surface:18-24 hour ave fcst: -111:1699133:d=2021032312:CSULF:surface:18-24 hour ave fcst: -112:1717704:d=2021032312:CSULF:top of atmosphere:18-24 hour ave fcst: -113:1734169:d=2021032312:VBDSF:surface:18-24 hour ave fcst: -114:1755463:d=2021032312:VDDSF:surface:18-24 hour ave fcst: -115:1777495:d=2021032312:NBDSF:surface:18-24 hour ave fcst: -116:1800083:d=2021032312:NDDSF:surface:18-24 hour ave fcst: -117:1824432:d=2021032312:HPBL:surface:24 hour fcst: -118:1855586:d=2021032312:LAND:surface:24 hour fcst: -119:1857527:d=2021032312:ICEC:surface:24 hour fcst: -120:1860692:d=2021032312:ALBDO:surface:18-24 hour ave fcst: -121:1875454:d=2021032312:FDNSSTMP:surface:24 hour fcst: -+ interp_atmos_sflux.sh[43]export err=0 -+ interp_atmos_sflux.sh[43]err=0 -+ interp_atmos_sflux.sh[44][[ 0 -ne 0 ]] -+ interp_atmos_sflux.sh[49]exit 0 -+ exglobal_atmos_products.sh[191]true -+ exglobal_atmos_products.sh[192]export err=0 -+ exglobal_atmos_products.sh[192]err=0 -+ exglobal_atmos_products.sh[193][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[198]IFS=: -+ exglobal_atmos_products.sh[198]read -ra grids -+ exglobal_atmos_products.sh[199]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[200]prod_dir=COMOUT_ATMOS_GRIB_1p00 -+ exglobal_atmos_products.sh[201]cpfs sflux_f024_1p00 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f024 -+ cpfs[3]'[' 2 -ne 2 ']' -+ cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f024 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f024 = ./ ']' -+ cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f024 ']' -+ cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f024 -+ cpfs[16]cp sflux_f024_1p00 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f024.cptmp -+ cpfs[18]'[' 0 -ne 0 ']' -+ cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f024.cptmp -+ cpfs[23]'[' 0 -ne 0 ']' -+ cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f024.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f024 -+ cpfs[28]'[' 0 -ne 0 ']' -+ exglobal_atmos_products.sh[202]wgrib2 -s sflux_f024_1p00 -+ exglobal_atmos_products.sh[207][[ YES == \Y\E\S ]] -+ exglobal_atmos_products.sh[208]grp= -+ exglobal_atmos_products.sh[209][[ 24 -gt 0 ]] -+ exglobal_atmos_products.sh[209][[ 24 -le 180 ]] -+ exglobal_atmos_products.sh[211]wgrib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f024 -d 597 -grib /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.wgne.f024 -597:257531724:d=2021032312:APCP:surface:18-24 hour acc fcst: -+ exglobal_atmos_products.sh[219][[ NO == \Y\E\S ]] -+ exglobal_atmos_products.sh[290]exit 0 -+ JGLOBAL_ATMOS_PRODUCTS[27]true -+ JGLOBAL_ATMOS_PRODUCTS[28]export err=0 -+ JGLOBAL_ATMOS_PRODUCTS[28]err=0 -+ JGLOBAL_ATMOS_PRODUCTS[29][[ 0 -ne 0 ]] -+ JGLOBAL_ATMOS_PRODUCTS[40][[ -e OUTPUT.534058 ]] -+ JGLOBAL_ATMOS_PRODUCTS[47]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312 -+ JGLOBAL_ATMOS_PRODUCTS[48][[ NO == \N\O ]] -+ JGLOBAL_ATMOS_PRODUCTS[49]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f024.533601 -+ JGLOBAL_ATMOS_PRODUCTS[53]exit 0 -+ JGLOBAL_ATMOS_PRODUCTS[1]postamble /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_ATMOS_PRODUCTS 1753755861 0 -+ preamble.sh[62]set +x -End /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_ATMOS_PRODUCTS at 02:26:09 with error code 0 (time elapsed: 00:01:48) -+ status=0 -+ [[ 0 -ne 0 ]] -+ for FORECAST_HOUR in "${fhr_list[@]}" -++ printf %03d 25 -+ fhr3=025 -+ jobid=atmos_products_f025.533601 -+ /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_ATMOS_PRODUCTS -Begin /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_ATMOS_PRODUCTS at Tue Jul 29 02:26:09 UTC 2025 -++ jjob_header.sh[46]OPTIND=1 -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[50]env_job=atmos_products -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[49]read -ra configs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[61]shift 4 -++ jjob_header.sh[63][[ -z atmos_products ]] -++ jjob_header.sh[71]export DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601 -++ jjob_header.sh[71]DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601 -++ jjob_header.sh[72][[ YES == \Y\E\S ]] -++ jjob_header.sh[73]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601 -++ jjob_header.sh[75]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601 -++ jjob_header.sh[76]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601 -++ jjob_header.sh[85]export pid=548446 -++ jjob_header.sh[85]pid=548446 -++ jjob_header.sh[86]export pgmout=OUTPUT.548446 -++ jjob_header.sh[86]pgmout=OUTPUT.548446 -++ jjob_header.sh[87]export pgmerr=errfile -++ jjob_header.sh[87]pgmerr=errfile -++ jjob_header.sh[90]export pgm= -++ jjob_header.sh[90]pgm= -++ jjob_header.sh[96]export cycle=t12z -++ jjob_header.sh[96]cycle=t12z -++ jjob_header.sh[97]setpdy.sh -+ setpdy.sh[20]'[' /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601 == /home/mterry ']' -+ setpdy.sh[25][[ ! t12z =~ t??z ]] -+ setpdy.sh[30]case $# in -+ setpdy.sh[31]dates_before_PDY=7 -+ setpdy.sh[32]dates_after_PDY=7 -+ setpdy.sh[50]COMDATEROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+ setpdy.sh[53]'[' -z 20210323 ']' -+ setpdy.sh[57]sed 's/[0-9]\{8\}/20210323/' /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z -sed: can't read /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z: No such file or directory -++ jjob_header.sh[97]true -++ jjob_header.sh[98]source ./PDY -/work2/noaa/global/mterry/global-workflow_forked/ush/jjob_header.sh: line 98: ./PDY: No such file or directory -++ jjob_header.sh[98]true -++ jjob_header.sh[104]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[104]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.base -+++ config.base[6]echo 'BEGIN: config.base' -BEGIN: config.base -+++ config.base[9]export machine=HERCULES -+++ config.base[9]machine=HERCULES -+++ config.base[12]export RUN_ENVIR=emc -+++ config.base[12]RUN_ENVIR=emc -+++ config.base[15]export ACCOUNT=fv3-cpu -+++ config.base[15]ACCOUNT=fv3-cpu -+++ config.base[16]export QUEUE=batch -+++ config.base[16]QUEUE=batch -+++ config.base[17]export QUEUE_SERVICE=batch -+++ config.base[17]QUEUE_SERVICE=batch -+++ config.base[18]export QUEUE_DTN=batch -+++ config.base[18]QUEUE_DTN=batch -+++ config.base[19]export PARTITION_BATCH=hercules -+++ config.base[19]PARTITION_BATCH=hercules -+++ config.base[20]export PARTITION_SERVICE=service -+++ config.base[20]PARTITION_SERVICE=service -+++ config.base[21]export PARTITION_DTN= -+++ config.base[21]PARTITION_DTN= -+++ config.base[22]export RESERVATION= -+++ config.base[22]RESERVATION= -+++ config.base[23]export CLUSTERS= -+++ config.base[23]CLUSTERS= -+++ config.base[24]export CLUSTERS_SERVICE= -+++ config.base[24]CLUSTERS_SERVICE= -+++ config.base[25]export CLUSTERS_DTN= -+++ config.base[25]CLUSTERS_DTN= -+++ config.base[28]export HPSS_PROJECT=emc-global -+++ config.base[28]HPSS_PROJECT=emc-global -+++ config.base[31]export HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[31]HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[32]export EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[32]EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[33]export FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[33]FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[34]export PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[34]PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[35]export SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[35]SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[36]export USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[36]USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[38]export FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[38]FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[39]export FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[39]FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[40]export FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[40]FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[41]export FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[41]FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[42]export FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[42]FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[43]export FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[43]FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[44]export FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[44]FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[45]export FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[45]FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[50]export PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[50]PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[51]export COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[51]COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[52]export COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[52]COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[53]export DMPDIR=/work/noaa/rstprod/dump -+++ config.base[53]DMPDIR=/work/noaa/rstprod/dump -+++ config.base[57]export COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[57]COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[58]export COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[58]COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[59]export COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[59]COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[62]export HOMEDIR=/work2/noaa/global/mterry -+++ config.base[62]HOMEDIR=/work2/noaa/global/mterry -+++ config.base[63]export STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[63]STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]export PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[65]export NOSCRUB=/work2/noaa/global/mterry -+++ config.base[65]NOSCRUB=/work2/noaa/global/mterry -+++ config.base[68]export BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[68]BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[71]export BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[71]BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[74]export DO_PREP_SFC=NO -+++ config.base[74]DO_PREP_SFC=NO -+++ config.base[77]export DO_GOES=NO -+++ config.base[77]DO_GOES=NO -+++ config.base[78]export DO_BUFRSND=NO -+++ config.base[78]DO_BUFRSND=NO -+++ config.base[79]export DO_GEMPAK=NO -+++ config.base[79]DO_GEMPAK=NO -+++ config.base[80]export DO_AWIPS=NO -+++ config.base[80]DO_AWIPS=NO -+++ config.base[81]export DO_NPOESS=NO -+++ config.base[81]DO_NPOESS=NO -+++ config.base[82]export DO_TRACKER=YES -+++ config.base[82]DO_TRACKER=YES -+++ config.base[83]export DO_GENESIS=YES -+++ config.base[83]DO_GENESIS=YES -+++ config.base[84]export DO_GENESIS_FSU=NO -+++ config.base[84]DO_GENESIS_FSU=NO -+++ config.base[85]export DO_VERFOZN=YES -+++ config.base[85]DO_VERFOZN=YES -+++ config.base[86]export DO_VERFRAD=YES -+++ config.base[86]DO_VERFRAD=YES -+++ config.base[87]export DO_VMINMON=YES -+++ config.base[87]DO_VMINMON=YES -+++ config.base[88]export DO_ANLSTAT=NO -+++ config.base[88]DO_ANLSTAT=NO -+++ config.base[91]export MODE=forecast-only -+++ config.base[91]MODE=forecast-only -+++ config.base[92]export DO_TEST_MODE=YES -+++ config.base[92]DO_TEST_MODE=YES -+++ config.base[101]export FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[101]FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[102]export HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[102]HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[103]export HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v1.2.0 -+++ config.base[103]HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v1.2.0 -+++ config.base[106]export NMV=/bin/mv -+++ config.base[106]NMV=/bin/mv -+++ config.base[107]export 'NLN=/bin/ln -sf' -+++ config.base[107]NLN='/bin/ln -sf' -+++ config.base[108]export VERBOSE=YES -+++ config.base[108]VERBOSE=YES -+++ config.base[109]export KEEPDATA=NO -+++ config.base[109]KEEPDATA=NO -+++ config.base[110]export DEBUG_POSTSCRIPT=NO -+++ config.base[110]DEBUG_POSTSCRIPT=NO -+++ config.base[111]export CHGRP_RSTPROD=YES -+++ config.base[111]CHGRP_RSTPROD=YES -+++ config.base[112]export 'CHGRP_CMD=chgrp rstprod' -+++ config.base[112]CHGRP_CMD='chgrp rstprod' -+++ config.base[113]export NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[113]NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[114]export NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[114]NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[117]export BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[117]BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[120]export SDATE=2021032312 -+++ config.base[120]SDATE=2021032312 -+++ config.base[121]export EDATE=2021032312 -+++ config.base[121]EDATE=2021032312 -+++ config.base[122]export EXP_WARM_START=.false. -+++ config.base[122]EXP_WARM_START=.false. -+++ config.base[123]export assim_freq=6 -+++ config.base[123]assim_freq=6 -+++ config.base[124]export PSLOT=C48_S2SW -+++ config.base[124]PSLOT=C48_S2SW -+++ config.base[125]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[125]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[126]export ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[126]ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[127]export DUMP_SUFFIX= -+++ config.base[127]DUMP_SUFFIX= -+++ config.base[128][[ 2021032312 -ge 2019092100 ]] -+++ config.base[128][[ 2021032312 -le 2019110700 ]] -+++ config.base[131]export ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[131]ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[132]export ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[132]ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[133]export FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[133]FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[136]export envir=prod -+++ config.base[136]envir=prod -+++ config.base[137]export NET=gfs -+++ config.base[137]NET=gfs -+++ config.base[138]export RUN=gfs -+++ config.base[138]RUN=gfs -+++ config.base[141]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.com -++++ config.com[4]echo 'BEGIN: config.com' -BEGIN: config.com -++++ config.com[38][[ emc == \n\c\o ]] -++++ config.com[43]COM_OBSPROC_TMPL='${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos' -++++ config.com[44]COM_RTOFS_TMPL='${DMPDIR}' -++++ config.com[45]COM_TCVITAL_TMPL='${DMPDIR}/${RUN}.${YMD}/${HH}/atmos' -++++ config.com[47]declare -rx 'COM_OBS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/obs' -++++ config.com[48]declare -rx COM_OBSPROC_TMPL COM_RTOFS_TMPL -++++ config.com[50]COM_BASE='${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}' -++++ config.com[52]declare -rx 'COM_TOP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}' -++++ config.com[54]declare -rx 'COM_CONF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf' -++++ config.com[55]declare -rx 'COM_OBS_JEDI=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi' -++++ config.com[57]declare -rx 'COM_ATMOS_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input' -++++ config.com[58]declare -rx 'COM_ATMOS_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart' -++++ config.com[59]declare -rx 'COM_ATMOS_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos' -++++ config.com[60]declare -rx 'COM_SNOW_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow' -++++ config.com[61]declare -rx 'COM_SNOW_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon' -++++ config.com[62]declare -rx 'COM_ATMOS_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history' -++++ config.com[63]declare -rx 'COM_ATMOS_MASTER_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master' -++++ config.com[64]declare -rx 'COM_ATMOS_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2' -++++ config.com[65]declare -rx 'COM_ATMOS_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}' -++++ config.com[66]declare -rx 'COM_ATMOS_BUFR_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr' -++++ config.com[67]declare -rx 'COM_ATMOS_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}' -++++ config.com[68]declare -rx 'COM_ATMOS_GENESIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital' -++++ config.com[69]declare -rx 'COM_ATMOS_TRACK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks' -++++ config.com[70]declare -rx 'COM_ATMOS_GOES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim' -++++ config.com[71]declare -rx 'COM_ATMOS_IMAGERY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery' -++++ config.com[72]declare -rx 'COM_ATMOS_OZNMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon' -++++ config.com[73]declare -rx 'COM_ATMOS_RADMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon' -++++ config.com[74]declare -rx 'COM_ATMOS_MINMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon' -++++ config.com[75]declare -rx 'COM_ATMOS_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon' -++++ config.com[76]declare -rx 'COM_ATMOS_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo' -++++ config.com[78]declare -rx 'COM_WAVE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart' -++++ config.com[79]declare -rx 'COM_WAVE_PREP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep' -++++ config.com[80]declare -rx 'COM_WAVE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history' -++++ config.com[81]declare -rx 'COM_WAVE_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded' -++++ config.com[82]declare -rx 'COM_WAVE_GRID_RES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}' -++++ config.com[83]declare -rx 'COM_WAVE_STATION_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station' -++++ config.com[84]declare -rx 'COM_WAVE_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak' -++++ config.com[85]declare -rx 'COM_WAVE_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo' -++++ config.com[87]declare -rx 'COM_OCEAN_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history' -++++ config.com[88]declare -rx 'COM_OCEAN_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart' -++++ config.com[89]declare -rx 'COM_OCEAN_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input' -++++ config.com[90]declare -rx 'COM_OCEAN_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean' -++++ config.com[91]declare -rx 'COM_OCEAN_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon' -++++ config.com[92]declare -rx 'COM_OCEAN_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf' -++++ config.com[93]declare -rx 'COM_OCEAN_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean' -++++ config.com[94]declare -rx 'COM_OCEAN_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf' -++++ config.com[95]declare -rx 'COM_OCEAN_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2' -++++ config.com[96]declare -rx 'COM_OCEAN_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}' -++++ config.com[98]declare -rx 'COM_ICE_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice' -++++ config.com[99]declare -rx 'COM_ICE_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf' -++++ config.com[100]declare -rx 'COM_ICE_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon' -++++ config.com[101]declare -rx 'COM_ICE_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice' -++++ config.com[102]declare -rx 'COM_ICE_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input' -++++ config.com[103]declare -rx 'COM_ICE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history' -++++ config.com[104]declare -rx 'COM_ICE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart' -++++ config.com[105]declare -rx 'COM_ICE_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf' -++++ config.com[106]declare -rx 'COM_ICE_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2' -++++ config.com[107]declare -rx 'COM_ICE_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}' -++++ config.com[109]declare -rx 'COM_CHEM_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history' -++++ config.com[110]declare -rx 'COM_CHEM_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem' -++++ config.com[111]declare -rx 'COM_CHEM_BMAT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix' -++++ config.com[112]declare -rx 'COM_CHEM_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon' -++++ config.com[114]declare -rx 'COM_MED_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart' -+++ config.base[143]export LOGSCRIPT= -+++ config.base[143]LOGSCRIPT= -+++ config.base[145]export 'REDOUT=1>' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.atmos_products -+++ config.atmos_products[6]echo 'BEGIN: config.atmos_products' -BEGIN: config.atmos_products -+++ config.atmos_products[9]. /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources atmos_products -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=atmos_products -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[1019]walltime=00:15:00 -++++ config.resources[1020]ntasks=24 -++++ config.resources[1021]threads_per_task=1 -++++ config.resources[1022]tasks_per_node=24 -++++ config.resources[1023]export is_exclusive=True -++++ config.resources[1023]is_exclusive=True -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n '' ]] -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n '' ]] -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.atmos_products[12]export MAX_TASKS=25 -+++ config.atmos_products[12]MAX_TASKS=25 -+++ config.atmos_products[15]export INTERP_ATMOS_MASTERSH=/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh -+++ config.atmos_products[15]INTERP_ATMOS_MASTERSH=/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh -+++ config.atmos_products[16]export INTERP_ATMOS_SFLUXSH=/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_sflux.sh -+++ config.atmos_products[16]INTERP_ATMOS_SFLUXSH=/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_sflux.sh -+++ config.atmos_products[18][[ gfs == \g\d\a\s ]] -+++ config.atmos_products[24][[ gfs == \g\f\s ]] -+++ config.atmos_products[25]export downset=2 -+++ config.atmos_products[25]downset=2 -+++ config.atmos_products[26]export FHOUT_PGBS=3 -+++ config.atmos_products[26]FHOUT_PGBS=3 -+++ config.atmos_products[27]export FLXGF=YES -+++ config.atmos_products[27]FLXGF=YES -+++ config.atmos_products[28]export WGNE=YES -+++ config.atmos_products[28]WGNE=YES -+++ config.atmos_products[29]export FHMAX_WGNE=180 -+++ config.atmos_products[29]FHMAX_WGNE=180 -+++ config.atmos_products[32]export APCP_MSG=597 -+++ config.atmos_products[32]APCP_MSG=597 -+++ config.atmos_products[35]export paramlista=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.fFFF.paramlist.a.txt -+++ config.atmos_products[35]paramlista=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.fFFF.paramlist.a.txt -+++ config.atmos_products[36]export paramlista_anl=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.anl.paramlist.a.txt -+++ config.atmos_products[36]paramlista_anl=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.anl.paramlist.a.txt -+++ config.atmos_products[37]export paramlista_f000=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.f000.paramlist.a.txt -+++ config.atmos_products[37]paramlista_f000=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.f000.paramlist.a.txt -+++ config.atmos_products[38]export paramlistb=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.fFFF.paramlist.b.txt -+++ config.atmos_products[38]paramlistb=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.fFFF.paramlist.b.txt -+++ config.atmos_products[40]echo 'END: config.atmos_products' -END: config.atmos_products -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env atmos_products -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=atmos_products -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 520192000 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 24 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 24 ]] -+++ HERCULES.env[34]max_threads_per_task=3 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 3 ]] -+++ HERCULES.env[40][[ 1 -gt 3 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 24' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[281]export USE_CFP=YES -+++ HERCULES.env[281]USE_CFP=YES -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_ATMOS_PRODUCTS[11]YMD=20210323 -+ JGLOBAL_ATMOS_PRODUCTS[11]HH=12 -+ JGLOBAL_ATMOS_PRODUCTS[11]declare_from_tmpl -rx COMIN_ATMOS_ANALYSIS:COM_ATMOS_ANALYSIS_TMPL COMIN_ATMOS_HISTORY:COM_ATMOS_HISTORY_TMPL COMIN_ATMOS_MASTER:COM_ATMOS_MASTER_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_ATMOS_ANALYSIS=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//analysis/atmos -declare_from_tmpl :: COMIN_ATMOS_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/history -declare_from_tmpl :: COMIN_ATMOS_MASTER=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master -+ JGLOBAL_ATMOS_PRODUCTS[16]for grid in '0p25' '0p50' '1p00' -+ JGLOBAL_ATMOS_PRODUCTS[17]prod_dir=COMOUT_ATMOS_GRIB_0p25 -+ JGLOBAL_ATMOS_PRODUCTS[18]GRID=0p25 -+ JGLOBAL_ATMOS_PRODUCTS[18]YMD=20210323 -+ JGLOBAL_ATMOS_PRODUCTS[18]HH=12 -+ JGLOBAL_ATMOS_PRODUCTS[18]declare_from_tmpl -rx COMOUT_ATMOS_GRIB_0p25:COM_ATMOS_GRIB_GRID_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_ATMOS_GRIB_0p25=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25 -+ JGLOBAL_ATMOS_PRODUCTS[19][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25 ]] -+ JGLOBAL_ATMOS_PRODUCTS[16]for grid in '0p25' '0p50' '1p00' -+ JGLOBAL_ATMOS_PRODUCTS[17]prod_dir=COMOUT_ATMOS_GRIB_0p50 -+ JGLOBAL_ATMOS_PRODUCTS[18]GRID=0p50 -+ JGLOBAL_ATMOS_PRODUCTS[18]YMD=20210323 -+ JGLOBAL_ATMOS_PRODUCTS[18]HH=12 -+ JGLOBAL_ATMOS_PRODUCTS[18]declare_from_tmpl -rx COMOUT_ATMOS_GRIB_0p50:COM_ATMOS_GRIB_GRID_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_ATMOS_GRIB_0p50=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50 -+ JGLOBAL_ATMOS_PRODUCTS[19][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50 ]] -+ JGLOBAL_ATMOS_PRODUCTS[16]for grid in '0p25' '0p50' '1p00' -+ JGLOBAL_ATMOS_PRODUCTS[17]prod_dir=COMOUT_ATMOS_GRIB_1p00 -+ JGLOBAL_ATMOS_PRODUCTS[18]GRID=1p00 -+ JGLOBAL_ATMOS_PRODUCTS[18]YMD=20210323 -+ JGLOBAL_ATMOS_PRODUCTS[18]HH=12 -+ JGLOBAL_ATMOS_PRODUCTS[18]declare_from_tmpl -rx COMOUT_ATMOS_GRIB_1p00:COM_ATMOS_GRIB_GRID_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_ATMOS_GRIB_1p00=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00 -+ JGLOBAL_ATMOS_PRODUCTS[19][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00 ]] -+ JGLOBAL_ATMOS_PRODUCTS[23]export PREFIX=gfs.t12z. -+ JGLOBAL_ATMOS_PRODUCTS[23]PREFIX=gfs.t12z. -+ JGLOBAL_ATMOS_PRODUCTS[27]/work2/noaa/global/mterry/global-workflow_forked/scripts/exglobal_atmos_products.sh -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ exglobal_atmos_products.sh[4]INTERP_ATMOS_MASTERSH=/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh -+ exglobal_atmos_products.sh[5]INTERP_ATMOS_SFLUXSH=/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_sflux.sh -+ exglobal_atmos_products.sh[8]downset=2 -+ exglobal_atmos_products.sh[9]ntasks_atmos_products=8 -+ exglobal_atmos_products.sh[12]WGNE=YES -+ exglobal_atmos_products.sh[13]FHMAX_WGNE=180 -+ exglobal_atmos_products.sh[15]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601 -+ exglobal_atmos_products.sh[19][[ 25 -le 0 ]] -++ exglobal_atmos_products.sh[30]printf f%03d 25 -+ exglobal_atmos_products.sh[30]fhr3=f025 -+ exglobal_atmos_products.sh[31](( FORECAST_HOUR%FHOUT_PGBS == 0 )) -+ exglobal_atmos_products.sh[40]MASTER_FILE=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.master.grb2f025 -+ exglobal_atmos_products.sh[45]wgrib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.master.grb2f025 -+ exglobal_atmos_products.sh[45]wgrib2 -i -grib tmpfile_f025 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.master.grb2f025 -+ exglobal_atmos_products.sh[45]grep -F -f /work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.fFFF.paramlist.a.txt -1:0:d=2021032312:PRMSL:mean sea level:25 hour fcst: -2:25688:d=2021032312:CLMR:1 hybrid level:25 hour fcst: -3:29343:d=2021032312:ICMR:1 hybrid level:25 hour fcst: -4:36295:d=2021032312:RWMR:1 hybrid level:25 hour fcst: -5:48732:d=2021032312:SNMR:1 hybrid level:25 hour fcst: -6:58824:d=2021032312:GRLE:1 hybrid level:25 hour fcst: -7:61276:d=2021032312:REFD:1 hybrid level:25 hour fcst: -8:87031:d=2021032312:REFD:2 hybrid level:25 hour fcst: -9:112761:d=2021032312:REFC:entire atmosphere (considered as a single layer):25 hour fcst: -10:140221:d=2021032312:VIS:surface:25 hour fcst: -11:172542:d=2021032312:UGRD:planetary boundary layer:25 hour fcst: -12:186396:d=2021032312:VGRD:planetary boundary layer:25 hour fcst: -13:200985:d=2021032312:VRATE:planetary boundary layer:25 hour fcst: -14:216099:d=2021032312:GUST:surface:25 hour fcst: -15:230805:d=2021032312:HGT:0.01 mb:25 hour fcst: -16:254189:d=2021032312:TMP:0.01 mb:25 hour fcst: -17:264779:d=2021032312:RH:0.01 mb:25 hour fcst: -18:271064:d=2021032312:SPFH:0.01 mb:25 hour fcst: -19:294006:d=2021032312:VVEL:0.01 mb:25 hour fcst: -20:320813:d=2021032312:DZDT:0.01 mb:25 hour fcst: -21:348111:d=2021032312:UGRD:0.01 mb:25 hour fcst: -22:359231:d=2021032312:VGRD:0.01 mb:25 hour fcst: -23:369764:d=2021032312:ABSV:0.01 mb:25 hour fcst: -24:385189:d=2021032312:O3MR:0.01 mb:25 hour fcst: -25:399140:d=2021032312:HGT:0.02 mb:25 hour fcst: -26:424624:d=2021032312:TMP:0.02 mb:25 hour fcst: -27:439277:d=2021032312:RH:0.02 mb:25 hour fcst: -28:446600:d=2021032312:SPFH:0.02 mb:25 hour fcst: -29:470664:d=2021032312:VVEL:0.02 mb:25 hour fcst: -30:494846:d=2021032312:DZDT:0.02 mb:25 hour fcst: -31:522503:d=2021032312:UGRD:0.02 mb:25 hour fcst: -32:534851:d=2021032312:VGRD:0.02 mb:25 hour fcst: -33:546731:d=2021032312:ABSV:0.02 mb:25 hour fcst: -34:563548:d=2021032312:O3MR:0.02 mb:25 hour fcst: -35:579203:d=2021032312:HGT:0.04 mb:25 hour fcst: -36:603429:d=2021032312:TMP:0.04 mb:25 hour fcst: -37:620683:d=2021032312:RH:0.04 mb:25 hour fcst: -38:626661:d=2021032312:SPFH:0.04 mb:25 hour fcst: -39:652297:d=2021032312:VVEL:0.04 mb:25 hour fcst: -40:679819:d=2021032312:DZDT:0.04 mb:25 hour fcst: -41:704384:d=2021032312:UGRD:0.04 mb:25 hour fcst: -42:724315:d=2021032312:VGRD:0.04 mb:25 hour fcst: -43:744093:d=2021032312:ABSV:0.04 mb:25 hour fcst: -44:762514:d=2021032312:O3MR:0.04 mb:25 hour fcst: -45:786090:d=2021032312:HGT:0.07 mb:25 hour fcst: -46:810131:d=2021032312:TMP:0.07 mb:25 hour fcst: -47:827105:d=2021032312:RH:0.07 mb:25 hour fcst: -48:836678:d=2021032312:SPFH:0.07 mb:25 hour fcst: -49:860406:d=2021032312:VVEL:0.07 mb:25 hour fcst: -50:889126:d=2021032312:DZDT:0.07 mb:25 hour fcst: -51:914407:d=2021032312:UGRD:0.07 mb:25 hour fcst: -52:934816:d=2021032312:VGRD:0.07 mb:25 hour fcst: -53:955066:d=2021032312:ABSV:0.07 mb:25 hour fcst: -54:973823:d=2021032312:O3MR:0.07 mb:25 hour fcst: -55:993322:d=2021032312:HGT:0.1 mb:25 hour fcst: -56:1014866:d=2021032312:TMP:0.1 mb:25 hour fcst: -57:1031519:d=2021032312:RH:0.1 mb:25 hour fcst: -58:1039608:d=2021032312:SPFH:0.1 mb:25 hour fcst: -59:1060931:d=2021032312:VVEL:0.1 mb:25 hour fcst: -60:1090511:d=2021032312:DZDT:0.1 mb:25 hour fcst: -61:1116276:d=2021032312:UGRD:0.1 mb:25 hour fcst: -62:1136357:d=2021032312:VGRD:0.1 mb:25 hour fcst: -63:1156361:d=2021032312:ABSV:0.1 mb:25 hour fcst: -64:1174931:d=2021032312:O3MR:0.1 mb:25 hour fcst: -65:1198589:d=2021032312:HGT:0.2 mb:25 hour fcst: -66:1219583:d=2021032312:TMP:0.2 mb:25 hour fcst: -67:1235664:d=2021032312:RH:0.2 mb:25 hour fcst: -68:1246733:d=2021032312:SPFH:0.2 mb:25 hour fcst: -69:1266393:d=2021032312:VVEL:0.2 mb:25 hour fcst: -70:1295218:d=2021032312:DZDT:0.2 mb:25 hour fcst: -71:1322716:d=2021032312:UGRD:0.2 mb:25 hour fcst: -72:1342068:d=2021032312:VGRD:0.2 mb:25 hour fcst: -73:1360966:d=2021032312:ABSV:0.2 mb:25 hour fcst: -74:1378683:d=2021032312:O3MR:0.2 mb:25 hour fcst: -75:1404280:d=2021032312:HGT:0.4 mb:25 hour fcst: -76:1425336:d=2021032312:TMP:0.4 mb:25 hour fcst: -77:1441114:d=2021032312:RH:0.4 mb:25 hour fcst: -78:1450120:d=2021032312:SPFH:0.4 mb:25 hour fcst: -79:1470117:d=2021032312:VVEL:0.4 mb:25 hour fcst: -80:1494987:d=2021032312:DZDT:0.4 mb:25 hour fcst: -81:1525450:d=2021032312:UGRD:0.4 mb:25 hour fcst: -82:1544379:d=2021032312:VGRD:0.4 mb:25 hour fcst: -83:1562564:d=2021032312:ABSV:0.4 mb:25 hour fcst: -84:1579356:d=2021032312:O3MR:0.4 mb:25 hour fcst: -85:1604006:d=2021032312:HGT:0.7 mb:25 hour fcst: -86:1622145:d=2021032312:TMP:0.7 mb:25 hour fcst: -87:1637956:d=2021032312:RH:0.7 mb:25 hour fcst: -88:1646886:d=2021032312:SPFH:0.7 mb:25 hour fcst: -89:1666137:d=2021032312:VVEL:0.7 mb:25 hour fcst: -90:1692176:d=2021032312:DZDT:0.7 mb:25 hour fcst: -91:1721471:d=2021032312:UGRD:0.7 mb:25 hour fcst: -92:1740085:d=2021032312:VGRD:0.7 mb:25 hour fcst: -93:1757941:d=2021032312:ABSV:0.7 mb:25 hour fcst: -94:1774401:d=2021032312:O3MR:0.7 mb:25 hour fcst: -95:1794914:d=2021032312:HGT:1 mb:25 hour fcst: -96:1812911:d=2021032312:TMP:1 mb:25 hour fcst: -97:1828771:d=2021032312:RH:1 mb:25 hour fcst: -98:1838541:d=2021032312:SPFH:1 mb:25 hour fcst: -99:1857022:d=2021032312:VVEL:1 mb:25 hour fcst: -100:1883735:d=2021032312:DZDT:1 mb:25 hour fcst: -101:1909925:d=2021032312:UGRD:1 mb:25 hour fcst: -102:1928523:d=2021032312:VGRD:1 mb:25 hour fcst: -103:1946195:d=2021032312:ABSV:1 mb:25 hour fcst: -104:1962493:d=2021032312:O3MR:1 mb:25 hour fcst: -105:1983625:d=2021032312:HGT:2 mb:25 hour fcst: -106:2003828:d=2021032312:TMP:2 mb:25 hour fcst: -107:2020480:d=2021032312:RH:2 mb:25 hour fcst: -108:2029162:d=2021032312:SPFH:2 mb:25 hour fcst: -109:2048030:d=2021032312:VVEL:2 mb:25 hour fcst: -110:2076676:d=2021032312:DZDT:2 mb:25 hour fcst: -111:2103180:d=2021032312:UGRD:2 mb:25 hour fcst: -112:2114743:d=2021032312:VGRD:2 mb:25 hour fcst: -113:2133171:d=2021032312:ABSV:2 mb:25 hour fcst: -114:2150190:d=2021032312:O3MR:2 mb:25 hour fcst: -115:2172965:d=2021032312:HGT:3 mb:25 hour fcst: -116:2190765:d=2021032312:TMP:3 mb:25 hour fcst: -117:2207253:d=2021032312:RH:3 mb:25 hour fcst: -118:2213243:d=2021032312:SPFH:3 mb:25 hour fcst: -119:2234371:d=2021032312:VVEL:3 mb:25 hour fcst: -120:2264412:d=2021032312:DZDT:3 mb:25 hour fcst: -121:2290617:d=2021032312:UGRD:3 mb:25 hour fcst: -122:2301983:d=2021032312:VGRD:3 mb:25 hour fcst: -123:2319975:d=2021032312:ABSV:3 mb:25 hour fcst: -124:2336893:d=2021032312:O3MR:3 mb:25 hour fcst: -125:2360195:d=2021032312:HGT:5 mb:25 hour fcst: -126:2377852:d=2021032312:TMP:5 mb:25 hour fcst: -127:2394543:d=2021032312:RH:5 mb:25 hour fcst: -128:2403733:d=2021032312:SPFH:5 mb:25 hour fcst: -129:2424446:d=2021032312:VVEL:5 mb:25 hour fcst: -130:2453841:d=2021032312:DZDT:5 mb:25 hour fcst: -131:2479891:d=2021032312:UGRD:5 mb:25 hour fcst: -132:2491488:d=2021032312:VGRD:5 mb:25 hour fcst: -133:2510096:d=2021032312:ABSV:5 mb:25 hour fcst: -134:2527781:d=2021032312:O3MR:5 mb:25 hour fcst: -135:2550552:d=2021032312:HGT:7 mb:25 hour fcst: -136:2568217:d=2021032312:TMP:7 mb:25 hour fcst: -137:2584301:d=2021032312:RH:7 mb:25 hour fcst: -138:2590351:d=2021032312:SPFH:7 mb:25 hour fcst: -139:2609606:d=2021032312:VVEL:7 mb:25 hour fcst: -140:2634228:d=2021032312:DZDT:7 mb:25 hour fcst: -141:2659864:d=2021032312:UGRD:7 mb:25 hour fcst: -142:2671281:d=2021032312:VGRD:7 mb:25 hour fcst: -143:2690034:d=2021032312:ABSV:7 mb:25 hour fcst: -144:2707769:d=2021032312:O3MR:7 mb:25 hour fcst: -145:2731436:d=2021032312:HGT:10 mb:25 hour fcst: -146:2751273:d=2021032312:TMP:10 mb:25 hour fcst: -147:2767471:d=2021032312:RH:10 mb:25 hour fcst: -148:2775142:d=2021032312:SPFH:10 mb:25 hour fcst: -149:2793463:d=2021032312:VVEL:10 mb:25 hour fcst: -150:2819001:d=2021032312:DZDT:10 mb:25 hour fcst: -151:2844272:d=2021032312:UGRD:10 mb:25 hour fcst: -152:2855600:d=2021032312:VGRD:10 mb:25 hour fcst: -153:2874169:d=2021032312:ABSV:10 mb:25 hour fcst: -154:2891838:d=2021032312:O3MR:10 mb:25 hour fcst: -155:2910572:d=2021032312:HGT:15 mb:25 hour fcst: -156:2930102:d=2021032312:TMP:15 mb:25 hour fcst: -157:2946232:d=2021032312:RH:15 mb:25 hour fcst: -158:2951518:d=2021032312:SPFH:15 mb:25 hour fcst: -159:2966965:d=2021032312:VVEL:15 mb:25 hour fcst: -160:2993621:d=2021032312:DZDT:15 mb:25 hour fcst: -161:3024127:d=2021032312:UGRD:15 mb:25 hour fcst: -162:3035325:d=2021032312:VGRD:15 mb:25 hour fcst: -163:3053768:d=2021032312:ABSV:15 mb:25 hour fcst: -164:3071318:d=2021032312:O3MR:15 mb:25 hour fcst: -165:3090249:d=2021032312:HGT:20 mb:25 hour fcst: -166:3109661:d=2021032312:TMP:20 mb:25 hour fcst: -167:3125676:d=2021032312:RH:20 mb:25 hour fcst: -168:3131845:d=2021032312:SPFH:20 mb:25 hour fcst: -169:3146021:d=2021032312:VVEL:20 mb:25 hour fcst: -170:3173520:d=2021032312:DZDT:20 mb:25 hour fcst: -171:3203755:d=2021032312:UGRD:20 mb:25 hour fcst: -172:3222365:d=2021032312:VGRD:20 mb:25 hour fcst: -173:3240870:d=2021032312:ABSV:20 mb:25 hour fcst: -174:3258429:d=2021032312:O3MR:20 mb:25 hour fcst: -175:3283298:d=2021032312:HGT:30 mb:25 hour fcst: -176:3302405:d=2021032312:TMP:30 mb:25 hour fcst: -177:3318338:d=2021032312:RH:30 mb:25 hour fcst: -178:3325947:d=2021032312:SPFH:30 mb:25 hour fcst: -179:3341278:d=2021032312:VVEL:30 mb:25 hour fcst: -180:3369610:d=2021032312:DZDT:30 mb:25 hour fcst: -181:3399506:d=2021032312:UGRD:30 mb:25 hour fcst: -182:3418211:d=2021032312:VGRD:30 mb:25 hour fcst: -183:3436710:d=2021032312:ABSV:30 mb:25 hour fcst: -184:3454414:d=2021032312:O3MR:30 mb:25 hour fcst: -185:3481157:d=2021032312:HGT:40 mb:25 hour fcst: -186:3500141:d=2021032312:TMP:40 mb:25 hour fcst: -187:3516456:d=2021032312:RH:40 mb:25 hour fcst: -188:3525925:d=2021032312:SPFH:40 mb:25 hour fcst: -189:3542959:d=2021032312:VVEL:40 mb:25 hour fcst: -190:3572326:d=2021032312:DZDT:40 mb:25 hour fcst: -191:3602018:d=2021032312:UGRD:40 mb:25 hour fcst: -192:3620869:d=2021032312:VGRD:40 mb:25 hour fcst: -193:3639576:d=2021032312:ABSV:40 mb:25 hour fcst: -194:3657499:d=2021032312:O3MR:40 mb:25 hour fcst: -195:3684461:d=2021032312:HGT:50 mb:25 hour fcst: -196:3703406:d=2021032312:TMP:50 mb:25 hour fcst: -197:3719538:d=2021032312:RH:50 mb:25 hour fcst: -198:3730155:d=2021032312:TCDC:50 mb:25 hour fcst: -199:3730334:d=2021032312:SPFH:50 mb:25 hour fcst: -200:3748840:d=2021032312:VVEL:50 mb:25 hour fcst: -201:3779025:d=2021032312:DZDT:50 mb:25 hour fcst: -202:3811122:d=2021032312:UGRD:50 mb:25 hour fcst: -203:3829670:d=2021032312:VGRD:50 mb:25 hour fcst: -204:3848193:d=2021032312:ABSV:50 mb:25 hour fcst: -205:3865912:d=2021032312:CLMR:50 mb:25 hour fcst: -206:3866091:d=2021032312:ICMR:50 mb:25 hour fcst: -207:3866270:d=2021032312:RWMR:50 mb:25 hour fcst: -208:3866449:d=2021032312:SNMR:50 mb:25 hour fcst: -209:3866628:d=2021032312:GRLE:50 mb:25 hour fcst: -210:3866807:d=2021032312:O3MR:50 mb:25 hour fcst: -211:3893152:d=2021032312:HGT:70 mb:25 hour fcst: -212:3912162:d=2021032312:TMP:70 mb:25 hour fcst: -213:3928568:d=2021032312:RH:70 mb:25 hour fcst: -215:3937085:d=2021032312:SPFH:70 mb:25 hour fcst: -216:3957999:d=2021032312:VVEL:70 mb:25 hour fcst: -217:3989191:d=2021032312:DZDT:70 mb:25 hour fcst: -218:4018878:d=2021032312:UGRD:70 mb:25 hour fcst: -219:4037520:d=2021032312:VGRD:70 mb:25 hour fcst: -220:4056226:d=2021032312:ABSV:70 mb:25 hour fcst: -226:4074882:d=2021032312:O3MR:70 mb:25 hour fcst: -227:4100738:d=2021032312:HGT:100 mb:25 hour fcst: -228:4120344:d=2021032312:TMP:100 mb:25 hour fcst: -229:4137294:d=2021032312:RH:100 mb:25 hour fcst: -230:4147427:d=2021032312:TCDC:100 mb:25 hour fcst: -231:4147606:d=2021032312:SPFH:100 mb:25 hour fcst: -232:4173223:d=2021032312:VVEL:100 mb:25 hour fcst: -233:4198219:d=2021032312:DZDT:100 mb:25 hour fcst: -234:4228588:d=2021032312:UGRD:100 mb:25 hour fcst: -235:4247967:d=2021032312:VGRD:100 mb:25 hour fcst: -236:4267351:d=2021032312:ABSV:100 mb:25 hour fcst: -237:4285546:d=2021032312:CLMR:100 mb:25 hour fcst: -238:4285725:d=2021032312:ICMR:100 mb:25 hour fcst: -239:4286430:d=2021032312:RWMR:100 mb:25 hour fcst: -240:4286609:d=2021032312:SNMR:100 mb:25 hour fcst: -241:4288706:d=2021032312:GRLE:100 mb:25 hour fcst: -242:4288885:d=2021032312:O3MR:100 mb:25 hour fcst: -259:4529842:d=2021032312:HGT:150 mb:25 hour fcst: -260:4550185:d=2021032312:TMP:150 mb:25 hour fcst: -261:4567162:d=2021032312:RH:150 mb:25 hour fcst: -262:4578582:d=2021032312:TCDC:150 mb:25 hour fcst: -263:4582570:d=2021032312:SPFH:150 mb:25 hour fcst: -264:4606907:d=2021032312:VVEL:150 mb:25 hour fcst: -265:4634190:d=2021032312:DZDT:150 mb:25 hour fcst: -266:4659790:d=2021032312:UGRD:150 mb:25 hour fcst: -267:4679730:d=2021032312:VGRD:150 mb:25 hour fcst: -268:4699763:d=2021032312:ABSV:150 mb:25 hour fcst: -269:4718631:d=2021032312:CLMR:150 mb:25 hour fcst: -270:4718810:d=2021032312:ICMR:150 mb:25 hour fcst: -271:4724065:d=2021032312:RWMR:150 mb:25 hour fcst: -272:4724244:d=2021032312:SNMR:150 mb:25 hour fcst: -273:4727712:d=2021032312:GRLE:150 mb:25 hour fcst: -274:4728238:d=2021032312:O3MR:150 mb:25 hour fcst: -291:4972680:d=2021032312:HGT:200 mb:25 hour fcst: -292:4993730:d=2021032312:TMP:200 mb:25 hour fcst: -293:5011856:d=2021032312:RH:200 mb:25 hour fcst: -294:5026916:d=2021032312:TCDC:200 mb:25 hour fcst: -295:5033361:d=2021032312:SPFH:200 mb:25 hour fcst: -296:5059115:d=2021032312:VVEL:200 mb:25 hour fcst: -297:5085974:d=2021032312:DZDT:200 mb:25 hour fcst: -298:5112266:d=2021032312:UGRD:200 mb:25 hour fcst: -299:5125636:d=2021032312:VGRD:200 mb:25 hour fcst: -300:5139392:d=2021032312:ABSV:200 mb:25 hour fcst: -301:5159431:d=2021032312:CLMR:200 mb:25 hour fcst: -302:5159610:d=2021032312:ICMR:200 mb:25 hour fcst: -303:5168120:d=2021032312:RWMR:200 mb:25 hour fcst: -304:5168299:d=2021032312:SNMR:200 mb:25 hour fcst: -305:5173626:d=2021032312:GRLE:200 mb:25 hour fcst: -306:5173984:d=2021032312:O3MR:200 mb:25 hour fcst: -323:5430905:d=2021032312:HGT:250 mb:25 hour fcst: -324:5452688:d=2021032312:TMP:250 mb:25 hour fcst: -325:5471537:d=2021032312:RH:250 mb:25 hour fcst: -326:5489364:d=2021032312:TCDC:250 mb:25 hour fcst: -327:5499989:d=2021032312:SPFH:250 mb:25 hour fcst: -328:5528032:d=2021032312:VVEL:250 mb:25 hour fcst: -329:5556410:d=2021032312:DZDT:250 mb:25 hour fcst: -330:5583557:d=2021032312:UGRD:250 mb:25 hour fcst: -331:5597687:d=2021032312:VGRD:250 mb:25 hour fcst: -332:5612542:d=2021032312:ABSV:250 mb:25 hour fcst: -333:5634115:d=2021032312:CLMR:250 mb:25 hour fcst: -334:5634294:d=2021032312:ICMR:250 mb:25 hour fcst: -335:5648520:d=2021032312:RWMR:250 mb:25 hour fcst: -336:5648699:d=2021032312:SNMR:250 mb:25 hour fcst: -337:5657525:d=2021032312:GRLE:250 mb:25 hour fcst: -338:5657830:d=2021032312:O3MR:250 mb:25 hour fcst: -355:5934883:d=2021032312:HGT:300 mb:25 hour fcst: -356:5957176:d=2021032312:TMP:300 mb:25 hour fcst: -357:5975610:d=2021032312:RH:300 mb:25 hour fcst: -358:5994541:d=2021032312:TCDC:300 mb:25 hour fcst: -359:6007294:d=2021032312:SPFH:300 mb:25 hour fcst: -360:6034006:d=2021032312:VVEL:300 mb:25 hour fcst: -361:6057911:d=2021032312:DZDT:300 mb:25 hour fcst: -362:6086022:d=2021032312:UGRD:300 mb:25 hour fcst: -363:6100686:d=2021032312:VGRD:300 mb:25 hour fcst: -364:6116014:d=2021032312:ABSV:300 mb:25 hour fcst: -365:6138341:d=2021032312:CLMR:300 mb:25 hour fcst: -366:6138520:d=2021032312:ICMR:300 mb:25 hour fcst: -367:6154312:d=2021032312:RWMR:300 mb:25 hour fcst: -368:6154491:d=2021032312:SNMR:300 mb:25 hour fcst: -369:6166020:d=2021032312:GRLE:300 mb:25 hour fcst: -370:6166431:d=2021032312:O3MR:300 mb:25 hour fcst: -387:6454162:d=2021032312:HGT:350 mb:25 hour fcst: -388:6476106:d=2021032312:TMP:350 mb:25 hour fcst: -389:6494357:d=2021032312:RH:350 mb:25 hour fcst: -390:6513442:d=2021032312:TCDC:350 mb:25 hour fcst: -391:6526620:d=2021032312:SPFH:350 mb:25 hour fcst: -392:6554833:d=2021032312:VVEL:350 mb:25 hour fcst: -393:6579609:d=2021032312:DZDT:350 mb:25 hour fcst: -394:6607871:d=2021032312:UGRD:350 mb:25 hour fcst: -395:6622478:d=2021032312:VGRD:350 mb:25 hour fcst: -396:6637689:d=2021032312:ABSV:350 mb:25 hour fcst: -397:6659929:d=2021032312:CLMR:350 mb:25 hour fcst: -398:6660108:d=2021032312:ICMR:350 mb:25 hour fcst: -399:6677096:d=2021032312:RWMR:350 mb:25 hour fcst: -400:6677275:d=2021032312:SNMR:350 mb:25 hour fcst: -401:6689300:d=2021032312:GRLE:350 mb:25 hour fcst: -402:6689824:d=2021032312:O3MR:350 mb:25 hour fcst: -419:6975586:d=2021032312:HGT:400 mb:25 hour fcst: -420:6997474:d=2021032312:TMP:400 mb:25 hour fcst: -421:7015795:d=2021032312:RH:400 mb:25 hour fcst: -422:7034985:d=2021032312:TCDC:400 mb:25 hour fcst: -423:7048203:d=2021032312:SPFH:400 mb:25 hour fcst: -424:7075976:d=2021032312:VVEL:400 mb:25 hour fcst: -425:7100949:d=2021032312:DZDT:400 mb:25 hour fcst: -426:7129039:d=2021032312:UGRD:400 mb:25 hour fcst: -427:7151017:d=2021032312:VGRD:400 mb:25 hour fcst: -428:7173737:d=2021032312:ABSV:400 mb:25 hour fcst: -429:7195681:d=2021032312:CLMR:400 mb:25 hour fcst: -430:7196064:d=2021032312:ICMR:400 mb:25 hour fcst: -431:7213043:d=2021032312:RWMR:400 mb:25 hour fcst: -432:7213339:d=2021032312:SNMR:400 mb:25 hour fcst: -433:7227255:d=2021032312:GRLE:400 mb:25 hour fcst: -434:7228140:d=2021032312:O3MR:400 mb:25 hour fcst: -451:7531365:d=2021032312:HGT:450 mb:25 hour fcst: -452:7555317:d=2021032312:TMP:450 mb:25 hour fcst: -453:7573882:d=2021032312:RH:450 mb:25 hour fcst: -454:7593269:d=2021032312:TCDC:450 mb:25 hour fcst: -455:7606053:d=2021032312:SPFH:450 mb:25 hour fcst: -456:7635374:d=2021032312:VVEL:450 mb:25 hour fcst: -457:7660549:d=2021032312:DZDT:450 mb:25 hour fcst: -458:7688577:d=2021032312:UGRD:450 mb:25 hour fcst: -459:7710256:d=2021032312:VGRD:450 mb:25 hour fcst: -460:7732571:d=2021032312:ABSV:450 mb:25 hour fcst: -461:7753987:d=2021032312:CLMR:450 mb:25 hour fcst: -462:7754935:d=2021032312:ICMR:450 mb:25 hour fcst: -463:7771755:d=2021032312:RWMR:450 mb:25 hour fcst: -464:7772309:d=2021032312:SNMR:450 mb:25 hour fcst: -465:7785947:d=2021032312:GRLE:450 mb:25 hour fcst: -466:7787032:d=2021032312:O3MR:450 mb:25 hour fcst: -483:8088589:d=2021032312:HGT:500 mb:25 hour fcst: -484:8112309:d=2021032312:TMP:500 mb:25 hour fcst: -485:8130981:d=2021032312:RH:500 mb:25 hour fcst: -486:8150523:d=2021032312:TCDC:500 mb:25 hour fcst: -487:8162865:d=2021032312:SPFH:500 mb:25 hour fcst: -488:8191168:d=2021032312:VVEL:500 mb:25 hour fcst: -489:8216421:d=2021032312:DZDT:500 mb:25 hour fcst: -490:8244144:d=2021032312:UGRD:500 mb:25 hour fcst: -491:8265613:d=2021032312:VGRD:500 mb:25 hour fcst: -492:8287649:d=2021032312:ABSV:500 mb:25 hour fcst: -493:8308806:d=2021032312:CLMR:500 mb:25 hour fcst: -494:8310449:d=2021032312:ICMR:500 mb:25 hour fcst: -495:8326378:d=2021032312:RWMR:500 mb:25 hour fcst: -496:8327296:d=2021032312:SNMR:500 mb:25 hour fcst: -497:8341587:d=2021032312:GRLE:500 mb:25 hour fcst: -498:8343372:d=2021032312:O3MR:500 mb:25 hour fcst: -515:8646565:d=2021032312:HGT:550 mb:25 hour fcst: -516:8670137:d=2021032312:TMP:550 mb:25 hour fcst: -517:8688926:d=2021032312:RH:550 mb:25 hour fcst: -518:8708721:d=2021032312:TCDC:550 mb:25 hour fcst: -519:8720796:d=2021032312:SPFH:550 mb:25 hour fcst: -520:8750443:d=2021032312:VVEL:550 mb:25 hour fcst: -521:8775765:d=2021032312:DZDT:550 mb:25 hour fcst: -522:8803256:d=2021032312:UGRD:550 mb:25 hour fcst: -523:8824256:d=2021032312:VGRD:550 mb:25 hour fcst: -524:8846032:d=2021032312:ABSV:550 mb:25 hour fcst: -525:8866982:d=2021032312:CLMR:550 mb:25 hour fcst: -526:8870055:d=2021032312:ICMR:550 mb:25 hour fcst: -527:8886683:d=2021032312:RWMR:550 mb:25 hour fcst: -528:8888309:d=2021032312:SNMR:550 mb:25 hour fcst: -529:8902884:d=2021032312:GRLE:550 mb:25 hour fcst: -530:8905072:d=2021032312:O3MR:550 mb:25 hour fcst: -547:9210647:d=2021032312:HGT:600 mb:25 hour fcst: -548:9235057:d=2021032312:TMP:600 mb:25 hour fcst: -549:9254108:d=2021032312:RH:600 mb:25 hour fcst: -550:9274181:d=2021032312:TCDC:600 mb:25 hour fcst: -551:9286175:d=2021032312:SPFH:600 mb:25 hour fcst: -552:9316634:d=2021032312:VVEL:600 mb:25 hour fcst: -553:9342041:d=2021032312:DZDT:600 mb:25 hour fcst: -554:9369469:d=2021032312:UGRD:600 mb:25 hour fcst: -555:9390548:d=2021032312:VGRD:600 mb:25 hour fcst: -556:9412151:d=2021032312:ABSV:600 mb:25 hour fcst: -557:9432890:d=2021032312:CLMR:600 mb:25 hour fcst: -558:9437858:d=2021032312:ICMR:600 mb:25 hour fcst: -559:9451705:d=2021032312:RWMR:600 mb:25 hour fcst: -560:9455425:d=2021032312:SNMR:600 mb:25 hour fcst: -561:9471205:d=2021032312:GRLE:600 mb:25 hour fcst: -562:9473669:d=2021032312:O3MR:600 mb:25 hour fcst: -579:9789460:d=2021032312:HGT:650 mb:25 hour fcst: -580:9813603:d=2021032312:TMP:650 mb:25 hour fcst: -581:9832911:d=2021032312:RH:650 mb:25 hour fcst: -582:9853153:d=2021032312:TCDC:650 mb:25 hour fcst: -583:9864916:d=2021032312:SPFH:650 mb:25 hour fcst: -584:9893796:d=2021032312:VVEL:650 mb:25 hour fcst: -585:9919155:d=2021032312:DZDT:650 mb:25 hour fcst: -586:9946560:d=2021032312:UGRD:650 mb:25 hour fcst: -587:9967486:d=2021032312:VGRD:650 mb:25 hour fcst: -588:9989025:d=2021032312:ABSV:650 mb:25 hour fcst: -589:10009880:d=2021032312:CLMR:650 mb:25 hour fcst: -590:10015109:d=2021032312:ICMR:650 mb:25 hour fcst: -591:10027526:d=2021032312:RWMR:650 mb:25 hour fcst: -592:10032179:d=2021032312:SNMR:650 mb:25 hour fcst: -593:10047105:d=2021032312:GRLE:650 mb:25 hour fcst: -594:10049406:d=2021032312:O3MR:650 mb:25 hour fcst: -611:10366632:d=2021032312:HGT:700 mb:25 hour fcst: -612:10390786:d=2021032312:TMP:700 mb:25 hour fcst: -613:10410569:d=2021032312:RH:700 mb:25 hour fcst: -614:10430898:d=2021032312:TCDC:700 mb:25 hour fcst: -615:10442489:d=2021032312:SPFH:700 mb:25 hour fcst: -616:10471096:d=2021032312:VVEL:700 mb:25 hour fcst: -617:10496666:d=2021032312:DZDT:700 mb:25 hour fcst: -618:10524039:d=2021032312:UGRD:700 mb:25 hour fcst: -619:10544895:d=2021032312:VGRD:700 mb:25 hour fcst: -620:10566501:d=2021032312:ABSV:700 mb:25 hour fcst: -621:10587536:d=2021032312:CLMR:700 mb:25 hour fcst: -622:10594547:d=2021032312:ICMR:700 mb:25 hour fcst: -623:10605742:d=2021032312:RWMR:700 mb:25 hour fcst: -624:10611355:d=2021032312:SNMR:700 mb:25 hour fcst: -625:10625483:d=2021032312:GRLE:700 mb:25 hour fcst: -626:10627374:d=2021032312:O3MR:700 mb:25 hour fcst: -643:10944547:d=2021032312:HGT:750 mb:25 hour fcst: -644:10968981:d=2021032312:TMP:750 mb:25 hour fcst: -645:10988889:d=2021032312:RH:750 mb:25 hour fcst: -646:11009300:d=2021032312:TCDC:750 mb:25 hour fcst: -647:11021694:d=2021032312:SPFH:750 mb:25 hour fcst: -648:11050844:d=2021032312:VVEL:750 mb:25 hour fcst: -649:11076606:d=2021032312:DZDT:750 mb:25 hour fcst: -650:11103876:d=2021032312:UGRD:750 mb:25 hour fcst: -651:11124892:d=2021032312:VGRD:750 mb:25 hour fcst: -652:11146621:d=2021032312:ABSV:750 mb:25 hour fcst: -653:11167777:d=2021032312:CLMR:750 mb:25 hour fcst: -654:11176919:d=2021032312:ICMR:750 mb:25 hour fcst: -655:11188565:d=2021032312:RWMR:750 mb:25 hour fcst: -656:11194453:d=2021032312:SNMR:750 mb:25 hour fcst: -657:11208076:d=2021032312:GRLE:750 mb:25 hour fcst: -658:11210231:d=2021032312:O3MR:750 mb:25 hour fcst: -675:11531841:d=2021032312:HGT:800 mb:25 hour fcst: -676:11556371:d=2021032312:TMP:800 mb:25 hour fcst: -677:11576924:d=2021032312:RH:800 mb:25 hour fcst: -678:11597375:d=2021032312:TCDC:800 mb:25 hour fcst: -679:11610818:d=2021032312:SPFH:800 mb:25 hour fcst: -680:11640285:d=2021032312:VVEL:800 mb:25 hour fcst: -681:11665968:d=2021032312:DZDT:800 mb:25 hour fcst: -682:11693022:d=2021032312:UGRD:800 mb:25 hour fcst: -683:11714138:d=2021032312:VGRD:800 mb:25 hour fcst: -684:11735889:d=2021032312:ABSV:800 mb:25 hour fcst: -685:11757073:d=2021032312:CLMR:800 mb:25 hour fcst: -686:11768828:d=2021032312:ICMR:800 mb:25 hour fcst: -687:11779867:d=2021032312:RWMR:800 mb:25 hour fcst: -688:11788229:d=2021032312:SNMR:800 mb:25 hour fcst: -689:11800977:d=2021032312:GRLE:800 mb:25 hour fcst: -690:11803392:d=2021032312:O3MR:800 mb:25 hour fcst: -707:12131463:d=2021032312:HGT:850 mb:25 hour fcst: -708:12156213:d=2021032312:TMP:850 mb:25 hour fcst: -709:12177091:d=2021032312:RH:850 mb:25 hour fcst: -710:12197340:d=2021032312:TCDC:850 mb:25 hour fcst: -711:12211948:d=2021032312:SPFH:850 mb:25 hour fcst: -712:12241500:d=2021032312:VVEL:850 mb:25 hour fcst: -713:12266819:d=2021032312:DZDT:850 mb:25 hour fcst: -714:12293310:d=2021032312:UGRD:850 mb:25 hour fcst: -715:12314493:d=2021032312:VGRD:850 mb:25 hour fcst: -716:12336317:d=2021032312:ABSV:850 mb:25 hour fcst: -717:12357575:d=2021032312:CLMR:850 mb:25 hour fcst: -718:12371002:d=2021032312:ICMR:850 mb:25 hour fcst: -719:12380715:d=2021032312:RWMR:850 mb:25 hour fcst: -720:12391220:d=2021032312:SNMR:850 mb:25 hour fcst: -721:12402857:d=2021032312:GRLE:850 mb:25 hour fcst: -722:12405492:d=2021032312:O3MR:850 mb:25 hour fcst: -739:12731639:d=2021032312:HGT:900 mb:25 hour fcst: -740:12756666:d=2021032312:TMP:900 mb:25 hour fcst: -741:12777648:d=2021032312:RH:900 mb:25 hour fcst: -742:12797472:d=2021032312:TCDC:900 mb:25 hour fcst: -743:12810577:d=2021032312:SPFH:900 mb:25 hour fcst: -744:12840049:d=2021032312:VVEL:900 mb:25 hour fcst: -745:12864928:d=2021032312:DZDT:900 mb:25 hour fcst: -746:12891032:d=2021032312:UGRD:900 mb:25 hour fcst: -747:12912234:d=2021032312:VGRD:900 mb:25 hour fcst: -748:12934126:d=2021032312:ABSV:900 mb:25 hour fcst: -749:12955310:d=2021032312:CLMR:900 mb:25 hour fcst: -750:12965347:d=2021032312:ICMR:900 mb:25 hour fcst: -751:12973729:d=2021032312:RWMR:900 mb:25 hour fcst: -752:12985942:d=2021032312:SNMR:900 mb:25 hour fcst: -753:12996161:d=2021032312:GRLE:900 mb:25 hour fcst: -754:12998989:d=2021032312:O3MR:900 mb:25 hour fcst: -755:13025024:d=2021032312:HGT:925 mb:25 hour fcst: -756:13050180:d=2021032312:TMP:925 mb:25 hour fcst: -757:13071358:d=2021032312:RH:925 mb:25 hour fcst: -758:13091122:d=2021032312:TCDC:925 mb:25 hour fcst: -759:13103035:d=2021032312:SPFH:925 mb:25 hour fcst: -760:13132575:d=2021032312:VVEL:925 mb:25 hour fcst: -761:13157000:d=2021032312:DZDT:925 mb:25 hour fcst: -762:13182601:d=2021032312:UGRD:925 mb:25 hour fcst: -763:13203867:d=2021032312:VGRD:925 mb:25 hour fcst: -764:13225785:d=2021032312:ABSV:925 mb:25 hour fcst: -765:13247061:d=2021032312:CLMR:925 mb:25 hour fcst: -766:13255650:d=2021032312:ICMR:925 mb:25 hour fcst: -767:13263265:d=2021032312:RWMR:925 mb:25 hour fcst: -768:13275564:d=2021032312:SNMR:925 mb:25 hour fcst: -769:13285173:d=2021032312:GRLE:925 mb:25 hour fcst: -770:13287935:d=2021032312:O3MR:925 mb:25 hour fcst: -771:13313884:d=2021032312:HGT:950 mb:25 hour fcst: -772:13339195:d=2021032312:TMP:950 mb:25 hour fcst: -773:13360277:d=2021032312:RH:950 mb:25 hour fcst: -774:13379750:d=2021032312:TCDC:950 mb:25 hour fcst: -775:13389698:d=2021032312:SPFH:950 mb:25 hour fcst: -776:13416866:d=2021032312:VVEL:950 mb:25 hour fcst: -777:13440689:d=2021032312:DZDT:950 mb:25 hour fcst: -778:13465553:d=2021032312:UGRD:950 mb:25 hour fcst: -779:13486845:d=2021032312:VGRD:950 mb:25 hour fcst: -780:13508788:d=2021032312:ABSV:950 mb:25 hour fcst: -781:13529865:d=2021032312:CLMR:950 mb:25 hour fcst: -782:13536401:d=2021032312:ICMR:950 mb:25 hour fcst: -783:13543504:d=2021032312:RWMR:950 mb:25 hour fcst: -784:13555853:d=2021032312:SNMR:950 mb:25 hour fcst: -785:13564662:d=2021032312:GRLE:950 mb:25 hour fcst: -786:13567514:d=2021032312:O3MR:950 mb:25 hour fcst: -787:13593429:d=2021032312:HINDEX:surface:25 hour fcst: -788:13598526:d=2021032312:HGT:975 mb:25 hour fcst: -789:13623986:d=2021032312:TMP:975 mb:25 hour fcst: -790:13644893:d=2021032312:RH:975 mb:25 hour fcst: -791:13664012:d=2021032312:TCDC:975 mb:25 hour fcst: -792:13672151:d=2021032312:SPFH:975 mb:25 hour fcst: -793:13699225:d=2021032312:VVEL:975 mb:25 hour fcst: -794:13721921:d=2021032312:DZDT:975 mb:25 hour fcst: -795:13745434:d=2021032312:UGRD:975 mb:25 hour fcst: -796:13766794:d=2021032312:VGRD:975 mb:25 hour fcst: -797:13788667:d=2021032312:ABSV:975 mb:25 hour fcst: -798:13809649:d=2021032312:CLMR:975 mb:25 hour fcst: -799:13814309:d=2021032312:ICMR:975 mb:25 hour fcst: -800:13819644:d=2021032312:RWMR:975 mb:25 hour fcst: -801:13831343:d=2021032312:SNMR:975 mb:25 hour fcst: -802:13838387:d=2021032312:GRLE:975 mb:25 hour fcst: -803:13841131:d=2021032312:O3MR:975 mb:25 hour fcst: -804:13867005:d=2021032312:TMP:1000 mb:25 hour fcst: -805:13887777:d=2021032312:RH:1000 mb:25 hour fcst: -806:13906637:d=2021032312:TCDC:1000 mb:25 hour fcst: -807:13911098:d=2021032312:SPFH:1000 mb:25 hour fcst: -808:13938064:d=2021032312:VVEL:1000 mb:25 hour fcst: -809:13959187:d=2021032312:DZDT:1000 mb:25 hour fcst: -810:13980530:d=2021032312:UGRD:1000 mb:25 hour fcst: -811:14001320:d=2021032312:VGRD:1000 mb:25 hour fcst: -812:14022752:d=2021032312:ABSV:1000 mb:25 hour fcst: -813:14043325:d=2021032312:CLMR:1000 mb:25 hour fcst: -814:14045208:d=2021032312:ICMR:1000 mb:25 hour fcst: -815:14047671:d=2021032312:RWMR:1000 mb:25 hour fcst: -816:14057309:d=2021032312:SNMR:1000 mb:25 hour fcst: -817:14060741:d=2021032312:GRLE:1000 mb:25 hour fcst: -818:14061919:d=2021032312:O3MR:1000 mb:25 hour fcst: -819:14087784:d=2021032312:MSLET:mean sea level:25 hour fcst: -820:14114388:d=2021032312:HGT:1000 mb:25 hour fcst: -821:14140062:d=2021032312:SDEN:surface:25 hour fcst: -822:14151064:d=2021032312:REFD:4000 m above ground:25 hour fcst: -823:14164095:d=2021032312:REFD:1000 m above ground:25 hour fcst: -824:14188253:d=2021032312:PRES:surface:25 hour fcst: -825:14213861:d=2021032312:HGT:surface:25 hour fcst: -826:14229931:d=2021032312:TMP:surface:25 hour fcst: -827:14246574:d=2021032312:TSOIL:0-0.1 m below ground:25 hour fcst: -828:14258498:d=2021032312:SOILW:0-0.1 m below ground:25 hour fcst: -829:14267412:d=2021032312:SOILL:0-0.1 m below ground:25 hour fcst: -830:14276218:d=2021032312:TSOIL:0.1-0.4 m below ground:25 hour fcst: -831:14291694:d=2021032312:SOILW:0.1-0.4 m below ground:25 hour fcst: -832:14300710:d=2021032312:SOILL:0.1-0.4 m below ground:25 hour fcst: -833:14309604:d=2021032312:TSOIL:0.4-1 m below ground:25 hour fcst: -834:14324050:d=2021032312:SOILW:0.4-1 m below ground:25 hour fcst: -835:14333187:d=2021032312:SOILL:0.4-1 m below ground:25 hour fcst: -836:14342106:d=2021032312:TSOIL:1-2 m below ground:25 hour fcst: -837:14356392:d=2021032312:SOILW:1-2 m below ground:25 hour fcst: -838:14365455:d=2021032312:SOILL:1-2 m below ground:25 hour fcst: -839:14374533:d=2021032312:CNWAT:surface:25 hour fcst: -840:14380077:d=2021032312:WEASD:surface:25 hour fcst: -841:14392890:d=2021032312:SNOD:surface:25 hour fcst: -842:14407252:d=2021032312:PEVPR:surface:25 hour fcst: -843:14423020:d=2021032312:ICETK:surface:25 hour fcst: -845:14434458:d=2021032312:TMP:2 m above ground:25 hour fcst: -846:14455259:d=2021032312:SPFH:2 m above ground:25 hour fcst: -847:14484177:d=2021032312:DPT:2 m above ground:25 hour fcst: -848:14505443:d=2021032312:RH:2 m above ground:25 hour fcst: -849:14523975:d=2021032312:APTMP:2 m above ground:25 hour fcst: -850:14538373:d=2021032312:TMAX:2 m above ground:24-25 hour max fcst: -851:14559132:d=2021032312:TMIN:2 m above ground:24-25 hour min fcst: -852:14579952:d=2021032312:UGRD:10 m above ground:25 hour fcst: -853:14600463:d=2021032312:VGRD:10 m above ground:25 hour fcst: -854:14621361:d=2021032312:ICEG:10 m above mean sea level:25 hour fcst: -855:14621996:d=2021032312:CPOFP:surface:25 hour fcst: -856:14642369:d=2021032312:CPRAT:surface:25 hour fcst: -857:14663870:d=2021032312:PRATE:surface:25 hour fcst: -858:14690820:d=2021032312:CPRAT:surface:24-25 hour ave fcst: -859:14714624:d=2021032312:PRATE:surface:24-25 hour ave fcst: -860:14739939:d=2021032312:APCP:surface:24-25 hour acc fcst: -861:14746124:d=2021032312:APCP:surface:0-25 hour acc fcst: -862:14759398:d=2021032312:ACPCP:surface:24-25 hour acc fcst: -863:14764389:d=2021032312:ACPCP:surface:0-25 hour acc fcst: -866:14787708:d=2021032312:FROZR:surface:0-25 hour acc fcst: -867:14794059:d=2021032312:FRZR:surface:0-25 hour acc fcst: -868:14797201:d=2021032312:TSNOWP:surface:0-25 hour acc fcst: -869:14804809:d=2021032312:WATR:surface:24-25 hour acc fcst: -870:14817552:d=2021032312:FROZR:surface:24-25 hour acc fcst: -871:14820492:d=2021032312:FRZR:surface:24-25 hour acc fcst: -872:14821641:d=2021032312:TSNOWP:surface:24-25 hour acc fcst: -873:14827203:d=2021032312:CSNOW:surface:25 hour fcst: -874:14829003:d=2021032312:CICEP:surface:25 hour fcst: -875:14829182:d=2021032312:CFRZR:surface:25 hour fcst: -876:14829402:d=2021032312:CRAIN:surface:25 hour fcst: -877:14832908:d=2021032312:LHTFL:surface:24-25 hour ave fcst: -878:14854714:d=2021032312:SHTFL:surface:24-25 hour ave fcst: -879:14875710:d=2021032312:GFLUX:surface:24-25 hour ave fcst: -880:14887782:d=2021032312:UFLX:surface:24-25 hour ave fcst: -881:14904183:d=2021032312:VFLX:surface:24-25 hour ave fcst: -882:14920699:d=2021032312:SFCR:surface:25 hour fcst: -883:14939122:d=2021032312:FRICV:surface:25 hour fcst: -884:14960859:d=2021032312:U-GWD:surface:24-25 hour ave fcst: -885:14975731:d=2021032312:V-GWD:surface:24-25 hour ave fcst: -886:14990671:d=2021032312:VEG:surface:25 hour fcst: -887:14998705:d=2021032312:SOTYP:surface:25 hour fcst: -888:15009754:d=2021032312:WILT:surface:25 hour fcst: -889:15019664:d=2021032312:FLDCP:surface:25 hour fcst: -890:15029870:d=2021032312:SUNSD:surface:25 hour fcst: -891:15037033:d=2021032312:LFTX:surface:25 hour fcst: -892:15050860:d=2021032312:CAPE:surface:25 hour fcst: -893:15066040:d=2021032312:CIN:surface:25 hour fcst: -894:15078904:d=2021032312:PWAT:entire atmosphere (considered as a single layer):25 hour fcst: -895:15105845:d=2021032312:CWAT:entire atmosphere (considered as a single layer):25 hour fcst: -896:15132359:d=2021032312:RH:entire atmosphere (considered as a single layer):25 hour fcst: -897:15147477:d=2021032312:TOZNE:entire atmosphere (considered as a single layer):25 hour fcst: -898:15161816:d=2021032312:LCDC:low cloud layer:25 hour fcst: -899:15179696:d=2021032312:LCDC:low cloud layer:24-25 hour ave fcst: -900:15197513:d=2021032312:MCDC:middle cloud layer:25 hour fcst: -901:15213338:d=2021032312:MCDC:middle cloud layer:24-25 hour ave fcst: -902:15228803:d=2021032312:HCDC:high cloud layer:25 hour fcst: -903:15243606:d=2021032312:HCDC:high cloud layer:24-25 hour ave fcst: -904:15259449:d=2021032312:TCDC:entire atmosphere (considered as a single layer):25 hour fcst: -905:15276653:d=2021032312:TCDC:entire atmosphere (considered as a single layer):24-25 hour ave fcst: -906:15293883:d=2021032312:HGT:cloud ceiling:25 hour fcst: -907:15325976:d=2021032312:PRES:convective cloud bottom level:25 hour fcst: -908:15339634:d=2021032312:PRES:low cloud bottom level:24-25 hour ave fcst: -909:15360722:d=2021032312:PRES:middle cloud bottom level:24-25 hour ave fcst: -910:15377012:d=2021032312:PRES:high cloud bottom level:24-25 hour ave fcst: -912:15414266:d=2021032312:PRES:convective cloud top level:25 hour fcst: -913:15429629:d=2021032312:PRES:low cloud top level:24-25 hour ave fcst: -914:15450717:d=2021032312:PRES:middle cloud top level:24-25 hour ave fcst: -915:15466828:d=2021032312:PRES:high cloud top level:24-25 hour ave fcst: -916:15484123:d=2021032312:TMP:low cloud top level:24-25 hour ave fcst: -917:15500006:d=2021032312:TMP:middle cloud top level:24-25 hour ave fcst: -918:15513051:d=2021032312:TMP:high cloud top level:24-25 hour ave fcst: -919:15527419:d=2021032312:TCDC:convective cloud layer:25 hour fcst: -920:15548054:d=2021032312:TCDC:boundary layer cloud layer:24-25 hour ave fcst: -921:15563410:d=2021032312:CWORK:entire atmosphere (considered as a single layer):24-25 hour ave fcst: -922:15575959:d=2021032312:DSWRF:surface:24-25 hour ave fcst: -925:15623698:d=2021032312:DLWRF:surface:24-25 hour ave fcst: -926:15645409:d=2021032312:USWRF:surface:24-25 hour ave fcst: -927:15662701:d=2021032312:ULWRF:surface:24-25 hour ave fcst: -928:15681655:d=2021032312:USWRF:top of atmosphere:24-25 hour ave fcst: -929:15699739:d=2021032312:ULWRF:top of atmosphere:24-25 hour ave fcst: -930:15720885:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Total Aerosol:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -931:15748168:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Dust Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -932:15766287:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Sea Salt Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -933:15793657:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Sulphate Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -934:15819336:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Particulate Organic Matter Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -935:15844963:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Black Carbon Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -936:15867236:d=2021032312:HLCY:3000-0 m above ground:25 hour fcst: -937:15883100:d=2021032312:HLCY:1000-0 m above ground:25 hour fcst: -938:15905356:d=2021032312:USTM:6000-0 m above ground:25 hour fcst: -939:15928447:d=2021032312:VSTM:6000-0 m above ground:25 hour fcst: -940:15951279:d=2021032312:PRES:tropopause:25 hour fcst: -941:15981078:d=2021032312:ICAHT:tropopause:25 hour fcst: -942:16011907:d=2021032312:HGT:tropopause:25 hour fcst: -943:16040124:d=2021032312:TMP:tropopause:25 hour fcst: -944:16061935:d=2021032312:UGRD:tropopause:25 hour fcst: -945:16077924:d=2021032312:VGRD:tropopause:25 hour fcst: -946:16093818:d=2021032312:VWSH:tropopause:25 hour fcst: -947:16112326:d=2021032312:PRES:max wind:25 hour fcst: -948:16146814:d=2021032312:ICAHT:max wind:25 hour fcst: -949:16182016:d=2021032312:HGT:max wind:25 hour fcst: -950:16217367:d=2021032312:UGRD:max wind:25 hour fcst: -951:16233821:d=2021032312:VGRD:max wind:25 hour fcst: -952:16251269:d=2021032312:TMP:max wind:25 hour fcst: -953:16278018:d=2021032312:UGRD:20 m above ground:25 hour fcst: -954:16298568:d=2021032312:VGRD:20 m above ground:25 hour fcst: -955:16319528:d=2021032312:UGRD:30 m above ground:25 hour fcst: -956:16340418:d=2021032312:VGRD:30 m above ground:25 hour fcst: -957:16361738:d=2021032312:UGRD:40 m above ground:25 hour fcst: -958:16382735:d=2021032312:VGRD:40 m above ground:25 hour fcst: -959:16404130:d=2021032312:UGRD:50 m above ground:25 hour fcst: -960:16425249:d=2021032312:VGRD:50 m above ground:25 hour fcst: -961:16446757:d=2021032312:TMP:80 m above ground:25 hour fcst: -962:16467533:d=2021032312:SPFH:80 m above ground:25 hour fcst: -963:16496230:d=2021032312:PRES:80 m above ground:25 hour fcst: -964:16521812:d=2021032312:UGRD:80 m above ground:25 hour fcst: -965:16543083:d=2021032312:VGRD:80 m above ground:25 hour fcst: -966:16564787:d=2021032312:TMP:100 m above ground:25 hour fcst: -967:16585539:d=2021032312:UGRD:100 m above ground:25 hour fcst: -968:16606881:d=2021032312:VGRD:100 m above ground:25 hour fcst: -981:16872311:d=2021032312:TMP:1829 m above mean sea level:25 hour fcst: -982:16893957:d=2021032312:UGRD:1829 m above mean sea level:25 hour fcst: -983:16916203:d=2021032312:VGRD:1829 m above mean sea level:25 hour fcst: -984:16939103:d=2021032312:TMP:2743 m above mean sea level:25 hour fcst: -985:16960750:d=2021032312:UGRD:2743 m above mean sea level:25 hour fcst: -986:16983411:d=2021032312:VGRD:2743 m above mean sea level:25 hour fcst: -987:17006725:d=2021032312:TMP:3658 m above mean sea level:25 hour fcst: -988:17028015:d=2021032312:UGRD:3658 m above mean sea level:25 hour fcst: -989:17051060:d=2021032312:VGRD:3658 m above mean sea level:25 hour fcst: -993:17143174:d=2021032312:HGT:0C isotherm:25 hour fcst: -994:17164818:d=2021032312:RH:0C isotherm:25 hour fcst: -995:17184056:d=2021032312:HGT:highest tropospheric freezing level:25 hour fcst: -996:17205544:d=2021032312:RH:highest tropospheric freezing level:25 hour fcst: -997:17224554:d=2021032312:TMP:30-0 mb above ground:25 hour fcst: -998:17245372:d=2021032312:RH:30-0 mb above ground:25 hour fcst: -1000:17284735:d=2021032312:SPFH:30-0 mb above ground:25 hour fcst: -1001:17313560:d=2021032312:UGRD:30-0 mb above ground:25 hour fcst: -1002:17334722:d=2021032312:VGRD:30-0 mb above ground:25 hour fcst: -1030:17959881:d=2021032312:4LFTX:surface:25 hour fcst: -1031:17973429:d=2021032312:CAPE:180-0 mb above ground:25 hour fcst: -1032:17987407:d=2021032312:CIN:180-0 mb above ground:25 hour fcst: -1033:18001447:d=2021032312:HPBL:surface:25 hour fcst: -1034:18032478:d=2021032312:RH:0.33-1 sigma layer:25 hour fcst: -1035:18050850:d=2021032312:RH:0.44-1 sigma layer:25 hour fcst: -1036:18068847:d=2021032312:RH:0.72-0.94 sigma layer:25 hour fcst: -1037:18087288:d=2021032312:RH:0.44-0.72 sigma layer:25 hour fcst: -1038:18106623:d=2021032312:TMP:0.995 sigma level:25 hour fcst: -1039:18127459:d=2021032312:POT:0.995 sigma level:25 hour fcst: -1040:18148287:d=2021032312:RH:0.995 sigma level:25 hour fcst: -1041:18166076:d=2021032312:UGRD:0.995 sigma level:25 hour fcst: -1042:18187268:d=2021032312:VGRD:0.995 sigma level:25 hour fcst: -1043:18208794:d=2021032312:VVEL:0.995 sigma level:25 hour fcst: -1044:18229512:d=2021032312:CAPE:90-0 mb above ground:25 hour fcst: -1045:18242296:d=2021032312:CIN:90-0 mb above ground:25 hour fcst: -1046:18256384:d=2021032312:CAPE:255-0 mb above ground:25 hour fcst: -1047:18269731:d=2021032312:CIN:255-0 mb above ground:25 hour fcst: -1048:18281126:d=2021032312:PLPL:255-0 mb above ground:25 hour fcst: -1049:18305078:d=2021032312:LAND:surface:25 hour fcst: -1050:18307019:d=2021032312:ICEC:surface:25 hour fcst: -1051:18310188:d=2021032312:ALBDO:surface:24-25 hour ave fcst: -1052:18324672:d=2021032312:ICETMP:surface:25 hour fcst: -1094:18937440:d=2021032312:UGRD:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -1095:18948167:d=2021032312:VGRD:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -1096:18958883:d=2021032312:TMP:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -1097:18973228:d=2021032312:HGT:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -1098:18990587:d=2021032312:PRES:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -1099:19008956:d=2021032312:VWSH:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -1100:19020875:d=2021032312:UGRD:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -1101:19031753:d=2021032312:VGRD:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -1102:19046534:d=2021032312:TMP:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -1103:19060946:d=2021032312:HGT:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -1104:19078507:d=2021032312:PRES:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -1105:19096811:d=2021032312:VWSH:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -+ exglobal_atmos_products.sh[45]true -+ exglobal_atmos_products.sh[46]export err=0 -+ exglobal_atmos_products.sh[46]err=0 -+ exglobal_atmos_products.sh[47][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[52][[ 2 -eq 2 ]] -+ exglobal_atmos_products.sh[54]wgrib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.master.grb2f025 -+ exglobal_atmos_products.sh[54]wgrib2 -i -grib tmpfileb_f025 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.master.grb2f025 -+ exglobal_atmos_products.sh[54]grep -F -f /work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.fFFF.paramlist.b.txt -95:1794914:d=2021032312:HGT:1 mb:25 hour fcst: -96:1812911:d=2021032312:TMP:1 mb:25 hour fcst: -97:1828771:d=2021032312:RH:1 mb:25 hour fcst: -101:1909925:d=2021032312:UGRD:1 mb:25 hour fcst: -102:1928523:d=2021032312:VGRD:1 mb:25 hour fcst: -103:1946195:d=2021032312:ABSV:1 mb:25 hour fcst: -104:1962493:d=2021032312:O3MR:1 mb:25 hour fcst: -105:1983625:d=2021032312:HGT:2 mb:25 hour fcst: -106:2003828:d=2021032312:TMP:2 mb:25 hour fcst: -107:2020480:d=2021032312:RH:2 mb:25 hour fcst: -111:2103180:d=2021032312:UGRD:2 mb:25 hour fcst: -112:2114743:d=2021032312:VGRD:2 mb:25 hour fcst: -113:2133171:d=2021032312:ABSV:2 mb:25 hour fcst: -114:2150190:d=2021032312:O3MR:2 mb:25 hour fcst: -115:2172965:d=2021032312:HGT:3 mb:25 hour fcst: -116:2190765:d=2021032312:TMP:3 mb:25 hour fcst: -117:2207253:d=2021032312:RH:3 mb:25 hour fcst: -121:2290617:d=2021032312:UGRD:3 mb:25 hour fcst: -122:2301983:d=2021032312:VGRD:3 mb:25 hour fcst: -123:2319975:d=2021032312:ABSV:3 mb:25 hour fcst: -124:2336893:d=2021032312:O3MR:3 mb:25 hour fcst: -125:2360195:d=2021032312:HGT:5 mb:25 hour fcst: -126:2377852:d=2021032312:TMP:5 mb:25 hour fcst: -127:2394543:d=2021032312:RH:5 mb:25 hour fcst: -131:2479891:d=2021032312:UGRD:5 mb:25 hour fcst: -132:2491488:d=2021032312:VGRD:5 mb:25 hour fcst: -133:2510096:d=2021032312:ABSV:5 mb:25 hour fcst: -134:2527781:d=2021032312:O3MR:5 mb:25 hour fcst: -135:2550552:d=2021032312:HGT:7 mb:25 hour fcst: -136:2568217:d=2021032312:TMP:7 mb:25 hour fcst: -137:2584301:d=2021032312:RH:7 mb:25 hour fcst: -141:2659864:d=2021032312:UGRD:7 mb:25 hour fcst: -142:2671281:d=2021032312:VGRD:7 mb:25 hour fcst: -143:2690034:d=2021032312:ABSV:7 mb:25 hour fcst: -144:2707769:d=2021032312:O3MR:7 mb:25 hour fcst: -214:3936906:d=2021032312:TCDC:70 mb:25 hour fcst: -221:4073987:d=2021032312:CLMR:70 mb:25 hour fcst: -222:4074166:d=2021032312:ICMR:70 mb:25 hour fcst: -223:4074345:d=2021032312:RWMR:70 mb:25 hour fcst: -224:4074524:d=2021032312:SNMR:70 mb:25 hour fcst: -225:4074703:d=2021032312:GRLE:70 mb:25 hour fcst: -243:4312898:d=2021032312:HGT:125 mb:25 hour fcst: -244:4332973:d=2021032312:TMP:125 mb:25 hour fcst: -245:4349982:d=2021032312:RH:125 mb:25 hour fcst: -246:4360492:d=2021032312:TCDC:125 mb:25 hour fcst: -248:4389073:d=2021032312:VVEL:125 mb:25 hour fcst: -249:4415268:d=2021032312:DZDT:125 mb:25 hour fcst: -250:4440360:d=2021032312:UGRD:125 mb:25 hour fcst: -251:4459977:d=2021032312:VGRD:125 mb:25 hour fcst: -252:4479606:d=2021032312:ABSV:125 mb:25 hour fcst: -253:4498113:d=2021032312:CLMR:125 mb:25 hour fcst: -254:4498292:d=2021032312:ICMR:125 mb:25 hour fcst: -255:4502785:d=2021032312:RWMR:125 mb:25 hour fcst: -256:4502964:d=2021032312:SNMR:125 mb:25 hour fcst: -257:4506781:d=2021032312:GRLE:125 mb:25 hour fcst: -275:4750119:d=2021032312:HGT:175 mb:25 hour fcst: -276:4770749:d=2021032312:TMP:175 mb:25 hour fcst: -277:4788231:d=2021032312:RH:175 mb:25 hour fcst: -278:4801363:d=2021032312:TCDC:175 mb:25 hour fcst: -280:4832701:d=2021032312:VVEL:175 mb:25 hour fcst: -281:4860903:d=2021032312:DZDT:175 mb:25 hour fcst: -282:4886963:d=2021032312:UGRD:175 mb:25 hour fcst: -283:4899962:d=2021032312:VGRD:175 mb:25 hour fcst: -284:4920704:d=2021032312:ABSV:175 mb:25 hour fcst: -285:4940239:d=2021032312:CLMR:175 mb:25 hour fcst: -286:4940418:d=2021032312:ICMR:175 mb:25 hour fcst: -287:4947048:d=2021032312:RWMR:175 mb:25 hour fcst: -288:4947227:d=2021032312:SNMR:175 mb:25 hour fcst: -289:4950897:d=2021032312:GRLE:175 mb:25 hour fcst: -307:5194590:d=2021032312:HGT:225 mb:25 hour fcst: -308:5216023:d=2021032312:TMP:225 mb:25 hour fcst: -309:5234622:d=2021032312:RH:225 mb:25 hour fcst: -310:5251269:d=2021032312:TCDC:225 mb:25 hour fcst: -312:5288124:d=2021032312:VVEL:225 mb:25 hour fcst: -313:5315735:d=2021032312:DZDT:225 mb:25 hour fcst: -314:5342485:d=2021032312:UGRD:225 mb:25 hour fcst: -315:5356447:d=2021032312:VGRD:225 mb:25 hour fcst: -316:5370820:d=2021032312:ABSV:225 mb:25 hour fcst: -317:5391750:d=2021032312:CLMR:225 mb:25 hour fcst: -318:5391929:d=2021032312:ICMR:225 mb:25 hour fcst: -319:5403437:d=2021032312:RWMR:225 mb:25 hour fcst: -320:5403616:d=2021032312:SNMR:225 mb:25 hour fcst: -321:5410430:d=2021032312:GRLE:225 mb:25 hour fcst: -339:5683439:d=2021032312:HGT:275 mb:25 hour fcst: -340:5705625:d=2021032312:TMP:275 mb:25 hour fcst: -341:5724462:d=2021032312:RH:275 mb:25 hour fcst: -342:5742901:d=2021032312:TCDC:275 mb:25 hour fcst: -344:5782229:d=2021032312:VVEL:275 mb:25 hour fcst: -345:5805541:d=2021032312:DZDT:275 mb:25 hour fcst: -346:5833174:d=2021032312:UGRD:275 mb:25 hour fcst: -347:5847489:d=2021032312:VGRD:275 mb:25 hour fcst: -348:5862661:d=2021032312:ABSV:275 mb:25 hour fcst: -349:5884715:d=2021032312:CLMR:275 mb:25 hour fcst: -350:5884894:d=2021032312:ICMR:275 mb:25 hour fcst: -351:5899414:d=2021032312:RWMR:275 mb:25 hour fcst: -352:5899593:d=2021032312:SNMR:275 mb:25 hour fcst: -353:5909542:d=2021032312:GRLE:275 mb:25 hour fcst: -371:6193576:d=2021032312:HGT:325 mb:25 hour fcst: -372:6215760:d=2021032312:TMP:325 mb:25 hour fcst: -373:6234114:d=2021032312:RH:325 mb:25 hour fcst: -374:6253075:d=2021032312:TCDC:325 mb:25 hour fcst: -376:6293408:d=2021032312:VVEL:325 mb:25 hour fcst: -377:6317692:d=2021032312:DZDT:325 mb:25 hour fcst: -378:6346035:d=2021032312:UGRD:325 mb:25 hour fcst: -379:6360712:d=2021032312:VGRD:325 mb:25 hour fcst: -380:6376043:d=2021032312:ABSV:325 mb:25 hour fcst: -381:6398357:d=2021032312:CLMR:325 mb:25 hour fcst: -382:6398536:d=2021032312:ICMR:325 mb:25 hour fcst: -383:6415300:d=2021032312:RWMR:325 mb:25 hour fcst: -384:6415479:d=2021032312:SNMR:325 mb:25 hour fcst: -385:6427449:d=2021032312:GRLE:325 mb:25 hour fcst: -403:6715382:d=2021032312:HGT:375 mb:25 hour fcst: -404:6737230:d=2021032312:TMP:375 mb:25 hour fcst: -405:6755562:d=2021032312:RH:375 mb:25 hour fcst: -406:6774681:d=2021032312:TCDC:375 mb:25 hour fcst: -408:6814772:d=2021032312:VVEL:375 mb:25 hour fcst: -409:6839603:d=2021032312:DZDT:375 mb:25 hour fcst: -410:6867674:d=2021032312:UGRD:375 mb:25 hour fcst: -411:6882121:d=2021032312:VGRD:375 mb:25 hour fcst: -412:6897229:d=2021032312:ABSV:375 mb:25 hour fcst: -413:6919395:d=2021032312:CLMR:375 mb:25 hour fcst: -414:6919574:d=2021032312:ICMR:375 mb:25 hour fcst: -415:6936674:d=2021032312:RWMR:375 mb:25 hour fcst: -416:6936853:d=2021032312:SNMR:375 mb:25 hour fcst: -417:6950021:d=2021032312:GRLE:375 mb:25 hour fcst: -435:7252706:d=2021032312:HGT:425 mb:25 hour fcst: -436:7276715:d=2021032312:TMP:425 mb:25 hour fcst: -437:7295236:d=2021032312:RH:425 mb:25 hour fcst: -438:7314538:d=2021032312:TCDC:425 mb:25 hour fcst: -440:7356087:d=2021032312:VVEL:425 mb:25 hour fcst: -441:7380982:d=2021032312:DZDT:425 mb:25 hour fcst: -442:7409042:d=2021032312:UGRD:425 mb:25 hour fcst: -443:7430879:d=2021032312:VGRD:425 mb:25 hour fcst: -444:7453336:d=2021032312:ABSV:425 mb:25 hour fcst: -445:7474950:d=2021032312:CLMR:425 mb:25 hour fcst: -446:7475588:d=2021032312:ICMR:425 mb:25 hour fcst: -447:7492503:d=2021032312:RWMR:425 mb:25 hour fcst: -448:7492902:d=2021032312:SNMR:425 mb:25 hour fcst: -449:7506006:d=2021032312:GRLE:425 mb:25 hour fcst: -467:7811081:d=2021032312:HGT:475 mb:25 hour fcst: -468:7834969:d=2021032312:TMP:475 mb:25 hour fcst: -469:7853614:d=2021032312:RH:475 mb:25 hour fcst: -470:7873120:d=2021032312:TCDC:475 mb:25 hour fcst: -472:7913259:d=2021032312:VVEL:475 mb:25 hour fcst: -473:7938276:d=2021032312:DZDT:475 mb:25 hour fcst: -474:7966103:d=2021032312:UGRD:475 mb:25 hour fcst: -475:7987669:d=2021032312:VGRD:475 mb:25 hour fcst: -476:8009868:d=2021032312:ABSV:475 mb:25 hour fcst: -477:8031133:d=2021032312:CLMR:475 mb:25 hour fcst: -478:8032392:d=2021032312:ICMR:475 mb:25 hour fcst: -479:8048678:d=2021032312:RWMR:475 mb:25 hour fcst: -480:8049392:d=2021032312:SNMR:475 mb:25 hour fcst: -481:8063417:d=2021032312:GRLE:475 mb:25 hour fcst: -499:8366775:d=2021032312:HGT:525 mb:25 hour fcst: -500:8390411:d=2021032312:TMP:525 mb:25 hour fcst: -501:8409232:d=2021032312:RH:525 mb:25 hour fcst: -502:8428909:d=2021032312:TCDC:525 mb:25 hour fcst: -504:8469974:d=2021032312:VVEL:525 mb:25 hour fcst: -505:8495065:d=2021032312:DZDT:525 mb:25 hour fcst: -506:8522662:d=2021032312:UGRD:525 mb:25 hour fcst: -507:8543803:d=2021032312:VGRD:525 mb:25 hour fcst: -508:8565699:d=2021032312:ABSV:525 mb:25 hour fcst: -509:8586711:d=2021032312:CLMR:525 mb:25 hour fcst: -510:8588797:d=2021032312:ICMR:525 mb:25 hour fcst: -511:8605824:d=2021032312:RWMR:525 mb:25 hour fcst: -512:8606926:d=2021032312:SNMR:525 mb:25 hour fcst: -513:8621365:d=2021032312:GRLE:525 mb:25 hour fcst: -531:8927931:d=2021032312:HGT:575 mb:25 hour fcst: -532:8951413:d=2021032312:TMP:575 mb:25 hour fcst: -533:8970291:d=2021032312:RH:575 mb:25 hour fcst: -534:8990209:d=2021032312:TCDC:575 mb:25 hour fcst: -536:9032373:d=2021032312:VVEL:575 mb:25 hour fcst: -537:9057597:d=2021032312:DZDT:575 mb:25 hour fcst: -538:9085016:d=2021032312:UGRD:575 mb:25 hour fcst: -539:9105916:d=2021032312:VGRD:575 mb:25 hour fcst: -540:9127593:d=2021032312:ABSV:575 mb:25 hour fcst: -541:9148449:d=2021032312:CLMR:575 mb:25 hour fcst: -542:9152630:d=2021032312:ICMR:575 mb:25 hour fcst: -543:9168683:d=2021032312:RWMR:575 mb:25 hour fcst: -544:9171065:d=2021032312:SNMR:575 mb:25 hour fcst: -545:9185650:d=2021032312:GRLE:575 mb:25 hour fcst: -563:9501616:d=2021032312:HGT:625 mb:25 hour fcst: -564:9525966:d=2021032312:TMP:625 mb:25 hour fcst: -565:9545100:d=2021032312:RH:625 mb:25 hour fcst: -566:9565223:d=2021032312:TCDC:625 mb:25 hour fcst: -568:9605371:d=2021032312:VVEL:625 mb:25 hour fcst: -569:9630859:d=2021032312:DZDT:625 mb:25 hour fcst: -570:9658280:d=2021032312:UGRD:625 mb:25 hour fcst: -571:9679150:d=2021032312:VGRD:625 mb:25 hour fcst: -572:9700784:d=2021032312:ABSV:625 mb:25 hour fcst: -573:9721575:d=2021032312:CLMR:625 mb:25 hour fcst: -574:9726507:d=2021032312:ICMR:625 mb:25 hour fcst: -575:9739599:d=2021032312:RWMR:625 mb:25 hour fcst: -576:9743830:d=2021032312:SNMR:625 mb:25 hour fcst: -577:9759163:d=2021032312:GRLE:625 mb:25 hour fcst: -595:10076919:d=2021032312:HGT:675 mb:25 hour fcst: -596:10101066:d=2021032312:TMP:675 mb:25 hour fcst: -597:10120511:d=2021032312:RH:675 mb:25 hour fcst: -598:10140754:d=2021032312:TCDC:675 mb:25 hour fcst: -600:10183119:d=2021032312:VVEL:675 mb:25 hour fcst: -601:10208811:d=2021032312:DZDT:675 mb:25 hour fcst: -602:10236202:d=2021032312:UGRD:675 mb:25 hour fcst: -603:10257254:d=2021032312:VGRD:675 mb:25 hour fcst: -604:10278761:d=2021032312:ABSV:675 mb:25 hour fcst: -605:10299805:d=2021032312:CLMR:675 mb:25 hour fcst: -606:10305343:d=2021032312:ICMR:675 mb:25 hour fcst: -607:10317205:d=2021032312:RWMR:675 mb:25 hour fcst: -608:10322253:d=2021032312:SNMR:675 mb:25 hour fcst: -609:10336830:d=2021032312:GRLE:675 mb:25 hour fcst: -627:10654696:d=2021032312:HGT:725 mb:25 hour fcst: -628:10678878:d=2021032312:TMP:725 mb:25 hour fcst: -629:10698609:d=2021032312:RH:725 mb:25 hour fcst: -630:10718917:d=2021032312:TCDC:725 mb:25 hour fcst: -632:10759704:d=2021032312:VVEL:725 mb:25 hour fcst: -633:10785450:d=2021032312:DZDT:725 mb:25 hour fcst: -634:10812789:d=2021032312:UGRD:725 mb:25 hour fcst: -635:10833634:d=2021032312:VGRD:725 mb:25 hour fcst: -636:10855288:d=2021032312:ABSV:725 mb:25 hour fcst: -637:10876389:d=2021032312:CLMR:725 mb:25 hour fcst: -638:10884239:d=2021032312:ICMR:725 mb:25 hour fcst: -639:10895825:d=2021032312:RWMR:725 mb:25 hour fcst: -640:10901525:d=2021032312:SNMR:725 mb:25 hour fcst: -641:10915334:d=2021032312:GRLE:725 mb:25 hour fcst: -659:11237484:d=2021032312:HGT:775 mb:25 hour fcst: -660:11261932:d=2021032312:TMP:775 mb:25 hour fcst: -661:11282066:d=2021032312:RH:775 mb:25 hour fcst: -662:11302535:d=2021032312:TCDC:775 mb:25 hour fcst: -664:11344629:d=2021032312:VVEL:775 mb:25 hour fcst: -665:11370372:d=2021032312:DZDT:775 mb:25 hour fcst: -666:11397363:d=2021032312:UGRD:775 mb:25 hour fcst: -667:11418218:d=2021032312:VGRD:775 mb:25 hour fcst: -668:11439942:d=2021032312:ABSV:775 mb:25 hour fcst: -669:11461131:d=2021032312:CLMR:775 mb:25 hour fcst: -670:11471493:d=2021032312:ICMR:775 mb:25 hour fcst: -671:11482553:d=2021032312:RWMR:775 mb:25 hour fcst: -672:11489268:d=2021032312:SNMR:775 mb:25 hour fcst: -673:11502400:d=2021032312:GRLE:775 mb:25 hour fcst: -691:11830379:d=2021032312:HGT:825 mb:25 hour fcst: -692:11854994:d=2021032312:TMP:825 mb:25 hour fcst: -693:11875805:d=2021032312:RH:825 mb:25 hour fcst: -694:11896346:d=2021032312:TCDC:825 mb:25 hour fcst: -696:11940265:d=2021032312:VVEL:825 mb:25 hour fcst: -697:11965875:d=2021032312:DZDT:825 mb:25 hour fcst: -698:11992738:d=2021032312:UGRD:825 mb:25 hour fcst: -699:12013918:d=2021032312:VGRD:825 mb:25 hour fcst: -700:12035766:d=2021032312:ABSV:825 mb:25 hour fcst: -701:12057013:d=2021032312:CLMR:825 mb:25 hour fcst: -702:12070018:d=2021032312:ICMR:825 mb:25 hour fcst: -703:12080432:d=2021032312:RWMR:825 mb:25 hour fcst: -704:12089833:d=2021032312:SNMR:825 mb:25 hour fcst: -705:12102195:d=2021032312:GRLE:825 mb:25 hour fcst: -723:12432091:d=2021032312:HGT:875 mb:25 hour fcst: -724:12456959:d=2021032312:TMP:875 mb:25 hour fcst: -725:12477780:d=2021032312:RH:875 mb:25 hour fcst: -726:12497760:d=2021032312:TCDC:875 mb:25 hour fcst: -728:12541586:d=2021032312:VVEL:875 mb:25 hour fcst: -729:12566921:d=2021032312:DZDT:875 mb:25 hour fcst: -730:12593331:d=2021032312:UGRD:875 mb:25 hour fcst: -731:12614541:d=2021032312:VGRD:875 mb:25 hour fcst: -732:12636392:d=2021032312:ABSV:875 mb:25 hour fcst: -733:12657629:d=2021032312:CLMR:875 mb:25 hour fcst: -734:12669932:d=2021032312:ICMR:875 mb:25 hour fcst: -735:12679842:d=2021032312:RWMR:875 mb:25 hour fcst: -736:12691365:d=2021032312:SNMR:875 mb:25 hour fcst: -737:12702332:d=2021032312:GRLE:875 mb:25 hour fcst: -829:14267412:d=2021032312:SOILL:0-0.1 m below ground:25 hour fcst: -832:14300710:d=2021032312:SOILL:0.1-0.4 m below ground:25 hour fcst: -835:14333187:d=2021032312:SOILL:0.4-1 m below ground:25 hour fcst: -838:14365455:d=2021032312:SOILL:1-2 m below ground:25 hour fcst: -839:14374533:d=2021032312:CNWAT:surface:25 hour fcst: -843:14423020:d=2021032312:ICETK:surface:25 hour fcst: -923:15593816:d=2021032312:DUVB:surface:24-25 hour ave fcst: -924:15610532:d=2021032312:CDUVB:surface:24-25 hour ave fcst: -969:16628689:d=2021032312:TMP:305 m above mean sea level:25 hour fcst: -970:16646890:d=2021032312:UGRD:305 m above mean sea level:25 hour fcst: -971:16666045:d=2021032312:VGRD:305 m above mean sea level:25 hour fcst: -972:16685682:d=2021032312:TMP:457 m above mean sea level:25 hour fcst: -973:16705207:d=2021032312:UGRD:457 m above mean sea level:25 hour fcst: -974:16725414:d=2021032312:VGRD:457 m above mean sea level:25 hour fcst: -975:16745965:d=2021032312:TMP:610 m above mean sea level:25 hour fcst: -976:16766247:d=2021032312:UGRD:610 m above mean sea level:25 hour fcst: -977:16787083:d=2021032312:VGRD:610 m above mean sea level:25 hour fcst: -978:16808233:d=2021032312:TMP:914 m above mean sea level:25 hour fcst: -979:16829083:d=2021032312:UGRD:914 m above mean sea level:25 hour fcst: -980:16850476:d=2021032312:VGRD:914 m above mean sea level:25 hour fcst: -990:17074846:d=2021032312:TMP:4572 m above mean sea level:25 hour fcst: -991:17095996:d=2021032312:UGRD:4572 m above mean sea level:25 hour fcst: -992:17119217:d=2021032312:VGRD:4572 m above mean sea level:25 hour fcst: -1005:17397894:d=2021032312:TMP:60-30 mb above ground:25 hour fcst: -1006:17418700:d=2021032312:RH:60-30 mb above ground:25 hour fcst: -1007:17437372:d=2021032312:SPFH:60-30 mb above ground:25 hour fcst: -1008:17466113:d=2021032312:UGRD:60-30 mb above ground:25 hour fcst: -1009:17487696:d=2021032312:VGRD:60-30 mb above ground:25 hour fcst: -1010:17509830:d=2021032312:TMP:90-60 mb above ground:25 hour fcst: -1011:17530719:d=2021032312:RH:90-60 mb above ground:25 hour fcst: -1012:17549798:d=2021032312:SPFH:90-60 mb above ground:25 hour fcst: -1013:17578754:d=2021032312:UGRD:90-60 mb above ground:25 hour fcst: -1014:17600157:d=2021032312:VGRD:90-60 mb above ground:25 hour fcst: -1015:17622118:d=2021032312:TMP:120-90 mb above ground:25 hour fcst: -1016:17643060:d=2021032312:RH:120-90 mb above ground:25 hour fcst: -1017:17662408:d=2021032312:SPFH:120-90 mb above ground:25 hour fcst: -1018:17691342:d=2021032312:UGRD:120-90 mb above ground:25 hour fcst: -1019:17712520:d=2021032312:VGRD:120-90 mb above ground:25 hour fcst: -1020:17734310:d=2021032312:TMP:150-120 mb above ground:25 hour fcst: -1021:17755216:d=2021032312:RH:150-120 mb above ground:25 hour fcst: -1022:17774920:d=2021032312:SPFH:150-120 mb above ground:25 hour fcst: -1023:17804036:d=2021032312:UGRD:150-120 mb above ground:25 hour fcst: -1024:17825058:d=2021032312:VGRD:150-120 mb above ground:25 hour fcst: -1025:17846743:d=2021032312:TMP:180-150 mb above ground:25 hour fcst: -1026:17867907:d=2021032312:RH:180-150 mb above ground:25 hour fcst: -1027:17887895:d=2021032312:SPFH:180-150 mb above ground:25 hour fcst: -1028:17917154:d=2021032312:UGRD:180-150 mb above ground:25 hour fcst: -1029:17938236:d=2021032312:VGRD:180-150 mb above ground:25 hour fcst: -1058:18420018:d=2021032312:UGRD:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -1059:18430616:d=2021032312:VGRD:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -1060:18441181:d=2021032312:TMP:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -1061:18455789:d=2021032312:HGT:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -1062:18472521:d=2021032312:PRES:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -1063:18489362:d=2021032312:VWSH:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -1064:18500326:d=2021032312:UGRD:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -1065:18511294:d=2021032312:VGRD:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -1066:18525805:d=2021032312:TMP:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -1067:18541168:d=2021032312:HGT:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -1068:18558813:d=2021032312:PRES:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -1069:18576506:d=2021032312:VWSH:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -1070:18587918:d=2021032312:UGRD:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -1071:18599243:d=2021032312:VGRD:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -1072:18610555:d=2021032312:TMP:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -1073:18626492:d=2021032312:HGT:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -1074:18645060:d=2021032312:PRES:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -1075:18663738:d=2021032312:VWSH:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -1076:18675952:d=2021032312:UGRD:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -1077:18691320:d=2021032312:VGRD:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -1078:18702645:d=2021032312:TMP:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -1079:18718439:d=2021032312:HGT:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -1080:18736836:d=2021032312:PRES:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -1081:18755297:d=2021032312:VWSH:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -1082:18767628:d=2021032312:UGRD:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -1083:18778674:d=2021032312:VGRD:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -1084:18789660:d=2021032312:TMP:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -1085:18804642:d=2021032312:HGT:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -1086:18822571:d=2021032312:PRES:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -1087:18840383:d=2021032312:VWSH:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -1088:18852505:d=2021032312:UGRD:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -1089:18863552:d=2021032312:VGRD:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -1090:18874506:d=2021032312:TMP:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -1091:18889568:d=2021032312:HGT:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -1092:18907490:d=2021032312:PRES:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -1093:18925241:d=2021032312:VWSH:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -+ exglobal_atmos_products.sh[54]true -+ exglobal_atmos_products.sh[55]export err=0 -+ exglobal_atmos_products.sh[55]err=0 -+ exglobal_atmos_products.sh[56][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[62]grid_string=0p25 -+ exglobal_atmos_products.sh[63][[ '' == \Y\E\S ]] -+ exglobal_atmos_products.sh[66]echo 'Supplemental product generation is disable for fhr = f025' -Supplemental product generation is disable for fhr = f025 -+ exglobal_atmos_products.sh[67]PGBS=NO -+ exglobal_atmos_products.sh[70]IFS=: -+ exglobal_atmos_products.sh[70]read -ra grids -+ exglobal_atmos_products.sh[154](( nset=1 )) -+ exglobal_atmos_products.sh[154](( nset <= downset )) -+ exglobal_atmos_products.sh[74]echo 'Begin processing nset = 1' -Begin processing nset = 1 -+ exglobal_atmos_products.sh[77]nproc=24 -+ exglobal_atmos_products.sh[80][[ 1 == 1 ]] -+ exglobal_atmos_products.sh[81]grp= -+ exglobal_atmos_products.sh[87]tmpfile=tmpfile_f025 -++ exglobal_atmos_products.sh[90]wgrib2 tmpfile_f025 -++ exglobal_atmos_products.sh[90]wc -l -+ exglobal_atmos_products.sh[90]ncount=753 -+ exglobal_atmos_products.sh[91][[ 24 -gt 753 ]] -+ exglobal_atmos_products.sh[95]inv=31 -+ exglobal_atmos_products.sh[96]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601/poescript -+ exglobal_atmos_products.sh[98]last=0 -+ exglobal_atmos_products.sh[134](( iproc = 1 )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=1 -+ exglobal_atmos_products.sh[101]last=31 -+ exglobal_atmos_products.sh[102][[ 31 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 31 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -31:522503:d=2021032312:UGRD:0.02 mb:25 hour fcst: -+ exglobal_atmos_products.sh[112]rc=0 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 0 == 0 ]] -+ exglobal_atmos_products.sh[115]last=32 -+ exglobal_atmos_products.sh[117][[ 1 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 1:32 -grib tmpfile_f025_1 -1:0:d=2021032312:PRMSL:mean sea level:25 hour fcst: -2:25688:d=2021032312:CLMR:1 hybrid level:25 hour fcst: -3:29343:d=2021032312:ICMR:1 hybrid level:25 hour fcst: -4:36295:d=2021032312:RWMR:1 hybrid level:25 hour fcst: -5:48732:d=2021032312:SNMR:1 hybrid level:25 hour fcst: -6:58824:d=2021032312:GRLE:1 hybrid level:25 hour fcst: -7:61276:d=2021032312:REFD:1 hybrid level:25 hour fcst: -8:87031:d=2021032312:REFD:2 hybrid level:25 hour fcst: -9:112761:d=2021032312:REFC:entire atmosphere (considered as a single layer):25 hour fcst: -10:140221:d=2021032312:VIS:surface:25 hour fcst: -11:172542:d=2021032312:UGRD:planetary boundary layer:25 hour fcst: -12:186396:d=2021032312:VGRD:planetary boundary layer:25 hour fcst: -13:200985:d=2021032312:VRATE:planetary boundary layer:25 hour fcst: -14:216099:d=2021032312:GUST:surface:25 hour fcst: -15:230805:d=2021032312:HGT:0.01 mb:25 hour fcst: -16:254189:d=2021032312:TMP:0.01 mb:25 hour fcst: -17:264779:d=2021032312:RH:0.01 mb:25 hour fcst: -18:271064:d=2021032312:SPFH:0.01 mb:25 hour fcst: -19:294006:d=2021032312:VVEL:0.01 mb:25 hour fcst: -20:320813:d=2021032312:DZDT:0.01 mb:25 hour fcst: -21:348111:d=2021032312:UGRD:0.01 mb:25 hour fcst: -22:359231:d=2021032312:VGRD:0.01 mb:25 hour fcst: -23:369764:d=2021032312:ABSV:0.01 mb:25 hour fcst: -24:385189:d=2021032312:O3MR:0.01 mb:25 hour fcst: -25:399140:d=2021032312:HGT:0.02 mb:25 hour fcst: -26:424624:d=2021032312:TMP:0.02 mb:25 hour fcst: -27:439277:d=2021032312:RH:0.02 mb:25 hour fcst: -28:446600:d=2021032312:SPFH:0.02 mb:25 hour fcst: -29:470664:d=2021032312:VVEL:0.02 mb:25 hour fcst: -30:494846:d=2021032312:DZDT:0.02 mb:25 hour fcst: -31:522503:d=2021032312:UGRD:0.02 mb:25 hour fcst: -32:534851:d=2021032312:VGRD:0.02 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_1 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_1 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_1 pgb2file_f025_1 0p25' -+ exglobal_atmos_products.sh[133][[ 32 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=33 -+ exglobal_atmos_products.sh[101]last=63 -+ exglobal_atmos_products.sh[102][[ 63 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 63 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 2 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 33:63 -grib tmpfile_f025_2 -33:546731:d=2021032312:ABSV:0.02 mb:25 hour fcst: -34:563548:d=2021032312:O3MR:0.02 mb:25 hour fcst: -35:579203:d=2021032312:HGT:0.04 mb:25 hour fcst: -36:603429:d=2021032312:TMP:0.04 mb:25 hour fcst: -37:620683:d=2021032312:RH:0.04 mb:25 hour fcst: -38:626661:d=2021032312:SPFH:0.04 mb:25 hour fcst: -39:652297:d=2021032312:VVEL:0.04 mb:25 hour fcst: -40:679819:d=2021032312:DZDT:0.04 mb:25 hour fcst: -41:704384:d=2021032312:UGRD:0.04 mb:25 hour fcst: -42:724315:d=2021032312:VGRD:0.04 mb:25 hour fcst: -43:744093:d=2021032312:ABSV:0.04 mb:25 hour fcst: -44:762514:d=2021032312:O3MR:0.04 mb:25 hour fcst: -45:786090:d=2021032312:HGT:0.07 mb:25 hour fcst: -46:810131:d=2021032312:TMP:0.07 mb:25 hour fcst: -47:827105:d=2021032312:RH:0.07 mb:25 hour fcst: -48:836678:d=2021032312:SPFH:0.07 mb:25 hour fcst: -49:860406:d=2021032312:VVEL:0.07 mb:25 hour fcst: -50:889126:d=2021032312:DZDT:0.07 mb:25 hour fcst: -51:914407:d=2021032312:UGRD:0.07 mb:25 hour fcst: -52:934816:d=2021032312:VGRD:0.07 mb:25 hour fcst: -53:955066:d=2021032312:ABSV:0.07 mb:25 hour fcst: -54:973823:d=2021032312:O3MR:0.07 mb:25 hour fcst: -55:993322:d=2021032312:HGT:0.1 mb:25 hour fcst: -56:1014866:d=2021032312:TMP:0.1 mb:25 hour fcst: -57:1031519:d=2021032312:RH:0.1 mb:25 hour fcst: -58:1039608:d=2021032312:SPFH:0.1 mb:25 hour fcst: -59:1060931:d=2021032312:VVEL:0.1 mb:25 hour fcst: -60:1090511:d=2021032312:DZDT:0.1 mb:25 hour fcst: -61:1116276:d=2021032312:UGRD:0.1 mb:25 hour fcst: -62:1136357:d=2021032312:VGRD:0.1 mb:25 hour fcst: -63:1156361:d=2021032312:ABSV:0.1 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_2 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_2 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_2 pgb2file_f025_2 0p25' -+ exglobal_atmos_products.sh[133][[ 63 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=64 -+ exglobal_atmos_products.sh[101]last=94 -+ exglobal_atmos_products.sh[102][[ 94 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 94 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 3 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 64:94 -grib tmpfile_f025_3 -64:1174931:d=2021032312:O3MR:0.1 mb:25 hour fcst: -65:1198589:d=2021032312:HGT:0.2 mb:25 hour fcst: -66:1219583:d=2021032312:TMP:0.2 mb:25 hour fcst: -67:1235664:d=2021032312:RH:0.2 mb:25 hour fcst: -68:1246733:d=2021032312:SPFH:0.2 mb:25 hour fcst: -69:1266393:d=2021032312:VVEL:0.2 mb:25 hour fcst: -70:1295218:d=2021032312:DZDT:0.2 mb:25 hour fcst: -71:1322716:d=2021032312:UGRD:0.2 mb:25 hour fcst: -72:1342068:d=2021032312:VGRD:0.2 mb:25 hour fcst: -73:1360966:d=2021032312:ABSV:0.2 mb:25 hour fcst: -74:1378683:d=2021032312:O3MR:0.2 mb:25 hour fcst: -75:1404280:d=2021032312:HGT:0.4 mb:25 hour fcst: -76:1425336:d=2021032312:TMP:0.4 mb:25 hour fcst: -77:1441114:d=2021032312:RH:0.4 mb:25 hour fcst: -78:1450120:d=2021032312:SPFH:0.4 mb:25 hour fcst: -79:1470117:d=2021032312:VVEL:0.4 mb:25 hour fcst: -80:1494987:d=2021032312:DZDT:0.4 mb:25 hour fcst: -81:1525450:d=2021032312:UGRD:0.4 mb:25 hour fcst: -82:1544379:d=2021032312:VGRD:0.4 mb:25 hour fcst: -83:1562564:d=2021032312:ABSV:0.4 mb:25 hour fcst: -84:1579356:d=2021032312:O3MR:0.4 mb:25 hour fcst: -85:1604006:d=2021032312:HGT:0.7 mb:25 hour fcst: -86:1622145:d=2021032312:TMP:0.7 mb:25 hour fcst: -87:1637956:d=2021032312:RH:0.7 mb:25 hour fcst: -88:1646886:d=2021032312:SPFH:0.7 mb:25 hour fcst: -89:1666137:d=2021032312:VVEL:0.7 mb:25 hour fcst: -90:1692176:d=2021032312:DZDT:0.7 mb:25 hour fcst: -91:1721471:d=2021032312:UGRD:0.7 mb:25 hour fcst: -92:1740085:d=2021032312:VGRD:0.7 mb:25 hour fcst: -93:1757941:d=2021032312:ABSV:0.7 mb:25 hour fcst: -94:1774401:d=2021032312:O3MR:0.7 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_3 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_3 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_3 pgb2file_f025_3 0p25' -+ exglobal_atmos_products.sh[133][[ 94 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=95 -+ exglobal_atmos_products.sh[101]last=125 -+ exglobal_atmos_products.sh[102][[ 125 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 125 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 4 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 95:125 -grib tmpfile_f025_4 -95:1794914:d=2021032312:HGT:1 mb:25 hour fcst: -96:1812911:d=2021032312:TMP:1 mb:25 hour fcst: -97:1828771:d=2021032312:RH:1 mb:25 hour fcst: -98:1838541:d=2021032312:SPFH:1 mb:25 hour fcst: -99:1857022:d=2021032312:VVEL:1 mb:25 hour fcst: -100:1883735:d=2021032312:DZDT:1 mb:25 hour fcst: -101:1909925:d=2021032312:UGRD:1 mb:25 hour fcst: -102:1928523:d=2021032312:VGRD:1 mb:25 hour fcst: -103:1946195:d=2021032312:ABSV:1 mb:25 hour fcst: -104:1962493:d=2021032312:O3MR:1 mb:25 hour fcst: -105:1983625:d=2021032312:HGT:2 mb:25 hour fcst: -106:2003828:d=2021032312:TMP:2 mb:25 hour fcst: -107:2020480:d=2021032312:RH:2 mb:25 hour fcst: -108:2029162:d=2021032312:SPFH:2 mb:25 hour fcst: -109:2048030:d=2021032312:VVEL:2 mb:25 hour fcst: -110:2076676:d=2021032312:DZDT:2 mb:25 hour fcst: -111:2103180:d=2021032312:UGRD:2 mb:25 hour fcst: -112:2114743:d=2021032312:VGRD:2 mb:25 hour fcst: -113:2133171:d=2021032312:ABSV:2 mb:25 hour fcst: -114:2150190:d=2021032312:O3MR:2 mb:25 hour fcst: -115:2172965:d=2021032312:HGT:3 mb:25 hour fcst: -116:2190765:d=2021032312:TMP:3 mb:25 hour fcst: -117:2207253:d=2021032312:RH:3 mb:25 hour fcst: -118:2213243:d=2021032312:SPFH:3 mb:25 hour fcst: -119:2234371:d=2021032312:VVEL:3 mb:25 hour fcst: -120:2264412:d=2021032312:DZDT:3 mb:25 hour fcst: -121:2290617:d=2021032312:UGRD:3 mb:25 hour fcst: -122:2301983:d=2021032312:VGRD:3 mb:25 hour fcst: -123:2319975:d=2021032312:ABSV:3 mb:25 hour fcst: -124:2336893:d=2021032312:O3MR:3 mb:25 hour fcst: -125:2360195:d=2021032312:HGT:5 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_4 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_4 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_4 pgb2file_f025_4 0p25' -+ exglobal_atmos_products.sh[133][[ 125 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=126 -+ exglobal_atmos_products.sh[101]last=156 -+ exglobal_atmos_products.sh[102][[ 156 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 156 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 5 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 126:156 -grib tmpfile_f025_5 -126:2377852:d=2021032312:TMP:5 mb:25 hour fcst: -127:2394543:d=2021032312:RH:5 mb:25 hour fcst: -128:2403733:d=2021032312:SPFH:5 mb:25 hour fcst: -129:2424446:d=2021032312:VVEL:5 mb:25 hour fcst: -130:2453841:d=2021032312:DZDT:5 mb:25 hour fcst: -131:2479891:d=2021032312:UGRD:5 mb:25 hour fcst: -132:2491488:d=2021032312:VGRD:5 mb:25 hour fcst: -133:2510096:d=2021032312:ABSV:5 mb:25 hour fcst: -134:2527781:d=2021032312:O3MR:5 mb:25 hour fcst: -135:2550552:d=2021032312:HGT:7 mb:25 hour fcst: -136:2568217:d=2021032312:TMP:7 mb:25 hour fcst: -137:2584301:d=2021032312:RH:7 mb:25 hour fcst: -138:2590351:d=2021032312:SPFH:7 mb:25 hour fcst: -139:2609606:d=2021032312:VVEL:7 mb:25 hour fcst: -140:2634228:d=2021032312:DZDT:7 mb:25 hour fcst: -141:2659864:d=2021032312:UGRD:7 mb:25 hour fcst: -142:2671281:d=2021032312:VGRD:7 mb:25 hour fcst: -143:2690034:d=2021032312:ABSV:7 mb:25 hour fcst: -144:2707769:d=2021032312:O3MR:7 mb:25 hour fcst: -145:2731436:d=2021032312:HGT:10 mb:25 hour fcst: -146:2751273:d=2021032312:TMP:10 mb:25 hour fcst: -147:2767471:d=2021032312:RH:10 mb:25 hour fcst: -148:2775142:d=2021032312:SPFH:10 mb:25 hour fcst: -149:2793463:d=2021032312:VVEL:10 mb:25 hour fcst: -150:2819001:d=2021032312:DZDT:10 mb:25 hour fcst: -151:2844272:d=2021032312:UGRD:10 mb:25 hour fcst: -152:2855600:d=2021032312:VGRD:10 mb:25 hour fcst: -153:2874169:d=2021032312:ABSV:10 mb:25 hour fcst: -154:2891838:d=2021032312:O3MR:10 mb:25 hour fcst: -155:2910572:d=2021032312:HGT:15 mb:25 hour fcst: -156:2930102:d=2021032312:TMP:15 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_5 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_5 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_5 pgb2file_f025_5 0p25' -+ exglobal_atmos_products.sh[133][[ 156 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=157 -+ exglobal_atmos_products.sh[101]last=187 -+ exglobal_atmos_products.sh[102][[ 187 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 187 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 6 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 157:187 -grib tmpfile_f025_6 -157:2946232:d=2021032312:RH:15 mb:25 hour fcst: -158:2951518:d=2021032312:SPFH:15 mb:25 hour fcst: -159:2966965:d=2021032312:VVEL:15 mb:25 hour fcst: -160:2993621:d=2021032312:DZDT:15 mb:25 hour fcst: -161:3024127:d=2021032312:UGRD:15 mb:25 hour fcst: -162:3035325:d=2021032312:VGRD:15 mb:25 hour fcst: -163:3053768:d=2021032312:ABSV:15 mb:25 hour fcst: -164:3071318:d=2021032312:O3MR:15 mb:25 hour fcst: -165:3090249:d=2021032312:HGT:20 mb:25 hour fcst: -166:3109661:d=2021032312:TMP:20 mb:25 hour fcst: -167:3125676:d=2021032312:RH:20 mb:25 hour fcst: -168:3131845:d=2021032312:SPFH:20 mb:25 hour fcst: -169:3146021:d=2021032312:VVEL:20 mb:25 hour fcst: -170:3173520:d=2021032312:DZDT:20 mb:25 hour fcst: -171:3203755:d=2021032312:UGRD:20 mb:25 hour fcst: -172:3222365:d=2021032312:VGRD:20 mb:25 hour fcst: -173:3240870:d=2021032312:ABSV:20 mb:25 hour fcst: -174:3258429:d=2021032312:O3MR:20 mb:25 hour fcst: -175:3283298:d=2021032312:HGT:30 mb:25 hour fcst: -176:3302405:d=2021032312:TMP:30 mb:25 hour fcst: -177:3318338:d=2021032312:RH:30 mb:25 hour fcst: -178:3325947:d=2021032312:SPFH:30 mb:25 hour fcst: -179:3341278:d=2021032312:VVEL:30 mb:25 hour fcst: -180:3369610:d=2021032312:DZDT:30 mb:25 hour fcst: -181:3399506:d=2021032312:UGRD:30 mb:25 hour fcst: -182:3418211:d=2021032312:VGRD:30 mb:25 hour fcst: -183:3436710:d=2021032312:ABSV:30 mb:25 hour fcst: -184:3454414:d=2021032312:O3MR:30 mb:25 hour fcst: -185:3481157:d=2021032312:HGT:40 mb:25 hour fcst: -186:3500141:d=2021032312:TMP:40 mb:25 hour fcst: -187:3516456:d=2021032312:RH:40 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_6 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_6 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_6 pgb2file_f025_6 0p25' -+ exglobal_atmos_products.sh[133][[ 187 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=188 -+ exglobal_atmos_products.sh[101]last=218 -+ exglobal_atmos_products.sh[102][[ 218 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 218 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 7 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 188:218 -grib tmpfile_f025_7 -188:3525925:d=2021032312:SPFH:40 mb:25 hour fcst: -189:3542959:d=2021032312:VVEL:40 mb:25 hour fcst: -190:3572326:d=2021032312:DZDT:40 mb:25 hour fcst: -191:3602018:d=2021032312:UGRD:40 mb:25 hour fcst: -192:3620869:d=2021032312:VGRD:40 mb:25 hour fcst: -193:3639576:d=2021032312:ABSV:40 mb:25 hour fcst: -194:3657499:d=2021032312:O3MR:40 mb:25 hour fcst: -195:3684461:d=2021032312:HGT:50 mb:25 hour fcst: -196:3703406:d=2021032312:TMP:50 mb:25 hour fcst: -197:3719538:d=2021032312:RH:50 mb:25 hour fcst: -198:3730155:d=2021032312:TCDC:50 mb:25 hour fcst: -199:3730334:d=2021032312:SPFH:50 mb:25 hour fcst: -200:3748840:d=2021032312:VVEL:50 mb:25 hour fcst: -201:3779025:d=2021032312:DZDT:50 mb:25 hour fcst: -202:3811122:d=2021032312:UGRD:50 mb:25 hour fcst: -203:3829670:d=2021032312:VGRD:50 mb:25 hour fcst: -204:3848193:d=2021032312:ABSV:50 mb:25 hour fcst: -205:3865912:d=2021032312:CLMR:50 mb:25 hour fcst: -206:3866091:d=2021032312:ICMR:50 mb:25 hour fcst: -207:3866270:d=2021032312:RWMR:50 mb:25 hour fcst: -208:3866449:d=2021032312:SNMR:50 mb:25 hour fcst: -209:3866628:d=2021032312:GRLE:50 mb:25 hour fcst: -210:3866807:d=2021032312:O3MR:50 mb:25 hour fcst: -211:3893152:d=2021032312:HGT:70 mb:25 hour fcst: -212:3912162:d=2021032312:TMP:70 mb:25 hour fcst: -213:3928568:d=2021032312:RH:70 mb:25 hour fcst: -214:3936906:d=2021032312:SPFH:70 mb:25 hour fcst: -215:3957820:d=2021032312:VVEL:70 mb:25 hour fcst: -216:3989012:d=2021032312:DZDT:70 mb:25 hour fcst: -217:4018699:d=2021032312:UGRD:70 mb:25 hour fcst: -218:4037341:d=2021032312:VGRD:70 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_7 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_7 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_7 pgb2file_f025_7 0p25' -+ exglobal_atmos_products.sh[133][[ 218 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=219 -+ exglobal_atmos_products.sh[101]last=249 -+ exglobal_atmos_products.sh[102][[ 249 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 249 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 8 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 219:249 -grib tmpfile_f025_8 -219:4056047:d=2021032312:ABSV:70 mb:25 hour fcst: -220:4073808:d=2021032312:O3MR:70 mb:25 hour fcst: -221:4099664:d=2021032312:HGT:100 mb:25 hour fcst: -222:4119270:d=2021032312:TMP:100 mb:25 hour fcst: -223:4136220:d=2021032312:RH:100 mb:25 hour fcst: -224:4146353:d=2021032312:TCDC:100 mb:25 hour fcst: -225:4146532:d=2021032312:SPFH:100 mb:25 hour fcst: -226:4172149:d=2021032312:VVEL:100 mb:25 hour fcst: -227:4197145:d=2021032312:DZDT:100 mb:25 hour fcst: -228:4227514:d=2021032312:UGRD:100 mb:25 hour fcst: -229:4246893:d=2021032312:VGRD:100 mb:25 hour fcst: -230:4266277:d=2021032312:ABSV:100 mb:25 hour fcst: -231:4284472:d=2021032312:CLMR:100 mb:25 hour fcst: -232:4284651:d=2021032312:ICMR:100 mb:25 hour fcst: -233:4285356:d=2021032312:RWMR:100 mb:25 hour fcst: -234:4285535:d=2021032312:SNMR:100 mb:25 hour fcst: -235:4287632:d=2021032312:GRLE:100 mb:25 hour fcst: -236:4287811:d=2021032312:O3MR:100 mb:25 hour fcst: -237:4311824:d=2021032312:HGT:150 mb:25 hour fcst: -238:4332167:d=2021032312:TMP:150 mb:25 hour fcst: -239:4349144:d=2021032312:RH:150 mb:25 hour fcst: -240:4360564:d=2021032312:TCDC:150 mb:25 hour fcst: -241:4364552:d=2021032312:SPFH:150 mb:25 hour fcst: -242:4388889:d=2021032312:VVEL:150 mb:25 hour fcst: -243:4416172:d=2021032312:DZDT:150 mb:25 hour fcst: -244:4441772:d=2021032312:UGRD:150 mb:25 hour fcst: -245:4461712:d=2021032312:VGRD:150 mb:25 hour fcst: -246:4481745:d=2021032312:ABSV:150 mb:25 hour fcst: -247:4500613:d=2021032312:CLMR:150 mb:25 hour fcst: -248:4500792:d=2021032312:ICMR:150 mb:25 hour fcst: -249:4506047:d=2021032312:RWMR:150 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_8 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_8 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_8 pgb2file_f025_8 0p25' -+ exglobal_atmos_products.sh[133][[ 249 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=250 -+ exglobal_atmos_products.sh[101]last=280 -+ exglobal_atmos_products.sh[102][[ 280 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 280 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 9 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 250:280 -grib tmpfile_f025_9 -250:4506226:d=2021032312:SNMR:150 mb:25 hour fcst: -251:4509694:d=2021032312:GRLE:150 mb:25 hour fcst: -252:4510220:d=2021032312:O3MR:150 mb:25 hour fcst: -253:4532101:d=2021032312:HGT:200 mb:25 hour fcst: -254:4553151:d=2021032312:TMP:200 mb:25 hour fcst: -255:4571277:d=2021032312:RH:200 mb:25 hour fcst: -256:4586337:d=2021032312:TCDC:200 mb:25 hour fcst: -257:4592782:d=2021032312:SPFH:200 mb:25 hour fcst: -258:4618536:d=2021032312:VVEL:200 mb:25 hour fcst: -259:4645395:d=2021032312:DZDT:200 mb:25 hour fcst: -260:4671687:d=2021032312:UGRD:200 mb:25 hour fcst: -261:4685057:d=2021032312:VGRD:200 mb:25 hour fcst: -262:4698813:d=2021032312:ABSV:200 mb:25 hour fcst: -263:4718852:d=2021032312:CLMR:200 mb:25 hour fcst: -264:4719031:d=2021032312:ICMR:200 mb:25 hour fcst: -265:4727541:d=2021032312:RWMR:200 mb:25 hour fcst: -266:4727720:d=2021032312:SNMR:200 mb:25 hour fcst: -267:4733047:d=2021032312:GRLE:200 mb:25 hour fcst: -268:4733405:d=2021032312:O3MR:200 mb:25 hour fcst: -269:4754011:d=2021032312:HGT:250 mb:25 hour fcst: -270:4775794:d=2021032312:TMP:250 mb:25 hour fcst: -271:4794643:d=2021032312:RH:250 mb:25 hour fcst: -272:4812470:d=2021032312:TCDC:250 mb:25 hour fcst: -273:4823095:d=2021032312:SPFH:250 mb:25 hour fcst: -274:4851138:d=2021032312:VVEL:250 mb:25 hour fcst: -275:4879516:d=2021032312:DZDT:250 mb:25 hour fcst: -276:4906663:d=2021032312:UGRD:250 mb:25 hour fcst: -277:4920793:d=2021032312:VGRD:250 mb:25 hour fcst: -278:4935648:d=2021032312:ABSV:250 mb:25 hour fcst: -279:4957221:d=2021032312:CLMR:250 mb:25 hour fcst: -280:4957400:d=2021032312:ICMR:250 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_9 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_9 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_9 pgb2file_f025_9 0p25' -+ exglobal_atmos_products.sh[133][[ 280 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=281 -+ exglobal_atmos_products.sh[101]last=311 -+ exglobal_atmos_products.sh[102][[ 311 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 311 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 10 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 281:311 -grib tmpfile_f025_10 -281:4971626:d=2021032312:RWMR:250 mb:25 hour fcst: -282:4971805:d=2021032312:SNMR:250 mb:25 hour fcst: -283:4980631:d=2021032312:GRLE:250 mb:25 hour fcst: -284:4980936:d=2021032312:O3MR:250 mb:25 hour fcst: -285:5006545:d=2021032312:HGT:300 mb:25 hour fcst: -286:5028838:d=2021032312:TMP:300 mb:25 hour fcst: -287:5047272:d=2021032312:RH:300 mb:25 hour fcst: -288:5066203:d=2021032312:TCDC:300 mb:25 hour fcst: -289:5078956:d=2021032312:SPFH:300 mb:25 hour fcst: -290:5105668:d=2021032312:VVEL:300 mb:25 hour fcst: -291:5129573:d=2021032312:DZDT:300 mb:25 hour fcst: -292:5157684:d=2021032312:UGRD:300 mb:25 hour fcst: -293:5172348:d=2021032312:VGRD:300 mb:25 hour fcst: -294:5187676:d=2021032312:ABSV:300 mb:25 hour fcst: -295:5210003:d=2021032312:CLMR:300 mb:25 hour fcst: -296:5210182:d=2021032312:ICMR:300 mb:25 hour fcst: -297:5225974:d=2021032312:RWMR:300 mb:25 hour fcst: -298:5226153:d=2021032312:SNMR:300 mb:25 hour fcst: -299:5237682:d=2021032312:GRLE:300 mb:25 hour fcst: -300:5238093:d=2021032312:O3MR:300 mb:25 hour fcst: -301:5265238:d=2021032312:HGT:350 mb:25 hour fcst: -302:5287182:d=2021032312:TMP:350 mb:25 hour fcst: -303:5305433:d=2021032312:RH:350 mb:25 hour fcst: -304:5324518:d=2021032312:TCDC:350 mb:25 hour fcst: -305:5337696:d=2021032312:SPFH:350 mb:25 hour fcst: -306:5365909:d=2021032312:VVEL:350 mb:25 hour fcst: -307:5390685:d=2021032312:DZDT:350 mb:25 hour fcst: -308:5418947:d=2021032312:UGRD:350 mb:25 hour fcst: -309:5433554:d=2021032312:VGRD:350 mb:25 hour fcst: -310:5448765:d=2021032312:ABSV:350 mb:25 hour fcst: -311:5471005:d=2021032312:CLMR:350 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_10 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_10 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_10 pgb2file_f025_10 0p25' -+ exglobal_atmos_products.sh[133][[ 311 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=312 -+ exglobal_atmos_products.sh[101]last=342 -+ exglobal_atmos_products.sh[102][[ 342 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 342 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 11 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 312:342 -grib tmpfile_f025_11 -312:5471184:d=2021032312:ICMR:350 mb:25 hour fcst: -313:5488172:d=2021032312:RWMR:350 mb:25 hour fcst: -314:5488351:d=2021032312:SNMR:350 mb:25 hour fcst: -315:5500376:d=2021032312:GRLE:350 mb:25 hour fcst: -316:5500900:d=2021032312:O3MR:350 mb:25 hour fcst: -317:5526458:d=2021032312:HGT:400 mb:25 hour fcst: -318:5548346:d=2021032312:TMP:400 mb:25 hour fcst: -319:5566667:d=2021032312:RH:400 mb:25 hour fcst: -320:5585857:d=2021032312:TCDC:400 mb:25 hour fcst: -321:5599075:d=2021032312:SPFH:400 mb:25 hour fcst: -322:5626848:d=2021032312:VVEL:400 mb:25 hour fcst: -323:5651821:d=2021032312:DZDT:400 mb:25 hour fcst: -324:5679911:d=2021032312:UGRD:400 mb:25 hour fcst: -325:5701889:d=2021032312:VGRD:400 mb:25 hour fcst: -326:5724609:d=2021032312:ABSV:400 mb:25 hour fcst: -327:5746553:d=2021032312:CLMR:400 mb:25 hour fcst: -328:5746936:d=2021032312:ICMR:400 mb:25 hour fcst: -329:5763915:d=2021032312:RWMR:400 mb:25 hour fcst: -330:5764211:d=2021032312:SNMR:400 mb:25 hour fcst: -331:5778127:d=2021032312:GRLE:400 mb:25 hour fcst: -332:5779012:d=2021032312:O3MR:400 mb:25 hour fcst: -333:5803578:d=2021032312:HGT:450 mb:25 hour fcst: -334:5827530:d=2021032312:TMP:450 mb:25 hour fcst: -335:5846095:d=2021032312:RH:450 mb:25 hour fcst: -336:5865482:d=2021032312:TCDC:450 mb:25 hour fcst: -337:5878266:d=2021032312:SPFH:450 mb:25 hour fcst: -338:5907587:d=2021032312:VVEL:450 mb:25 hour fcst: -339:5932762:d=2021032312:DZDT:450 mb:25 hour fcst: -340:5960790:d=2021032312:UGRD:450 mb:25 hour fcst: -341:5982469:d=2021032312:VGRD:450 mb:25 hour fcst: -342:6004784:d=2021032312:ABSV:450 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_11 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_11 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_11 pgb2file_f025_11 0p25' -+ exglobal_atmos_products.sh[133][[ 342 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=343 -+ exglobal_atmos_products.sh[101]last=373 -+ exglobal_atmos_products.sh[102][[ 373 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 373 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 12 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 343:373 -grib tmpfile_f025_12 -343:6026200:d=2021032312:CLMR:450 mb:25 hour fcst: -344:6027148:d=2021032312:ICMR:450 mb:25 hour fcst: -345:6043968:d=2021032312:RWMR:450 mb:25 hour fcst: -346:6044522:d=2021032312:SNMR:450 mb:25 hour fcst: -347:6058160:d=2021032312:GRLE:450 mb:25 hour fcst: -348:6059245:d=2021032312:O3MR:450 mb:25 hour fcst: -349:6083294:d=2021032312:HGT:500 mb:25 hour fcst: -350:6107014:d=2021032312:TMP:500 mb:25 hour fcst: -351:6125686:d=2021032312:RH:500 mb:25 hour fcst: -352:6145228:d=2021032312:TCDC:500 mb:25 hour fcst: -353:6157570:d=2021032312:SPFH:500 mb:25 hour fcst: -354:6185873:d=2021032312:VVEL:500 mb:25 hour fcst: -355:6211126:d=2021032312:DZDT:500 mb:25 hour fcst: -356:6238849:d=2021032312:UGRD:500 mb:25 hour fcst: -357:6260318:d=2021032312:VGRD:500 mb:25 hour fcst: -358:6282354:d=2021032312:ABSV:500 mb:25 hour fcst: -359:6303511:d=2021032312:CLMR:500 mb:25 hour fcst: -360:6305154:d=2021032312:ICMR:500 mb:25 hour fcst: -361:6321083:d=2021032312:RWMR:500 mb:25 hour fcst: -362:6322001:d=2021032312:SNMR:500 mb:25 hour fcst: -363:6336292:d=2021032312:GRLE:500 mb:25 hour fcst: -364:6338077:d=2021032312:O3MR:500 mb:25 hour fcst: -365:6361480:d=2021032312:HGT:550 mb:25 hour fcst: -366:6385052:d=2021032312:TMP:550 mb:25 hour fcst: -367:6403841:d=2021032312:RH:550 mb:25 hour fcst: -368:6423636:d=2021032312:TCDC:550 mb:25 hour fcst: -369:6435711:d=2021032312:SPFH:550 mb:25 hour fcst: -370:6465358:d=2021032312:VVEL:550 mb:25 hour fcst: -371:6490680:d=2021032312:DZDT:550 mb:25 hour fcst: -372:6518171:d=2021032312:UGRD:550 mb:25 hour fcst: -373:6539171:d=2021032312:VGRD:550 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_12 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_12 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_12 pgb2file_f025_12 0p25' -+ exglobal_atmos_products.sh[133][[ 373 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=374 -+ exglobal_atmos_products.sh[101]last=404 -+ exglobal_atmos_products.sh[102][[ 404 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 404 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -404:7090915:d=2021032312:UGRD:650 mb:25 hour fcst: -+ exglobal_atmos_products.sh[112]rc=0 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 0 == 0 ]] -+ exglobal_atmos_products.sh[115]last=405 -+ exglobal_atmos_products.sh[117][[ 13 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 374:405 -grib tmpfile_f025_13 -374:6560947:d=2021032312:ABSV:550 mb:25 hour fcst: -375:6581897:d=2021032312:CLMR:550 mb:25 hour fcst: -376:6584970:d=2021032312:ICMR:550 mb:25 hour fcst: -377:6601598:d=2021032312:RWMR:550 mb:25 hour fcst: -378:6603224:d=2021032312:SNMR:550 mb:25 hour fcst: -379:6617799:d=2021032312:GRLE:550 mb:25 hour fcst: -380:6619987:d=2021032312:O3MR:550 mb:25 hour fcst: -381:6642846:d=2021032312:HGT:600 mb:25 hour fcst: -382:6667256:d=2021032312:TMP:600 mb:25 hour fcst: -383:6686307:d=2021032312:RH:600 mb:25 hour fcst: -384:6706380:d=2021032312:TCDC:600 mb:25 hour fcst: -385:6718374:d=2021032312:SPFH:600 mb:25 hour fcst: -386:6748833:d=2021032312:VVEL:600 mb:25 hour fcst: -387:6774240:d=2021032312:DZDT:600 mb:25 hour fcst: -388:6801668:d=2021032312:UGRD:600 mb:25 hour fcst: -389:6822747:d=2021032312:VGRD:600 mb:25 hour fcst: -390:6844350:d=2021032312:ABSV:600 mb:25 hour fcst: -391:6865089:d=2021032312:CLMR:600 mb:25 hour fcst: -392:6870057:d=2021032312:ICMR:600 mb:25 hour fcst: -393:6883904:d=2021032312:RWMR:600 mb:25 hour fcst: -394:6887624:d=2021032312:SNMR:600 mb:25 hour fcst: -395:6903404:d=2021032312:GRLE:600 mb:25 hour fcst: -396:6905868:d=2021032312:O3MR:600 mb:25 hour fcst: -397:6933815:d=2021032312:HGT:650 mb:25 hour fcst: -398:6957958:d=2021032312:TMP:650 mb:25 hour fcst: -399:6977266:d=2021032312:RH:650 mb:25 hour fcst: -400:6997508:d=2021032312:TCDC:650 mb:25 hour fcst: -401:7009271:d=2021032312:SPFH:650 mb:25 hour fcst: -402:7038151:d=2021032312:VVEL:650 mb:25 hour fcst: -403:7063510:d=2021032312:DZDT:650 mb:25 hour fcst: -404:7090915:d=2021032312:UGRD:650 mb:25 hour fcst: -405:7111841:d=2021032312:VGRD:650 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_13 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_13 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_13 pgb2file_f025_13 0p25' -+ exglobal_atmos_products.sh[133][[ 405 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=406 -+ exglobal_atmos_products.sh[101]last=436 -+ exglobal_atmos_products.sh[102][[ 436 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 436 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -436:7668667:d=2021032312:UGRD:750 mb:25 hour fcst: -+ exglobal_atmos_products.sh[112]rc=0 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 0 == 0 ]] -+ exglobal_atmos_products.sh[115]last=437 -+ exglobal_atmos_products.sh[117][[ 14 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 406:437 -grib tmpfile_f025_14 -406:7133380:d=2021032312:ABSV:650 mb:25 hour fcst: -407:7154235:d=2021032312:CLMR:650 mb:25 hour fcst: -408:7159464:d=2021032312:ICMR:650 mb:25 hour fcst: -409:7171881:d=2021032312:RWMR:650 mb:25 hour fcst: -410:7176534:d=2021032312:SNMR:650 mb:25 hour fcst: -411:7191460:d=2021032312:GRLE:650 mb:25 hour fcst: -412:7193761:d=2021032312:O3MR:650 mb:25 hour fcst: -413:7221274:d=2021032312:HGT:700 mb:25 hour fcst: -414:7245428:d=2021032312:TMP:700 mb:25 hour fcst: -415:7265211:d=2021032312:RH:700 mb:25 hour fcst: -416:7285540:d=2021032312:TCDC:700 mb:25 hour fcst: -417:7297131:d=2021032312:SPFH:700 mb:25 hour fcst: -418:7325738:d=2021032312:VVEL:700 mb:25 hour fcst: -419:7351308:d=2021032312:DZDT:700 mb:25 hour fcst: -420:7378681:d=2021032312:UGRD:700 mb:25 hour fcst: -421:7399537:d=2021032312:VGRD:700 mb:25 hour fcst: -422:7421143:d=2021032312:ABSV:700 mb:25 hour fcst: -423:7442178:d=2021032312:CLMR:700 mb:25 hour fcst: -424:7449189:d=2021032312:ICMR:700 mb:25 hour fcst: -425:7460384:d=2021032312:RWMR:700 mb:25 hour fcst: -426:7465997:d=2021032312:SNMR:700 mb:25 hour fcst: -427:7480125:d=2021032312:GRLE:700 mb:25 hour fcst: -428:7482016:d=2021032312:O3MR:700 mb:25 hour fcst: -429:7509338:d=2021032312:HGT:750 mb:25 hour fcst: -430:7533772:d=2021032312:TMP:750 mb:25 hour fcst: -431:7553680:d=2021032312:RH:750 mb:25 hour fcst: -432:7574091:d=2021032312:TCDC:750 mb:25 hour fcst: -433:7586485:d=2021032312:SPFH:750 mb:25 hour fcst: -434:7615635:d=2021032312:VVEL:750 mb:25 hour fcst: -435:7641397:d=2021032312:DZDT:750 mb:25 hour fcst: -436:7668667:d=2021032312:UGRD:750 mb:25 hour fcst: -437:7689683:d=2021032312:VGRD:750 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_14 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_14 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_14 pgb2file_f025_14 0p25' -+ exglobal_atmos_products.sh[133][[ 437 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=438 -+ exglobal_atmos_products.sh[101]last=468 -+ exglobal_atmos_products.sh[102][[ 468 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 468 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -468:8262660:d=2021032312:UGRD:850 mb:25 hour fcst: -+ exglobal_atmos_products.sh[112]rc=0 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 0 == 0 ]] -+ exglobal_atmos_products.sh[115]last=469 -+ exglobal_atmos_products.sh[117][[ 15 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 438:469 -grib tmpfile_f025_15 -438:7711412:d=2021032312:ABSV:750 mb:25 hour fcst: -439:7732568:d=2021032312:CLMR:750 mb:25 hour fcst: -440:7741710:d=2021032312:ICMR:750 mb:25 hour fcst: -441:7753356:d=2021032312:RWMR:750 mb:25 hour fcst: -442:7759244:d=2021032312:SNMR:750 mb:25 hour fcst: -443:7772867:d=2021032312:GRLE:750 mb:25 hour fcst: -444:7775022:d=2021032312:O3MR:750 mb:25 hour fcst: -445:7802275:d=2021032312:HGT:800 mb:25 hour fcst: -446:7826805:d=2021032312:TMP:800 mb:25 hour fcst: -447:7847358:d=2021032312:RH:800 mb:25 hour fcst: -448:7867809:d=2021032312:TCDC:800 mb:25 hour fcst: -449:7881252:d=2021032312:SPFH:800 mb:25 hour fcst: -450:7910719:d=2021032312:VVEL:800 mb:25 hour fcst: -451:7936402:d=2021032312:DZDT:800 mb:25 hour fcst: -452:7963456:d=2021032312:UGRD:800 mb:25 hour fcst: -453:7984572:d=2021032312:VGRD:800 mb:25 hour fcst: -454:8006323:d=2021032312:ABSV:800 mb:25 hour fcst: -455:8027507:d=2021032312:CLMR:800 mb:25 hour fcst: -456:8039262:d=2021032312:ICMR:800 mb:25 hour fcst: -457:8050301:d=2021032312:RWMR:800 mb:25 hour fcst: -458:8058663:d=2021032312:SNMR:800 mb:25 hour fcst: -459:8071411:d=2021032312:GRLE:800 mb:25 hour fcst: -460:8073826:d=2021032312:O3MR:800 mb:25 hour fcst: -461:8100813:d=2021032312:HGT:850 mb:25 hour fcst: -462:8125563:d=2021032312:TMP:850 mb:25 hour fcst: -463:8146441:d=2021032312:RH:850 mb:25 hour fcst: -464:8166690:d=2021032312:TCDC:850 mb:25 hour fcst: -465:8181298:d=2021032312:SPFH:850 mb:25 hour fcst: -466:8210850:d=2021032312:VVEL:850 mb:25 hour fcst: -467:8236169:d=2021032312:DZDT:850 mb:25 hour fcst: -468:8262660:d=2021032312:UGRD:850 mb:25 hour fcst: -469:8283843:d=2021032312:VGRD:850 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_15 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_15 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_15 pgb2file_f025_15 0p25' -+ exglobal_atmos_products.sh[133][[ 469 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=470 -+ exglobal_atmos_products.sh[101]last=500 -+ exglobal_atmos_products.sh[102][[ 500 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 500 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -500:8852403:d=2021032312:UGRD:925 mb:25 hour fcst: -+ exglobal_atmos_products.sh[112]rc=0 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 0 == 0 ]] -+ exglobal_atmos_products.sh[115]last=501 -+ exglobal_atmos_products.sh[117][[ 16 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 470:501 -grib tmpfile_f025_16 -470:8305667:d=2021032312:ABSV:850 mb:25 hour fcst: -471:8326925:d=2021032312:CLMR:850 mb:25 hour fcst: -472:8340352:d=2021032312:ICMR:850 mb:25 hour fcst: -473:8350065:d=2021032312:RWMR:850 mb:25 hour fcst: -474:8360570:d=2021032312:SNMR:850 mb:25 hour fcst: -475:8372207:d=2021032312:GRLE:850 mb:25 hour fcst: -476:8374842:d=2021032312:O3MR:850 mb:25 hour fcst: -477:8401441:d=2021032312:HGT:900 mb:25 hour fcst: -478:8426468:d=2021032312:TMP:900 mb:25 hour fcst: -479:8447450:d=2021032312:RH:900 mb:25 hour fcst: -480:8467274:d=2021032312:TCDC:900 mb:25 hour fcst: -481:8480379:d=2021032312:SPFH:900 mb:25 hour fcst: -482:8509851:d=2021032312:VVEL:900 mb:25 hour fcst: -483:8534730:d=2021032312:DZDT:900 mb:25 hour fcst: -484:8560834:d=2021032312:UGRD:900 mb:25 hour fcst: -485:8582036:d=2021032312:VGRD:900 mb:25 hour fcst: -486:8603928:d=2021032312:ABSV:900 mb:25 hour fcst: -487:8625112:d=2021032312:CLMR:900 mb:25 hour fcst: -488:8635149:d=2021032312:ICMR:900 mb:25 hour fcst: -489:8643531:d=2021032312:RWMR:900 mb:25 hour fcst: -490:8655744:d=2021032312:SNMR:900 mb:25 hour fcst: -491:8665963:d=2021032312:GRLE:900 mb:25 hour fcst: -492:8668791:d=2021032312:O3MR:900 mb:25 hour fcst: -493:8694826:d=2021032312:HGT:925 mb:25 hour fcst: -494:8719982:d=2021032312:TMP:925 mb:25 hour fcst: -495:8741160:d=2021032312:RH:925 mb:25 hour fcst: -496:8760924:d=2021032312:TCDC:925 mb:25 hour fcst: -497:8772837:d=2021032312:SPFH:925 mb:25 hour fcst: -498:8802377:d=2021032312:VVEL:925 mb:25 hour fcst: -499:8826802:d=2021032312:DZDT:925 mb:25 hour fcst: -500:8852403:d=2021032312:UGRD:925 mb:25 hour fcst: -501:8873669:d=2021032312:VGRD:925 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_16 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_16 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_16 pgb2file_f025_16 0p25' -+ exglobal_atmos_products.sh[133][[ 501 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=502 -+ exglobal_atmos_products.sh[101]last=532 -+ exglobal_atmos_products.sh[102][[ 532 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 532 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 17 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 502:532 -grib tmpfile_f025_17 -502:8895587:d=2021032312:ABSV:925 mb:25 hour fcst: -503:8916863:d=2021032312:CLMR:925 mb:25 hour fcst: -504:8925452:d=2021032312:ICMR:925 mb:25 hour fcst: -505:8933067:d=2021032312:RWMR:925 mb:25 hour fcst: -506:8945366:d=2021032312:SNMR:925 mb:25 hour fcst: -507:8954975:d=2021032312:GRLE:925 mb:25 hour fcst: -508:8957737:d=2021032312:O3MR:925 mb:25 hour fcst: -509:8983686:d=2021032312:HGT:950 mb:25 hour fcst: -510:9008997:d=2021032312:TMP:950 mb:25 hour fcst: -511:9030079:d=2021032312:RH:950 mb:25 hour fcst: -512:9049552:d=2021032312:TCDC:950 mb:25 hour fcst: -513:9059500:d=2021032312:SPFH:950 mb:25 hour fcst: -514:9086668:d=2021032312:VVEL:950 mb:25 hour fcst: -515:9110491:d=2021032312:DZDT:950 mb:25 hour fcst: -516:9135355:d=2021032312:UGRD:950 mb:25 hour fcst: -517:9156647:d=2021032312:VGRD:950 mb:25 hour fcst: -518:9178590:d=2021032312:ABSV:950 mb:25 hour fcst: -519:9199667:d=2021032312:CLMR:950 mb:25 hour fcst: -520:9206203:d=2021032312:ICMR:950 mb:25 hour fcst: -521:9213306:d=2021032312:RWMR:950 mb:25 hour fcst: -522:9225655:d=2021032312:SNMR:950 mb:25 hour fcst: -523:9234464:d=2021032312:GRLE:950 mb:25 hour fcst: -524:9237316:d=2021032312:O3MR:950 mb:25 hour fcst: -525:9263231:d=2021032312:HINDEX:surface:25 hour fcst: -526:9268328:d=2021032312:HGT:975 mb:25 hour fcst: -527:9293788:d=2021032312:TMP:975 mb:25 hour fcst: -528:9314695:d=2021032312:RH:975 mb:25 hour fcst: -529:9333814:d=2021032312:TCDC:975 mb:25 hour fcst: -530:9341953:d=2021032312:SPFH:975 mb:25 hour fcst: -531:9369027:d=2021032312:VVEL:975 mb:25 hour fcst: -532:9391723:d=2021032312:DZDT:975 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_17 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_17 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_17 pgb2file_f025_17 0p25' -+ exglobal_atmos_products.sh[133][[ 532 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=533 -+ exglobal_atmos_products.sh[101]last=563 -+ exglobal_atmos_products.sh[102][[ 563 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 563 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 18 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 533:563 -grib tmpfile_f025_18 -533:9415236:d=2021032312:UGRD:975 mb:25 hour fcst: -534:9436596:d=2021032312:VGRD:975 mb:25 hour fcst: -535:9458469:d=2021032312:ABSV:975 mb:25 hour fcst: -536:9479451:d=2021032312:CLMR:975 mb:25 hour fcst: -537:9484111:d=2021032312:ICMR:975 mb:25 hour fcst: -538:9489446:d=2021032312:RWMR:975 mb:25 hour fcst: -539:9501145:d=2021032312:SNMR:975 mb:25 hour fcst: -540:9508189:d=2021032312:GRLE:975 mb:25 hour fcst: -541:9510933:d=2021032312:O3MR:975 mb:25 hour fcst: -542:9536807:d=2021032312:TMP:1000 mb:25 hour fcst: -543:9557579:d=2021032312:RH:1000 mb:25 hour fcst: -544:9576439:d=2021032312:TCDC:1000 mb:25 hour fcst: -545:9580900:d=2021032312:SPFH:1000 mb:25 hour fcst: -546:9607866:d=2021032312:VVEL:1000 mb:25 hour fcst: -547:9628989:d=2021032312:DZDT:1000 mb:25 hour fcst: -548:9650332:d=2021032312:UGRD:1000 mb:25 hour fcst: -549:9671122:d=2021032312:VGRD:1000 mb:25 hour fcst: -550:9692554:d=2021032312:ABSV:1000 mb:25 hour fcst: -551:9713127:d=2021032312:CLMR:1000 mb:25 hour fcst: -552:9715010:d=2021032312:ICMR:1000 mb:25 hour fcst: -553:9717473:d=2021032312:RWMR:1000 mb:25 hour fcst: -554:9727111:d=2021032312:SNMR:1000 mb:25 hour fcst: -555:9730543:d=2021032312:GRLE:1000 mb:25 hour fcst: -556:9731721:d=2021032312:O3MR:1000 mb:25 hour fcst: -557:9757586:d=2021032312:MSLET:mean sea level:25 hour fcst: -558:9784190:d=2021032312:HGT:1000 mb:25 hour fcst: -559:9809864:d=2021032312:SDEN:surface:25 hour fcst: -560:9820866:d=2021032312:REFD:4000 m above ground:25 hour fcst: -561:9833897:d=2021032312:REFD:1000 m above ground:25 hour fcst: -562:9858055:d=2021032312:PRES:surface:25 hour fcst: -563:9883663:d=2021032312:HGT:surface:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_18 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_18 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_18 pgb2file_f025_18 0p25' -+ exglobal_atmos_products.sh[133][[ 563 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=564 -+ exglobal_atmos_products.sh[101]last=594 -+ exglobal_atmos_products.sh[102][[ 594 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 594 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 19 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 564:594 -grib tmpfile_f025_19 -564:9899733:d=2021032312:TMP:surface:25 hour fcst: -565:9916376:d=2021032312:TSOIL:0-0.1 m below ground:25 hour fcst: -566:9928300:d=2021032312:SOILW:0-0.1 m below ground:25 hour fcst: -567:9937214:d=2021032312:SOILL:0-0.1 m below ground:25 hour fcst: -568:9946020:d=2021032312:TSOIL:0.1-0.4 m below ground:25 hour fcst: -569:9961496:d=2021032312:SOILW:0.1-0.4 m below ground:25 hour fcst: -570:9970512:d=2021032312:SOILL:0.1-0.4 m below ground:25 hour fcst: -571:9979406:d=2021032312:TSOIL:0.4-1 m below ground:25 hour fcst: -572:9993852:d=2021032312:SOILW:0.4-1 m below ground:25 hour fcst: -573:10002989:d=2021032312:SOILL:0.4-1 m below ground:25 hour fcst: -574:10011908:d=2021032312:TSOIL:1-2 m below ground:25 hour fcst: -575:10026194:d=2021032312:SOILW:1-2 m below ground:25 hour fcst: -576:10035257:d=2021032312:SOILL:1-2 m below ground:25 hour fcst: -577:10044335:d=2021032312:CNWAT:surface:25 hour fcst: -578:10049879:d=2021032312:WEASD:surface:25 hour fcst: -579:10062692:d=2021032312:SNOD:surface:25 hour fcst: -580:10077054:d=2021032312:PEVPR:surface:25 hour fcst: -581:10092822:d=2021032312:ICETK:surface:25 hour fcst: -582:10096239:d=2021032312:TMP:2 m above ground:25 hour fcst: -583:10117040:d=2021032312:SPFH:2 m above ground:25 hour fcst: -584:10145958:d=2021032312:DPT:2 m above ground:25 hour fcst: -585:10167224:d=2021032312:RH:2 m above ground:25 hour fcst: -586:10185756:d=2021032312:APTMP:2 m above ground:25 hour fcst: -587:10200154:d=2021032312:TMAX:2 m above ground:24-25 hour max fcst: -588:10220913:d=2021032312:TMIN:2 m above ground:24-25 hour min fcst: -589:10241733:d=2021032312:UGRD:10 m above ground:25 hour fcst: -590:10262244:d=2021032312:VGRD:10 m above ground:25 hour fcst: -591:10283142:d=2021032312:ICEG:10 m above mean sea level:25 hour fcst: -592:10283777:d=2021032312:CPOFP:surface:25 hour fcst: -593:10304150:d=2021032312:CPRAT:surface:25 hour fcst: -594:10325651:d=2021032312:PRATE:surface:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_19 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_19 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_19 pgb2file_f025_19 0p25' -+ exglobal_atmos_products.sh[133][[ 594 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=595 -+ exglobal_atmos_products.sh[101]last=625 -+ exglobal_atmos_products.sh[102][[ 625 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 625 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 20 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 595:625 -grib tmpfile_f025_20 -595:10352601:d=2021032312:CPRAT:surface:24-25 hour ave fcst: -596:10376405:d=2021032312:PRATE:surface:24-25 hour ave fcst: -597:10401720:d=2021032312:APCP:surface:24-25 hour acc fcst: -598:10407905:d=2021032312:APCP:surface:0-25 hour acc fcst: -599:10421179:d=2021032312:ACPCP:surface:24-25 hour acc fcst: -600:10426170:d=2021032312:ACPCP:surface:0-25 hour acc fcst: -601:10437216:d=2021032312:FROZR:surface:0-25 hour acc fcst: -602:10443567:d=2021032312:FRZR:surface:0-25 hour acc fcst: -603:10446709:d=2021032312:TSNOWP:surface:0-25 hour acc fcst: -604:10454317:d=2021032312:WATR:surface:24-25 hour acc fcst: -605:10467060:d=2021032312:FROZR:surface:24-25 hour acc fcst: -606:10470000:d=2021032312:FRZR:surface:24-25 hour acc fcst: -607:10471149:d=2021032312:TSNOWP:surface:24-25 hour acc fcst: -608:10476711:d=2021032312:CSNOW:surface:25 hour fcst: -609:10478511:d=2021032312:CICEP:surface:25 hour fcst: -610:10478690:d=2021032312:CFRZR:surface:25 hour fcst: -611:10478910:d=2021032312:CRAIN:surface:25 hour fcst: -612:10482416:d=2021032312:LHTFL:surface:24-25 hour ave fcst: -613:10504222:d=2021032312:SHTFL:surface:24-25 hour ave fcst: -614:10525218:d=2021032312:GFLUX:surface:24-25 hour ave fcst: -615:10537290:d=2021032312:UFLX:surface:24-25 hour ave fcst: -616:10553691:d=2021032312:VFLX:surface:24-25 hour ave fcst: -617:10570207:d=2021032312:SFCR:surface:25 hour fcst: -618:10588630:d=2021032312:FRICV:surface:25 hour fcst: -619:10610367:d=2021032312:U-GWD:surface:24-25 hour ave fcst: -620:10625239:d=2021032312:V-GWD:surface:24-25 hour ave fcst: -621:10640179:d=2021032312:VEG:surface:25 hour fcst: -622:10648213:d=2021032312:SOTYP:surface:25 hour fcst: -623:10659262:d=2021032312:WILT:surface:25 hour fcst: -624:10669172:d=2021032312:FLDCP:surface:25 hour fcst: -625:10679378:d=2021032312:SUNSD:surface:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_20 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_20 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_20 pgb2file_f025_20 0p25' -+ exglobal_atmos_products.sh[133][[ 625 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=626 -+ exglobal_atmos_products.sh[101]last=656 -+ exglobal_atmos_products.sh[102][[ 656 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 656 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 21 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 626:656 -grib tmpfile_f025_21 -626:10686541:d=2021032312:LFTX:surface:25 hour fcst: -627:10700368:d=2021032312:CAPE:surface:25 hour fcst: -628:10715548:d=2021032312:CIN:surface:25 hour fcst: -629:10728412:d=2021032312:PWAT:entire atmosphere (considered as a single layer):25 hour fcst: -630:10755353:d=2021032312:CWAT:entire atmosphere (considered as a single layer):25 hour fcst: -631:10781867:d=2021032312:RH:entire atmosphere (considered as a single layer):25 hour fcst: -632:10796985:d=2021032312:TOZNE:entire atmosphere (considered as a single layer):25 hour fcst: -633:10811324:d=2021032312:LCDC:low cloud layer:25 hour fcst: -634:10829204:d=2021032312:LCDC:low cloud layer:24-25 hour ave fcst: -635:10847021:d=2021032312:MCDC:middle cloud layer:25 hour fcst: -636:10862846:d=2021032312:MCDC:middle cloud layer:24-25 hour ave fcst: -637:10878311:d=2021032312:HCDC:high cloud layer:25 hour fcst: -638:10893114:d=2021032312:HCDC:high cloud layer:24-25 hour ave fcst: -639:10908957:d=2021032312:TCDC:entire atmosphere (considered as a single layer):25 hour fcst: -640:10926161:d=2021032312:TCDC:entire atmosphere (considered as a single layer):24-25 hour ave fcst: -641:10943391:d=2021032312:HGT:cloud ceiling:25 hour fcst: -642:10975484:d=2021032312:PRES:convective cloud bottom level:25 hour fcst: -643:10989142:d=2021032312:PRES:low cloud bottom level:24-25 hour ave fcst: -644:11010230:d=2021032312:PRES:middle cloud bottom level:24-25 hour ave fcst: -645:11026520:d=2021032312:PRES:high cloud bottom level:24-25 hour ave fcst: -646:11043290:d=2021032312:PRES:convective cloud top level:25 hour fcst: -647:11058653:d=2021032312:PRES:low cloud top level:24-25 hour ave fcst: -648:11079741:d=2021032312:PRES:middle cloud top level:24-25 hour ave fcst: -649:11095852:d=2021032312:PRES:high cloud top level:24-25 hour ave fcst: -650:11113147:d=2021032312:TMP:low cloud top level:24-25 hour ave fcst: -651:11129030:d=2021032312:TMP:middle cloud top level:24-25 hour ave fcst: -652:11142075:d=2021032312:TMP:high cloud top level:24-25 hour ave fcst: -653:11156443:d=2021032312:TCDC:convective cloud layer:25 hour fcst: -654:11177078:d=2021032312:TCDC:boundary layer cloud layer:24-25 hour ave fcst: -655:11192434:d=2021032312:CWORK:entire atmosphere (considered as a single layer):24-25 hour ave fcst: -656:11204983:d=2021032312:DSWRF:surface:24-25 hour ave fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_21 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_21 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_21 pgb2file_f025_21 0p25' -+ exglobal_atmos_products.sh[133][[ 656 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=657 -+ exglobal_atmos_products.sh[101]last=687 -+ exglobal_atmos_products.sh[102][[ 687 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 687 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -687:11918670:d=2021032312:UGRD:30 m above ground:25 hour fcst: -+ exglobal_atmos_products.sh[112]rc=0 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 0 == 0 ]] -+ exglobal_atmos_products.sh[115]last=688 -+ exglobal_atmos_products.sh[117][[ 22 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 657:688 -grib tmpfile_f025_22 -657:11222840:d=2021032312:DLWRF:surface:24-25 hour ave fcst: -658:11244551:d=2021032312:USWRF:surface:24-25 hour ave fcst: -659:11261843:d=2021032312:ULWRF:surface:24-25 hour ave fcst: -660:11280797:d=2021032312:USWRF:top of atmosphere:24-25 hour ave fcst: -661:11298881:d=2021032312:ULWRF:top of atmosphere:24-25 hour ave fcst: -662:11320027:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Total Aerosol:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -663:11347310:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Dust Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -664:11365429:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Sea Salt Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -665:11392799:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Sulphate Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -666:11418478:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Particulate Organic Matter Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -667:11444105:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Black Carbon Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -668:11466378:d=2021032312:HLCY:3000-0 m above ground:25 hour fcst: -669:11482242:d=2021032312:HLCY:1000-0 m above ground:25 hour fcst: -670:11504498:d=2021032312:USTM:6000-0 m above ground:25 hour fcst: -671:11527589:d=2021032312:VSTM:6000-0 m above ground:25 hour fcst: -672:11550421:d=2021032312:PRES:tropopause:25 hour fcst: -673:11580220:d=2021032312:ICAHT:tropopause:25 hour fcst: -674:11611049:d=2021032312:HGT:tropopause:25 hour fcst: -675:11639266:d=2021032312:TMP:tropopause:25 hour fcst: -676:11661077:d=2021032312:UGRD:tropopause:25 hour fcst: -677:11677066:d=2021032312:VGRD:tropopause:25 hour fcst: -678:11692960:d=2021032312:VWSH:tropopause:25 hour fcst: -679:11711468:d=2021032312:PRES:max wind:25 hour fcst: -680:11745956:d=2021032312:ICAHT:max wind:25 hour fcst: -681:11781158:d=2021032312:HGT:max wind:25 hour fcst: -682:11816509:d=2021032312:UGRD:max wind:25 hour fcst: -683:11832963:d=2021032312:VGRD:max wind:25 hour fcst: -684:11850411:d=2021032312:TMP:max wind:25 hour fcst: -685:11877160:d=2021032312:UGRD:20 m above ground:25 hour fcst: -686:11897710:d=2021032312:VGRD:20 m above ground:25 hour fcst: -687:11918670:d=2021032312:UGRD:30 m above ground:25 hour fcst: -688:11939560:d=2021032312:VGRD:30 m above ground:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_22 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_22 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_22 pgb2file_f025_22 0p25' -+ exglobal_atmos_products.sh[133][[ 688 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=689 -+ exglobal_atmos_products.sh[101]last=719 -+ exglobal_atmos_products.sh[102][[ 719 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 719 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 23 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 689:719 -grib tmpfile_f025_23 -689:11960880:d=2021032312:UGRD:40 m above ground:25 hour fcst: -690:11981877:d=2021032312:VGRD:40 m above ground:25 hour fcst: -691:12003272:d=2021032312:UGRD:50 m above ground:25 hour fcst: -692:12024391:d=2021032312:VGRD:50 m above ground:25 hour fcst: -693:12045899:d=2021032312:TMP:80 m above ground:25 hour fcst: -694:12066675:d=2021032312:SPFH:80 m above ground:25 hour fcst: -695:12095372:d=2021032312:PRES:80 m above ground:25 hour fcst: -696:12120954:d=2021032312:UGRD:80 m above ground:25 hour fcst: -697:12142225:d=2021032312:VGRD:80 m above ground:25 hour fcst: -698:12163929:d=2021032312:TMP:100 m above ground:25 hour fcst: -699:12184681:d=2021032312:UGRD:100 m above ground:25 hour fcst: -700:12206023:d=2021032312:VGRD:100 m above ground:25 hour fcst: -701:12227831:d=2021032312:TMP:1829 m above mean sea level:25 hour fcst: -702:12249477:d=2021032312:UGRD:1829 m above mean sea level:25 hour fcst: -703:12271723:d=2021032312:VGRD:1829 m above mean sea level:25 hour fcst: -704:12294623:d=2021032312:TMP:2743 m above mean sea level:25 hour fcst: -705:12316270:d=2021032312:UGRD:2743 m above mean sea level:25 hour fcst: -706:12338931:d=2021032312:VGRD:2743 m above mean sea level:25 hour fcst: -707:12362245:d=2021032312:TMP:3658 m above mean sea level:25 hour fcst: -708:12383535:d=2021032312:UGRD:3658 m above mean sea level:25 hour fcst: -709:12406580:d=2021032312:VGRD:3658 m above mean sea level:25 hour fcst: -710:12430366:d=2021032312:HGT:0C isotherm:25 hour fcst: -711:12452010:d=2021032312:RH:0C isotherm:25 hour fcst: -712:12471248:d=2021032312:HGT:highest tropospheric freezing level:25 hour fcst: -713:12492736:d=2021032312:RH:highest tropospheric freezing level:25 hour fcst: -714:12511746:d=2021032312:TMP:30-0 mb above ground:25 hour fcst: -715:12532564:d=2021032312:RH:30-0 mb above ground:25 hour fcst: -716:12550759:d=2021032312:SPFH:30-0 mb above ground:25 hour fcst: -717:12579584:d=2021032312:UGRD:30-0 mb above ground:25 hour fcst: -718:12600746:d=2021032312:VGRD:30-0 mb above ground:25 hour fcst: -719:12622529:d=2021032312:4LFTX:surface:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_23 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_23 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_23 pgb2file_f025_23 0p25' -+ exglobal_atmos_products.sh[133][[ 719 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=720 -+ exglobal_atmos_products.sh[101]last=750 -+ exglobal_atmos_products.sh[102][[ 750 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 750 tmpfile_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 24 -eq 24 ]] -+ exglobal_atmos_products.sh[118]last=753 -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f025 -for 720:753 -grib tmpfile_f025_24 -720:12636077:d=2021032312:CAPE:180-0 mb above ground:25 hour fcst: -721:12650055:d=2021032312:CIN:180-0 mb above ground:25 hour fcst: -722:12664095:d=2021032312:HPBL:surface:25 hour fcst: -723:12695126:d=2021032312:RH:0.33-1 sigma layer:25 hour fcst: -724:12713498:d=2021032312:RH:0.44-1 sigma layer:25 hour fcst: -725:12731495:d=2021032312:RH:0.72-0.94 sigma layer:25 hour fcst: -726:12749936:d=2021032312:RH:0.44-0.72 sigma layer:25 hour fcst: -727:12769271:d=2021032312:TMP:0.995 sigma level:25 hour fcst: -728:12790107:d=2021032312:POT:0.995 sigma level:25 hour fcst: -729:12810935:d=2021032312:RH:0.995 sigma level:25 hour fcst: -730:12828724:d=2021032312:UGRD:0.995 sigma level:25 hour fcst: -731:12849916:d=2021032312:VGRD:0.995 sigma level:25 hour fcst: -732:12871442:d=2021032312:VVEL:0.995 sigma level:25 hour fcst: -733:12892160:d=2021032312:CAPE:90-0 mb above ground:25 hour fcst: -734:12904944:d=2021032312:CIN:90-0 mb above ground:25 hour fcst: -735:12919032:d=2021032312:CAPE:255-0 mb above ground:25 hour fcst: -736:12932379:d=2021032312:CIN:255-0 mb above ground:25 hour fcst: -737:12943774:d=2021032312:PLPL:255-0 mb above ground:25 hour fcst: -738:12967726:d=2021032312:LAND:surface:25 hour fcst: -739:12969667:d=2021032312:ICEC:surface:25 hour fcst: -740:12972836:d=2021032312:ALBDO:surface:24-25 hour ave fcst: -741:12987320:d=2021032312:ICETMP:surface:25 hour fcst: -742:12993999:d=2021032312:UGRD:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -743:13004726:d=2021032312:VGRD:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -744:13015442:d=2021032312:TMP:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -745:13029787:d=2021032312:HGT:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -746:13047146:d=2021032312:PRES:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -747:13065515:d=2021032312:VWSH:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -748:13077434:d=2021032312:UGRD:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -749:13088312:d=2021032312:VGRD:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -750:13103093:d=2021032312:TMP:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -751:13117505:d=2021032312:HGT:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -752:13135066:d=2021032312:PRES:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -753:13153370:d=2021032312:VWSH:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f025_24 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f025_24 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_24 pgb2file_f025_24 0p25' -+ exglobal_atmos_products.sh[133][[ 753 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( pproc = iproc+1 )) -+ exglobal_atmos_products.sh[134](( pproc < nproc )) -+ exglobal_atmos_products.sh[137]break -+ exglobal_atmos_products.sh[142]/work2/noaa/global/mterry/global-workflow_forked/ush/run_mpmd.sh /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601/poescript -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ run_mpmd.sh[31]source /work2/noaa/global/mterry/global-workflow_forked/ush/preamble.sh -++ preamble.sh[20]set +x -Begin run_mpmd.sh at Tue Jul 29 02:26:33 UTC 2025 -+ run_mpmd.sh[33]cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601/poescript -+ run_mpmd.sh[36][[ YES != \Y\E\S ]] -+ run_mpmd.sh[46]export OMP_NUM_THREADS=1 -+ run_mpmd.sh[46]OMP_NUM_THREADS=1 -++ run_mpmd.sh[49]wc -l -+ run_mpmd.sh[49]nprocs=24 -+ run_mpmd.sh[52]mpmd_cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601/mpmd_cmdfile -+ run_mpmd.sh[53][[ -s /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601/mpmd_cmdfile ]] -+ run_mpmd.sh[55]cat - INFO: Executing MPMD job, STDOUT redirected for each process separately - INFO: On failure, logs for each job will be available in /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601/mpmd.proc_num.out - INFO: The proc_num corresponds to the line in '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601/mpmd_cmdfile' -+ run_mpmd.sh[61][[ srun -l --export=ALL --hint=nomultithread =~ ^srun.* ]] -+ run_mpmd.sh[65]nm=0 -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '0 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_1 pgb2file_f025_1 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '1 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_2 pgb2file_f025_2 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '2 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_3 pgb2file_f025_3 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '3 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_4 pgb2file_f025_4 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '4 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_5 pgb2file_f025_5 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '5 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_6 pgb2file_f025_6 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '6 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_7 pgb2file_f025_7 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '7 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_8 pgb2file_f025_8 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '8 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_9 pgb2file_f025_9 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '9 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_10 pgb2file_f025_10 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '10 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_11 pgb2file_f025_11 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '11 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_12 pgb2file_f025_12 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '12 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_13 pgb2file_f025_13 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '13 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_14 pgb2file_f025_14 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '14 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_15 pgb2file_f025_15 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '15 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_16 pgb2file_f025_16 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '16 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_17 pgb2file_f025_17 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '17 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_18 pgb2file_f025_18 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '18 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_19 pgb2file_f025_19 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '19 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_20 pgb2file_f025_20 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '20 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_21 pgb2file_f025_21 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '21 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_22 pgb2file_f025_22 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '22 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_23 pgb2file_f025_23 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '23 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f025_24 pgb2file_f025_24 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[72]set +e -+ run_mpmd.sh[74]srun -l --export=ALL --hint=nomultithread --multi-prog --output=mpmd.%j.%t.out -n 24 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601/mpmd_cmdfile -+ run_mpmd.sh[75]err=0 -+ run_mpmd.sh[76]set_strict -+ preamble.sh[35][[ YES == \Y\E\S ]] -+ preamble.sh[37]set -eu -+ run_mpmd.sh[103][[ 0 -eq 0 ]] -+ run_mpmd.sh[104]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601/mpmd_cmdfile -++ run_mpmd.sh[105]find . -name 'mpmd.*.out' -+ run_mpmd.sh[105]out_files='./mpmd.5951061.17.out -./mpmd.5951061.0.out -./mpmd.5951061.6.out -./mpmd.5951061.15.out -./mpmd.5951061.9.out -./mpmd.5951061.2.out -./mpmd.5951061.3.out -./mpmd.5951061.23.out -./mpmd.5951061.4.out -./mpmd.5951061.22.out -./mpmd.5951061.1.out -./mpmd.5951061.7.out -./mpmd.5951061.21.out -./mpmd.5951061.12.out -./mpmd.5951061.19.out -./mpmd.5951061.8.out -./mpmd.5951061.16.out -./mpmd.5951061.18.out -./mpmd.5951061.13.out -./mpmd.5951061.10.out -./mpmd.5951061.20.out -./mpmd.5951061.14.out -./mpmd.5951061.5.out -./mpmd.5951061.11.out' -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.17.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.17.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.0.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.0.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.6.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.6.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.15.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.15.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.9.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.9.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.2.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.2.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.3.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.3.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.23.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.23.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.4.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.4.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.22.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.22.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.1.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.1.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.7.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.7.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.21.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.21.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.12.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.12.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.19.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.19.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.8.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.8.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.16.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.16.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.18.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.18.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.13.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.13.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.10.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.10.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.20.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.20.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.14.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.14.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.5.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.5.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.11.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.11.out -+ run_mpmd.sh[110]cat mpmd.out -17: + bash[8]'[' -z '' ']' -17: + bash[9]case "$-" in -17: + bash[12]__lmod_vx=x -17: + bash[16]'[' -n x ']' -17: + bash[16]set +x -17: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -17: Shell debugging restarted -17: + bash[224]unset __lmod_vx -17: + interp_atmos_master.sh[7]input_file=tmpfile_f025_18 -17: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_18 -17: + interp_atmos_master.sh[9]grid_string=0p25 -17: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -17: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -17: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -17: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -17: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -17: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -17: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -17: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -17: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -17: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -17: + interp_atmos_master.sh[31]IFS=: -17: + interp_atmos_master.sh[31]read -ra grids -17: + interp_atmos_master.sh[33]output_grids= -17: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -17: + interp_atmos_master.sh[35]gridopt=grid0p25 -17: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_18_0p25' -17: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_18 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_18_0p25 -17: 1:0:d=2021032312:UGRD:975 mb:25 hour fcst: -17: 2:21360:d=2021032312:VGRD:975 mb:25 hour fcst: -17: 3:43233:d=2021032312:ABSV:975 mb:25 hour fcst: -17: 4:64215:d=2021032312:CLMR:975 mb:25 hour fcst: -17: 5:68875:d=2021032312:ICMR:975 mb:25 hour fcst: -17: 6:74210:d=2021032312:RWMR:975 mb:25 hour fcst: -17: 7:85909:d=2021032312:SNMR:975 mb:25 hour fcst: -17: 8:92953:d=2021032312:GRLE:975 mb:25 hour fcst: -17: 9:95697:d=2021032312:O3MR:975 mb:25 hour fcst: -17: 10:121571:d=2021032312:TMP:1000 mb:25 hour fcst: -17: 11:142343:d=2021032312:RH:1000 mb:25 hour fcst: -17: 12:161203:d=2021032312:TCDC:1000 mb:25 hour fcst: -17: 13:165664:d=2021032312:SPFH:1000 mb:25 hour fcst: -17: 14:192630:d=2021032312:VVEL:1000 mb:25 hour fcst: -17: 15:213753:d=2021032312:DZDT:1000 mb:25 hour fcst: -17: 16:235096:d=2021032312:UGRD:1000 mb:25 hour fcst: -17: 17:255886:d=2021032312:VGRD:1000 mb:25 hour fcst: -17: 18:277318:d=2021032312:ABSV:1000 mb:25 hour fcst: -17: 19:297891:d=2021032312:CLMR:1000 mb:25 hour fcst: -17: 20:299774:d=2021032312:ICMR:1000 mb:25 hour fcst: -17: 21:302237:d=2021032312:RWMR:1000 mb:25 hour fcst: -17: 22:311875:d=2021032312:SNMR:1000 mb:25 hour fcst: -17: 23:315307:d=2021032312:GRLE:1000 mb:25 hour fcst: -17: 24:316485:d=2021032312:O3MR:1000 mb:25 hour fcst: -17: 25:342350:d=2021032312:MSLET:mean sea level:25 hour fcst: -17: 26:368954:d=2021032312:HGT:1000 mb:25 hour fcst: -17: 27:394628:d=2021032312:SDEN:surface:25 hour fcst: -17: 28:405630:d=2021032312:REFD:4000 m above ground:25 hour fcst: -17: 29:418661:d=2021032312:REFD:1000 m above ground:25 hour fcst: -17: 30:442819:d=2021032312:PRES:surface:25 hour fcst: -17: 31:468427:d=2021032312:HGT:surface:25 hour fcst: -17: + interp_atmos_master.sh[47]export err=0 -17: + interp_atmos_master.sh[47]err=0 -17: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -17: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -17: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_18_0p25 -17: + product_functions.sh[5]local filename=pgb2file_f025_18_0p25 -17: + product_functions.sh[6]wgrib2 pgb2file_f025_18_0p25 -not_if :RH: -grib pgb2file_f025_18_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_18_0p25.new -17: 1:0:d=2021032312:UGRD:975 mb:25 hour fcst: -17: 2:518427:d=2021032312:VGRD:975 mb:25 hour fcst: -17: 3:1061732:d=2021032312:ABSV:975 mb:25 hour fcst: -17: 4:1562054:d=2021032312:CLMR:975 mb:25 hour fcst: -17: 5:1663777:d=2021032312:ICMR:975 mb:25 hour fcst: -17: 6:1783444:d=2021032312:RWMR:975 mb:25 hour fcst: -17: 7:2026365:d=2021032312:SNMR:975 mb:25 hour fcst: -17: 8:2182077:d=2021032312:GRLE:975 mb:25 hour fcst: -17: 9:2236751:d=2021032312:O3MR:975 mb:25 hour fcst: -17: 10:2935562:d=2021032312:TMP:1000 mb:25 hour fcst: -17: 11:3418942:d=2021032312:RH:1000 mb:25 hour fcst: -17: 12:3839317:d=2021032312:TCDC:1000 mb:25 hour fcst: -17: 13:3919782:d=2021032312:SPFH:1000 mb:25 hour fcst: -17: 14:4651551:d=2021032312:VVEL:1000 mb:25 hour fcst: -17: 15:5166095:d=2021032312:DZDT:1000 mb:25 hour fcst: -17: 16:5680153:d=2021032312:UGRD:1000 mb:25 hour fcst: -17: 17:6185419:d=2021032312:VGRD:1000 mb:25 hour fcst: -17: 18:6706842:d=2021032312:ABSV:1000 mb:25 hour fcst: -17: 19:7184100:d=2021032312:CLMR:1000 mb:25 hour fcst: -17: 20:7224691:d=2021032312:ICMR:1000 mb:25 hour fcst: -17: 21:7276093:d=2021032312:RWMR:1000 mb:25 hour fcst: -17: 22:7469643:d=2021032312:SNMR:1000 mb:25 hour fcst: -17: 23:7540104:d=2021032312:GRLE:1000 mb:25 hour fcst: -17: 24:7559355:d=2021032312:O3MR:1000 mb:25 hour fcst: -17: 25:8256308:d=2021032312:MSLET:mean sea level:25 hour fcst: -17: 26:8976885:d=2021032312:HGT:1000 mb:25 hour fcst: -17: 27:9663803:d=2021032312:SDEN:surface:25 hour fcst: -17: 28:10005025:d=2021032312:REFD:4000 m above ground:25 hour fcst: -17: 29:10279449:d=2021032312:REFD:1000 m above ground:25 hour fcst: -17: 30:10847592:d=2021032312:PRES:surface:25 hour fcst: -17: 31:11491977:d=2021032312:HGT:surface:25 hour fcst: -17: + product_functions.sh[10]rc=0 -17: + product_functions.sh[11](( rc == 0 )) -17: + product_functions.sh[11]mv pgb2file_f025_18_0p25.new pgb2file_f025_18_0p25 -17: + product_functions.sh[12]return 0 -17: + interp_atmos_master.sh[56]export err=0 -17: + interp_atmos_master.sh[56]err=0 -17: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -17: ++ interp_atmos_master.sh[62]wc -l -17: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_18_0p25 -match 'LAND|ICEC' -17: + interp_atmos_master.sh[62]var_count=0 -17: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -17: + interp_atmos_master.sh[73]exit 0 - 0: + bash[8]'[' -z '' ']' - 0: + bash[9]case "$-" in - 0: + bash[12]__lmod_vx=x - 0: + bash[16]'[' -n x ']' - 0: + bash[16]set +x - 0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 0: Shell debugging restarted - 0: + bash[224]unset __lmod_vx - 0: + interp_atmos_master.sh[7]input_file=tmpfile_f025_1 - 0: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_1 - 0: + interp_atmos_master.sh[9]grid_string=0p25 - 0: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 0: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 0: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 0: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 0: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 0: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 0: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 0: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 0: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 0: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 0: + interp_atmos_master.sh[31]IFS=: - 0: + interp_atmos_master.sh[31]read -ra grids - 0: + interp_atmos_master.sh[33]output_grids= - 0: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 0: + interp_atmos_master.sh[35]gridopt=grid0p25 - 0: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_1_0p25' - 0: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_1 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_1_0p25 - 0: 1:0:d=2021032312:PRMSL:mean sea level:25 hour fcst: - 0: 2:25688:d=2021032312:CLMR:1 hybrid level:25 hour fcst: - 0: 3:29343:d=2021032312:ICMR:1 hybrid level:25 hour fcst: - 0: 4:36295:d=2021032312:RWMR:1 hybrid level:25 hour fcst: - 0: 5:48732:d=2021032312:SNMR:1 hybrid level:25 hour fcst: - 0: 6:58824:d=2021032312:GRLE:1 hybrid level:25 hour fcst: - 0: 7:61276:d=2021032312:REFD:1 hybrid level:25 hour fcst: - 0: 8:87031:d=2021032312:REFD:2 hybrid level:25 hour fcst: - 0: 9:112761:d=2021032312:REFC:entire atmosphere (considered as a single layer):25 hour fcst: - 0: 10:140221:d=2021032312:VIS:surface:25 hour fcst: - 0: 11:172542:d=2021032312:UGRD:planetary boundary layer:25 hour fcst: - 0: 12:186396:d=2021032312:VGRD:planetary boundary layer:25 hour fcst: - 0: 13:200985:d=2021032312:VRATE:planetary boundary layer:25 hour fcst: - 0: 14:216099:d=2021032312:GUST:surface:25 hour fcst: - 0: 15:230805:d=2021032312:HGT:0.01 mb:25 hour fcst: - 0: 16:254189:d=2021032312:TMP:0.01 mb:25 hour fcst: - 0: 17:264779:d=2021032312:RH:0.01 mb:25 hour fcst: - 0: 18:271064:d=2021032312:SPFH:0.01 mb:25 hour fcst: - 0: 19:294006:d=2021032312:VVEL:0.01 mb:25 hour fcst: - 0: 20:320813:d=2021032312:DZDT:0.01 mb:25 hour fcst: - 0: 21:348111:d=2021032312:UGRD:0.01 mb:25 hour fcst: - 0: 22:359231:d=2021032312:VGRD:0.01 mb:25 hour fcst: - 0: 23:369764:d=2021032312:ABSV:0.01 mb:25 hour fcst: - 0: 24:385189:d=2021032312:O3MR:0.01 mb:25 hour fcst: - 0: 25:399140:d=2021032312:HGT:0.02 mb:25 hour fcst: - 0: 26:424624:d=2021032312:TMP:0.02 mb:25 hour fcst: - 0: 27:439277:d=2021032312:RH:0.02 mb:25 hour fcst: - 0: 28:446600:d=2021032312:SPFH:0.02 mb:25 hour fcst: - 0: 29:470664:d=2021032312:VVEL:0.02 mb:25 hour fcst: - 0: 30:494846:d=2021032312:DZDT:0.02 mb:25 hour fcst: - 0: 31:522503:d=2021032312:UGRD:0.02 mb:25 hour fcst: - 0: 32:534851:d=2021032312:VGRD:0.02 mb:25 hour fcst: - 0: + interp_atmos_master.sh[47]export err=0 - 0: + interp_atmos_master.sh[47]err=0 - 0: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 0: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 0: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_1_0p25 - 0: + product_functions.sh[5]local filename=pgb2file_f025_1_0p25 - 0: + product_functions.sh[6]wgrib2 pgb2file_f025_1_0p25 -not_if :RH: -grib pgb2file_f025_1_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_1_0p25.new - 0: 1:0:d=2021032312:PRMSL:mean sea level:25 hour fcst: - 0: 2:689051:d=2021032312:CLMR:1 hybrid level:25 hour fcst: - 0: 3:771628:d=2021032312:ICMR:1 hybrid level:25 hour fcst: - 0: 4:925439:d=2021032312:RWMR:1 hybrid level:25 hour fcst: - 0: 5:1180067:d=2021032312:SNMR:1 hybrid level:25 hour fcst: - 0: 6:1410959:d=2021032312:GRLE:1 hybrid level:25 hour fcst: - 0: 7:1453667:d=2021032312:REFD:1 hybrid level:25 hour fcst: - 0: 8:2053534:d=2021032312:REFD:2 hybrid level:25 hour fcst: - 0: 9:2653499:d=2021032312:REFC:entire atmosphere (considered as a single layer):25 hour fcst: - 0: 10:3310214:d=2021032312:VIS:surface:25 hour fcst: - 0: 11:4191368:d=2021032312:UGRD:planetary boundary layer:25 hour fcst: - 0: 12:4490334:d=2021032312:VGRD:planetary boundary layer:25 hour fcst: - 0: 13:4800338:d=2021032312:VRATE:planetary boundary layer:25 hour fcst: - 0: 14:5114904:d=2021032312:GUST:surface:25 hour fcst: - 0: 15:5424061:d=2021032312:HGT:0.01 mb:25 hour fcst: - 0: 16:6032883:d=2021032312:TMP:0.01 mb:25 hour fcst: - 0: 17:6301175:d=2021032312:RH:0.01 mb:25 hour fcst: - 0: 18:6486278:d=2021032312:SPFH:0.01 mb:25 hour fcst: - 0: 19:7085423:d=2021032312:VVEL:0.01 mb:25 hour fcst: - 0: 20:7826281:d=2021032312:DZDT:0.01 mb:25 hour fcst: - 0: 21:8609803:d=2021032312:UGRD:0.01 mb:25 hour fcst: - 0: 22:8885306:d=2021032312:VGRD:0.01 mb:25 hour fcst: - 0: 23:9153681:d=2021032312:ABSV:0.01 mb:25 hour fcst: - 0: 24:9478158:d=2021032312:O3MR:0.01 mb:25 hour fcst: - 0: 25:9776868:d=2021032312:HGT:0.02 mb:25 hour fcst: - 0: 26:10471675:d=2021032312:TMP:0.02 mb:25 hour fcst: - 0: 27:10812391:d=2021032312:RH:0.02 mb:25 hour fcst: - 0: 28:10990876:d=2021032312:SPFH:0.02 mb:25 hour fcst: - 0: 29:11638795:d=2021032312:VVEL:0.02 mb:25 hour fcst: - 0: 30:12290396:d=2021032312:DZDT:0.02 mb:25 hour fcst: - 0: 31:13086835:d=2021032312:UGRD:0.02 mb:25 hour fcst: - 0: 32:13379832:d=2021032312:VGRD:0.02 mb:25 hour fcst: - 0: + product_functions.sh[10]rc=0 - 0: + product_functions.sh[11](( rc == 0 )) - 0: + product_functions.sh[11]mv pgb2file_f025_1_0p25.new pgb2file_f025_1_0p25 - 0: + product_functions.sh[12]return 0 - 0: + interp_atmos_master.sh[56]export err=0 - 0: + interp_atmos_master.sh[56]err=0 - 0: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 0: ++ interp_atmos_master.sh[62]wc -l - 0: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_1_0p25 -match 'LAND|ICEC' - 0: + interp_atmos_master.sh[62]var_count=0 - 0: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 0: + interp_atmos_master.sh[73]exit 0 - 6: + bash[8]'[' -z '' ']' - 6: + bash[9]case "$-" in - 6: + bash[12]__lmod_vx=x - 6: + bash[16]'[' -n x ']' - 6: + bash[16]set +x - 6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 6: Shell debugging restarted - 6: + bash[224]unset __lmod_vx - 6: + interp_atmos_master.sh[7]input_file=tmpfile_f025_7 - 6: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_7 - 6: + interp_atmos_master.sh[9]grid_string=0p25 - 6: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 6: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 6: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 6: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 6: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 6: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 6: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 6: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 6: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 6: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 6: + interp_atmos_master.sh[31]IFS=: - 6: + interp_atmos_master.sh[31]read -ra grids - 6: + interp_atmos_master.sh[33]output_grids= - 6: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 6: + interp_atmos_master.sh[35]gridopt=grid0p25 - 6: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_7_0p25' - 6: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_7 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_7_0p25 - 6: 1:0:d=2021032312:SPFH:40 mb:25 hour fcst: - 6: 2:17034:d=2021032312:VVEL:40 mb:25 hour fcst: - 6: 3:46401:d=2021032312:DZDT:40 mb:25 hour fcst: - 6: 4:76093:d=2021032312:UGRD:40 mb:25 hour fcst: - 6: 5:94944:d=2021032312:VGRD:40 mb:25 hour fcst: - 6: 6:113651:d=2021032312:ABSV:40 mb:25 hour fcst: - 6: 7:131574:d=2021032312:O3MR:40 mb:25 hour fcst: - 6: 8:158536:d=2021032312:HGT:50 mb:25 hour fcst: - 6: 9:177481:d=2021032312:TMP:50 mb:25 hour fcst: - 6: 10:193613:d=2021032312:RH:50 mb:25 hour fcst: - 6: 11:204230:d=2021032312:TCDC:50 mb:25 hour fcst: - 6: 12:204409:d=2021032312:SPFH:50 mb:25 hour fcst: - 6: 13:222915:d=2021032312:VVEL:50 mb:25 hour fcst: - 6: 14:253100:d=2021032312:DZDT:50 mb:25 hour fcst: - 6: 15:285197:d=2021032312:UGRD:50 mb:25 hour fcst: - 6: 16:303745:d=2021032312:VGRD:50 mb:25 hour fcst: - 6: 17:322268:d=2021032312:ABSV:50 mb:25 hour fcst: - 6: 18:339987:d=2021032312:CLMR:50 mb:25 hour fcst: - 6: 19:340166:d=2021032312:ICMR:50 mb:25 hour fcst: - 6: 20:340345:d=2021032312:RWMR:50 mb:25 hour fcst: - 6: 21:340524:d=2021032312:SNMR:50 mb:25 hour fcst: - 6: 22:340703:d=2021032312:GRLE:50 mb:25 hour fcst: - 6: 23:340882:d=2021032312:O3MR:50 mb:25 hour fcst: - 6: 24:367227:d=2021032312:HGT:70 mb:25 hour fcst: - 6: 25:386237:d=2021032312:TMP:70 mb:25 hour fcst: - 6: 26:402643:d=2021032312:RH:70 mb:25 hour fcst: - 6: 27:410981:d=2021032312:SPFH:70 mb:25 hour fcst: - 6: 28:431895:d=2021032312:VVEL:70 mb:25 hour fcst: - 6: 29:463087:d=2021032312:DZDT:70 mb:25 hour fcst: - 6: 30:492774:d=2021032312:UGRD:70 mb:25 hour fcst: - 6: 31:511416:d=2021032312:VGRD:70 mb:25 hour fcst: - 6: + interp_atmos_master.sh[47]export err=0 - 6: + interp_atmos_master.sh[47]err=0 - 6: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 6: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 6: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_7_0p25 - 6: + product_functions.sh[5]local filename=pgb2file_f025_7_0p25 - 6: + product_functions.sh[6]wgrib2 pgb2file_f025_7_0p25 -not_if :RH: -grib pgb2file_f025_7_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_7_0p25.new - 6: 1:0:d=2021032312:SPFH:40 mb:25 hour fcst: - 6: 2:415424:d=2021032312:VVEL:40 mb:25 hour fcst: - 6: 3:1238034:d=2021032312:DZDT:40 mb:25 hour fcst: - 6: 4:2111779:d=2021032312:UGRD:40 mb:25 hour fcst: - 6: 5:2552829:d=2021032312:VGRD:40 mb:25 hour fcst: - 6: 6:2991200:d=2021032312:ABSV:40 mb:25 hour fcst: - 6: 7:3394614:d=2021032312:O3MR:40 mb:25 hour fcst: - 6: 8:4129398:d=2021032312:HGT:50 mb:25 hour fcst: - 6: 9:4567646:d=2021032312:TMP:50 mb:25 hour fcst: - 6: 10:4923181:d=2021032312:RH:50 mb:25 hour fcst: - 6: 11:5179672:d=2021032312:TCDC:50 mb:25 hour fcst: - 6: 12:5179851:d=2021032312:SPFH:50 mb:25 hour fcst: - 6: 13:5633769:d=2021032312:VVEL:50 mb:25 hour fcst: - 6: 14:6490606:d=2021032312:DZDT:50 mb:25 hour fcst: - 6: 15:7473543:d=2021032312:UGRD:50 mb:25 hour fcst: - 6: 16:7905844:d=2021032312:VGRD:50 mb:25 hour fcst: - 6: 17:8336008:d=2021032312:ABSV:50 mb:25 hour fcst: - 6: 18:8730384:d=2021032312:CLMR:50 mb:25 hour fcst: - 6: 19:8730563:d=2021032312:ICMR:50 mb:25 hour fcst: - 6: 20:8730742:d=2021032312:RWMR:50 mb:25 hour fcst: - 6: 21:8730921:d=2021032312:SNMR:50 mb:25 hour fcst: - 6: 22:8731100:d=2021032312:GRLE:50 mb:25 hour fcst: - 6: 23:8731279:d=2021032312:O3MR:50 mb:25 hour fcst: - 6: 24:9449971:d=2021032312:HGT:70 mb:25 hour fcst: - 6: 25:9895050:d=2021032312:TMP:70 mb:25 hour fcst: - 6: 26:10264721:d=2021032312:RH:70 mb:25 hour fcst: - 6: 27:10453060:d=2021032312:SPFH:70 mb:25 hour fcst: - 6: 28:10976540:d=2021032312:VVEL:70 mb:25 hour fcst: - 6: 29:11849323:d=2021032312:DZDT:70 mb:25 hour fcst: - 6: 30:12719493:d=2021032312:UGRD:70 mb:25 hour fcst: - 6: 31:13163930:d=2021032312:VGRD:70 mb:25 hour fcst: - 6: + product_functions.sh[10]rc=0 - 6: + product_functions.sh[11](( rc == 0 )) - 6: + product_functions.sh[11]mv pgb2file_f025_7_0p25.new pgb2file_f025_7_0p25 - 6: + product_functions.sh[12]return 0 - 6: + interp_atmos_master.sh[56]export err=0 - 6: + interp_atmos_master.sh[56]err=0 - 6: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 6: ++ interp_atmos_master.sh[62]wc -l - 6: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_7_0p25 -match 'LAND|ICEC' - 6: + interp_atmos_master.sh[62]var_count=0 - 6: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 6: + interp_atmos_master.sh[73]exit 0 -15: + bash[8]'[' -z '' ']' -15: + bash[9]case "$-" in -15: + bash[12]__lmod_vx=x -15: + bash[16]'[' -n x ']' -15: + bash[16]set +x -15: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -15: Shell debugging restarted -15: + bash[224]unset __lmod_vx -15: + interp_atmos_master.sh[7]input_file=tmpfile_f025_16 -15: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_16 -15: + interp_atmos_master.sh[9]grid_string=0p25 -15: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -15: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -15: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -15: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -15: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -15: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -15: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -15: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -15: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -15: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -15: + interp_atmos_master.sh[31]IFS=: -15: + interp_atmos_master.sh[31]read -ra grids -15: + interp_atmos_master.sh[33]output_grids= -15: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -15: + interp_atmos_master.sh[35]gridopt=grid0p25 -15: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_16_0p25' -15: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_16 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_16_0p25 -15: 1:0:d=2021032312:ABSV:850 mb:25 hour fcst: -15: 2:21258:d=2021032312:CLMR:850 mb:25 hour fcst: -15: 3:34685:d=2021032312:ICMR:850 mb:25 hour fcst: -15: 4:44398:d=2021032312:RWMR:850 mb:25 hour fcst: -15: 5:54903:d=2021032312:SNMR:850 mb:25 hour fcst: -15: 6:66540:d=2021032312:GRLE:850 mb:25 hour fcst: -15: 7:69175:d=2021032312:O3MR:850 mb:25 hour fcst: -15: 8:95774:d=2021032312:HGT:900 mb:25 hour fcst: -15: 9:120801:d=2021032312:TMP:900 mb:25 hour fcst: -15: 10:141783:d=2021032312:RH:900 mb:25 hour fcst: -15: 11:161607:d=2021032312:TCDC:900 mb:25 hour fcst: -15: 12:174712:d=2021032312:SPFH:900 mb:25 hour fcst: -15: 13:204184:d=2021032312:VVEL:900 mb:25 hour fcst: -15: 14:229063:d=2021032312:DZDT:900 mb:25 hour fcst: -15: 15:255167:d=2021032312:UGRD:900 mb:25 hour fcst: -15: 16:276369:d=2021032312:VGRD:900 mb:25 hour fcst: -15: 17:298261:d=2021032312:ABSV:900 mb:25 hour fcst: -15: 18:319445:d=2021032312:CLMR:900 mb:25 hour fcst: -15: 19:329482:d=2021032312:ICMR:900 mb:25 hour fcst: -15: 20:337864:d=2021032312:RWMR:900 mb:25 hour fcst: -15: 21:350077:d=2021032312:SNMR:900 mb:25 hour fcst: -15: 22:360296:d=2021032312:GRLE:900 mb:25 hour fcst: -15: 23:363124:d=2021032312:O3MR:900 mb:25 hour fcst: -15: 24:389159:d=2021032312:HGT:925 mb:25 hour fcst: -15: 25:414315:d=2021032312:TMP:925 mb:25 hour fcst: -15: 26:435493:d=2021032312:RH:925 mb:25 hour fcst: -15: 27:455257:d=2021032312:TCDC:925 mb:25 hour fcst: -15: 28:467170:d=2021032312:SPFH:925 mb:25 hour fcst: -15: 29:496710:d=2021032312:VVEL:925 mb:25 hour fcst: -15: 30:521135:d=2021032312:DZDT:925 mb:25 hour fcst: -15: 31:546736:d=2021032312:UGRD:925 mb:25 hour fcst: -15: 32:568002:d=2021032312:VGRD:925 mb:25 hour fcst: -15: + interp_atmos_master.sh[47]export err=0 -15: + interp_atmos_master.sh[47]err=0 -15: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -15: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -15: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_16_0p25 -15: + product_functions.sh[5]local filename=pgb2file_f025_16_0p25 -15: + product_functions.sh[6]wgrib2 pgb2file_f025_16_0p25 -not_if :RH: -grib pgb2file_f025_16_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_16_0p25.new -15: 1:0:d=2021032312:ABSV:850 mb:25 hour fcst: -15: 2:506582:d=2021032312:CLMR:850 mb:25 hour fcst: -15: 3:795274:d=2021032312:ICMR:850 mb:25 hour fcst: -15: 4:1031240:d=2021032312:RWMR:850 mb:25 hour fcst: -15: 5:1233537:d=2021032312:SNMR:850 mb:25 hour fcst: -15: 6:1496409:d=2021032312:GRLE:850 mb:25 hour fcst: -15: 7:1546693:d=2021032312:O3MR:850 mb:25 hour fcst: -15: 8:2271552:d=2021032312:HGT:900 mb:25 hour fcst: -15: 9:2920789:d=2021032312:TMP:900 mb:25 hour fcst: -15: 10:3416545:d=2021032312:RH:900 mb:25 hour fcst: -15: 11:3868857:d=2021032312:TCDC:900 mb:25 hour fcst: -15: 12:4133782:d=2021032312:SPFH:900 mb:25 hour fcst: -15: 13:4936992:d=2021032312:VVEL:900 mb:25 hour fcst: -15: 14:5600263:d=2021032312:DZDT:900 mb:25 hour fcst: -15: 15:6303302:d=2021032312:UGRD:900 mb:25 hour fcst: -15: 16:6818832:d=2021032312:VGRD:900 mb:25 hour fcst: -15: 17:7362528:d=2021032312:ABSV:900 mb:25 hour fcst: -15: 18:7867172:d=2021032312:CLMR:900 mb:25 hour fcst: -15: 19:8089528:d=2021032312:ICMR:900 mb:25 hour fcst: -15: 20:8286045:d=2021032312:RWMR:900 mb:25 hour fcst: -15: 21:8529586:d=2021032312:SNMR:900 mb:25 hour fcst: -15: 22:8765518:d=2021032312:GRLE:900 mb:25 hour fcst: -15: 23:8818895:d=2021032312:O3MR:900 mb:25 hour fcst: -15: 24:9526544:d=2021032312:HGT:925 mb:25 hour fcst: -15: 25:10182003:d=2021032312:TMP:925 mb:25 hour fcst: -15: 26:10682636:d=2021032312:RH:925 mb:25 hour fcst: -15: 27:11134413:d=2021032312:TCDC:925 mb:25 hour fcst: -15: 28:11368263:d=2021032312:SPFH:925 mb:25 hour fcst: -15: 29:12193290:d=2021032312:VVEL:925 mb:25 hour fcst: -15: 30:12840592:d=2021032312:DZDT:925 mb:25 hour fcst: -15: 31:13530294:d=2021032312:UGRD:925 mb:25 hour fcst: -15: 32:14048976:d=2021032312:VGRD:925 mb:25 hour fcst: -15: + product_functions.sh[10]rc=0 -15: + product_functions.sh[11](( rc == 0 )) -15: + product_functions.sh[11]mv pgb2file_f025_16_0p25.new pgb2file_f025_16_0p25 -15: + product_functions.sh[12]return 0 -15: + interp_atmos_master.sh[56]export err=0 -15: + interp_atmos_master.sh[56]err=0 -15: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -15: ++ interp_atmos_master.sh[62]wc -l -15: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_16_0p25 -match 'LAND|ICEC' -15: + interp_atmos_master.sh[62]var_count=0 -15: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -15: + interp_atmos_master.sh[73]exit 0 - 9: + bash[8]'[' -z '' ']' - 9: + bash[9]case "$-" in - 9: + bash[12]__lmod_vx=x - 9: + bash[16]'[' -n x ']' - 9: + bash[16]set +x - 9: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 9: Shell debugging restarted - 9: + bash[224]unset __lmod_vx - 9: + interp_atmos_master.sh[7]input_file=tmpfile_f025_10 - 9: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_10 - 9: + interp_atmos_master.sh[9]grid_string=0p25 - 9: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 9: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 9: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 9: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 9: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 9: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 9: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 9: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 9: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 9: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 9: + interp_atmos_master.sh[31]IFS=: - 9: + interp_atmos_master.sh[31]read -ra grids - 9: + interp_atmos_master.sh[33]output_grids= - 9: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 9: + interp_atmos_master.sh[35]gridopt=grid0p25 - 9: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_10_0p25' - 9: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_10 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_10_0p25 - 9: 1:0:d=2021032312:RWMR:250 mb:25 hour fcst: - 9: 2:179:d=2021032312:SNMR:250 mb:25 hour fcst: - 9: 3:9005:d=2021032312:GRLE:250 mb:25 hour fcst: - 9: 4:9310:d=2021032312:O3MR:250 mb:25 hour fcst: - 9: 5:34919:d=2021032312:HGT:300 mb:25 hour fcst: - 9: 6:57212:d=2021032312:TMP:300 mb:25 hour fcst: - 9: 7:75646:d=2021032312:RH:300 mb:25 hour fcst: - 9: 8:94577:d=2021032312:TCDC:300 mb:25 hour fcst: - 9: 9:107330:d=2021032312:SPFH:300 mb:25 hour fcst: - 9: 10:134042:d=2021032312:VVEL:300 mb:25 hour fcst: - 9: 11:157947:d=2021032312:DZDT:300 mb:25 hour fcst: - 9: 12:186058:d=2021032312:UGRD:300 mb:25 hour fcst: - 9: 13:200722:d=2021032312:VGRD:300 mb:25 hour fcst: - 9: 14:216050:d=2021032312:ABSV:300 mb:25 hour fcst: - 9: 15:238377:d=2021032312:CLMR:300 mb:25 hour fcst: - 9: 16:238556:d=2021032312:ICMR:300 mb:25 hour fcst: - 9: 17:254348:d=2021032312:RWMR:300 mb:25 hour fcst: - 9: 18:254527:d=2021032312:SNMR:300 mb:25 hour fcst: - 9: 19:266056:d=2021032312:GRLE:300 mb:25 hour fcst: - 9: 20:266467:d=2021032312:O3MR:300 mb:25 hour fcst: - 9: 21:293612:d=2021032312:HGT:350 mb:25 hour fcst: - 9: 22:315556:d=2021032312:TMP:350 mb:25 hour fcst: - 9: 23:333807:d=2021032312:RH:350 mb:25 hour fcst: - 9: 24:352892:d=2021032312:TCDC:350 mb:25 hour fcst: - 9: 25:366070:d=2021032312:SPFH:350 mb:25 hour fcst: - 9: 26:394283:d=2021032312:VVEL:350 mb:25 hour fcst: - 9: 27:419059:d=2021032312:DZDT:350 mb:25 hour fcst: - 9: 28:447321:d=2021032312:UGRD:350 mb:25 hour fcst: - 9: 29:461928:d=2021032312:VGRD:350 mb:25 hour fcst: - 9: 30:477139:d=2021032312:ABSV:350 mb:25 hour fcst: - 9: 31:499379:d=2021032312:CLMR:350 mb:25 hour fcst: - 9: + interp_atmos_master.sh[47]export err=0 - 9: + interp_atmos_master.sh[47]err=0 - 9: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 9: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 9: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_10_0p25 - 9: + product_functions.sh[5]local filename=pgb2file_f025_10_0p25 - 9: + product_functions.sh[6]wgrib2 pgb2file_f025_10_0p25 -not_if :RH: -grib pgb2file_f025_10_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_10_0p25.new - 9: 1:0:d=2021032312:RWMR:250 mb:25 hour fcst: - 9: 2:179:d=2021032312:SNMR:250 mb:25 hour fcst: - 9: 3:174573:d=2021032312:GRLE:250 mb:25 hour fcst: - 9: 4:176480:d=2021032312:O3MR:250 mb:25 hour fcst: - 9: 5:842864:d=2021032312:HGT:300 mb:25 hour fcst: - 9: 6:1396784:d=2021032312:TMP:300 mb:25 hour fcst: - 9: 7:1808945:d=2021032312:RH:300 mb:25 hour fcst: - 9: 8:2234385:d=2021032312:TCDC:300 mb:25 hour fcst: - 9: 9:2471805:d=2021032312:SPFH:300 mb:25 hour fcst: - 9: 10:3178649:d=2021032312:VVEL:300 mb:25 hour fcst: - 9: 11:3811320:d=2021032312:DZDT:300 mb:25 hour fcst: - 9: 12:4602538:d=2021032312:UGRD:300 mb:25 hour fcst: - 9: 13:4919801:d=2021032312:VGRD:300 mb:25 hour fcst: - 9: 14:5253520:d=2021032312:ABSV:300 mb:25 hour fcst: - 9: 15:5805484:d=2021032312:CLMR:300 mb:25 hour fcst: - 9: 16:5805663:d=2021032312:ICMR:300 mb:25 hour fcst: - 9: 17:6155303:d=2021032312:RWMR:300 mb:25 hour fcst: - 9: 18:6155482:d=2021032312:SNMR:300 mb:25 hour fcst: - 9: 19:6390677:d=2021032312:GRLE:300 mb:25 hour fcst: - 9: 20:6395213:d=2021032312:O3MR:300 mb:25 hour fcst: - 9: 21:7114226:d=2021032312:HGT:350 mb:25 hour fcst: - 9: 22:7660430:d=2021032312:TMP:350 mb:25 hour fcst: - 9: 23:8067498:d=2021032312:RH:350 mb:25 hour fcst: - 9: 24:8496856:d=2021032312:TCDC:350 mb:25 hour fcst: - 9: 25:8737205:d=2021032312:SPFH:350 mb:25 hour fcst: - 9: 26:9482711:d=2021032312:VVEL:350 mb:25 hour fcst: - 9: 27:10136463:d=2021032312:DZDT:350 mb:25 hour fcst: - 9: 28:10930251:d=2021032312:UGRD:350 mb:25 hour fcst: - 9: 29:11242269:d=2021032312:VGRD:350 mb:25 hour fcst: - 9: 30:11573636:d=2021032312:ABSV:350 mb:25 hour fcst: - 9: 31:12121598:d=2021032312:CLMR:350 mb:25 hour fcst: - 9: + product_functions.sh[10]rc=0 - 9: + product_functions.sh[11](( rc == 0 )) - 9: + product_functions.sh[11]mv pgb2file_f025_10_0p25.new pgb2file_f025_10_0p25 - 9: + product_functions.sh[12]return 0 - 9: + interp_atmos_master.sh[56]export err=0 - 9: + interp_atmos_master.sh[56]err=0 - 9: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 9: ++ interp_atmos_master.sh[62]wc -l - 9: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_10_0p25 -match 'LAND|ICEC' - 9: + interp_atmos_master.sh[62]var_count=0 - 9: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 9: + interp_atmos_master.sh[73]exit 0 - 2: + bash[8]'[' -z '' ']' - 2: + bash[9]case "$-" in - 2: + bash[12]__lmod_vx=x - 2: + bash[16]'[' -n x ']' - 2: + bash[16]set +x - 2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 2: Shell debugging restarted - 2: + bash[224]unset __lmod_vx - 2: + interp_atmos_master.sh[7]input_file=tmpfile_f025_3 - 2: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_3 - 2: + interp_atmos_master.sh[9]grid_string=0p25 - 2: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 2: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 2: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 2: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 2: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 2: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 2: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 2: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 2: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 2: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 2: + interp_atmos_master.sh[31]IFS=: - 2: + interp_atmos_master.sh[31]read -ra grids - 2: + interp_atmos_master.sh[33]output_grids= - 2: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 2: + interp_atmos_master.sh[35]gridopt=grid0p25 - 2: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_3_0p25' - 2: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_3 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_3_0p25 - 2: 1:0:d=2021032312:O3MR:0.1 mb:25 hour fcst: - 2: 2:23658:d=2021032312:HGT:0.2 mb:25 hour fcst: - 2: 3:44652:d=2021032312:TMP:0.2 mb:25 hour fcst: - 2: 4:60733:d=2021032312:RH:0.2 mb:25 hour fcst: - 2: 5:71802:d=2021032312:SPFH:0.2 mb:25 hour fcst: - 2: 6:91462:d=2021032312:VVEL:0.2 mb:25 hour fcst: - 2: 7:120287:d=2021032312:DZDT:0.2 mb:25 hour fcst: - 2: 8:147785:d=2021032312:UGRD:0.2 mb:25 hour fcst: - 2: 9:167137:d=2021032312:VGRD:0.2 mb:25 hour fcst: - 2: 10:186035:d=2021032312:ABSV:0.2 mb:25 hour fcst: - 2: 11:203752:d=2021032312:O3MR:0.2 mb:25 hour fcst: - 2: 12:229349:d=2021032312:HGT:0.4 mb:25 hour fcst: - 2: 13:250405:d=2021032312:TMP:0.4 mb:25 hour fcst: - 2: 14:266183:d=2021032312:RH:0.4 mb:25 hour fcst: - 2: 15:275189:d=2021032312:SPFH:0.4 mb:25 hour fcst: - 2: 16:295186:d=2021032312:VVEL:0.4 mb:25 hour fcst: - 2: 17:320056:d=2021032312:DZDT:0.4 mb:25 hour fcst: - 2: 18:350519:d=2021032312:UGRD:0.4 mb:25 hour fcst: - 2: 19:369448:d=2021032312:VGRD:0.4 mb:25 hour fcst: - 2: 20:387633:d=2021032312:ABSV:0.4 mb:25 hour fcst: - 2: 21:404425:d=2021032312:O3MR:0.4 mb:25 hour fcst: - 2: 22:429075:d=2021032312:HGT:0.7 mb:25 hour fcst: - 2: 23:447214:d=2021032312:TMP:0.7 mb:25 hour fcst: - 2: 24:463025:d=2021032312:RH:0.7 mb:25 hour fcst: - 2: 25:471955:d=2021032312:SPFH:0.7 mb:25 hour fcst: - 2: 26:491206:d=2021032312:VVEL:0.7 mb:25 hour fcst: - 2: 27:517245:d=2021032312:DZDT:0.7 mb:25 hour fcst: - 2: 28:546540:d=2021032312:UGRD:0.7 mb:25 hour fcst: - 2: 29:565154:d=2021032312:VGRD:0.7 mb:25 hour fcst: - 2: 30:583010:d=2021032312:ABSV:0.7 mb:25 hour fcst: - 2: 31:599470:d=2021032312:O3MR:0.7 mb:25 hour fcst: - 2: + interp_atmos_master.sh[47]export err=0 - 2: + interp_atmos_master.sh[47]err=0 - 2: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 2: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 2: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_3_0p25 - 2: + product_functions.sh[5]local filename=pgb2file_f025_3_0p25 - 2: + product_functions.sh[6]wgrib2 pgb2file_f025_3_0p25 -not_if :RH: -grib pgb2file_f025_3_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_3_0p25.new - 2: 1:0:d=2021032312:O3MR:0.1 mb:25 hour fcst: - 2: 2:622596:d=2021032312:HGT:0.2 mb:25 hour fcst: - 2: 3:1143121:d=2021032312:TMP:0.2 mb:25 hour fcst: - 2: 4:1496887:d=2021032312:RH:0.2 mb:25 hour fcst: - 2: 5:1752129:d=2021032312:SPFH:0.2 mb:25 hour fcst: - 2: 6:2232021:d=2021032312:VVEL:0.2 mb:25 hour fcst: - 2: 7:3029183:d=2021032312:DZDT:0.2 mb:25 hour fcst: - 2: 8:3813047:d=2021032312:UGRD:0.2 mb:25 hour fcst: - 2: 9:4273868:d=2021032312:VGRD:0.2 mb:25 hour fcst: - 2: 10:4711563:d=2021032312:ABSV:0.2 mb:25 hour fcst: - 2: 11:5095291:d=2021032312:O3MR:0.2 mb:25 hour fcst: - 2: 12:5799368:d=2021032312:HGT:0.4 mb:25 hour fcst: - 2: 13:6313668:d=2021032312:TMP:0.4 mb:25 hour fcst: - 2: 14:6665603:d=2021032312:RH:0.4 mb:25 hour fcst: - 2: 15:6897925:d=2021032312:SPFH:0.4 mb:25 hour fcst: - 2: 16:7418296:d=2021032312:VVEL:0.4 mb:25 hour fcst: - 2: 17:8092450:d=2021032312:DZDT:0.4 mb:25 hour fcst: - 2: 18:9012747:d=2021032312:UGRD:0.4 mb:25 hour fcst: - 2: 19:9459074:d=2021032312:VGRD:0.4 mb:25 hour fcst: - 2: 20:9879971:d=2021032312:ABSV:0.4 mb:25 hour fcst: - 2: 21:10239167:d=2021032312:O3MR:0.4 mb:25 hour fcst: - 2: 22:10905232:d=2021032312:HGT:0.7 mb:25 hour fcst: - 2: 23:11319331:d=2021032312:TMP:0.7 mb:25 hour fcst: - 2: 24:11665876:d=2021032312:RH:0.7 mb:25 hour fcst: - 2: 25:11910459:d=2021032312:SPFH:0.7 mb:25 hour fcst: - 2: 26:12432255:d=2021032312:VVEL:0.7 mb:25 hour fcst: - 2: 27:13140762:d=2021032312:DZDT:0.7 mb:25 hour fcst: - 2: 28:14002311:d=2021032312:UGRD:0.7 mb:25 hour fcst: - 2: 29:14438410:d=2021032312:VGRD:0.7 mb:25 hour fcst: - 2: 30:14850148:d=2021032312:ABSV:0.7 mb:25 hour fcst: - 2: 31:15199397:d=2021032312:O3MR:0.7 mb:25 hour fcst: - 2: + product_functions.sh[10]rc=0 - 2: + product_functions.sh[11](( rc == 0 )) - 2: + product_functions.sh[11]mv pgb2file_f025_3_0p25.new pgb2file_f025_3_0p25 - 2: + product_functions.sh[12]return 0 - 2: + interp_atmos_master.sh[56]export err=0 - 2: + interp_atmos_master.sh[56]err=0 - 2: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 2: ++ interp_atmos_master.sh[62]wc -l - 2: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_3_0p25 -match 'LAND|ICEC' - 2: + interp_atmos_master.sh[62]var_count=0 - 2: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 2: + interp_atmos_master.sh[73]exit 0 - 3: + bash[8]'[' -z '' ']' - 3: + bash[9]case "$-" in - 3: + bash[12]__lmod_vx=x - 3: + bash[16]'[' -n x ']' - 3: + bash[16]set +x - 3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 3: Shell debugging restarted - 3: + bash[224]unset __lmod_vx - 3: + interp_atmos_master.sh[7]input_file=tmpfile_f025_4 - 3: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_4 - 3: + interp_atmos_master.sh[9]grid_string=0p25 - 3: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 3: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 3: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 3: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 3: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 3: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 3: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 3: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 3: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 3: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 3: + interp_atmos_master.sh[31]IFS=: - 3: + interp_atmos_master.sh[31]read -ra grids - 3: + interp_atmos_master.sh[33]output_grids= - 3: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 3: + interp_atmos_master.sh[35]gridopt=grid0p25 - 3: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_4_0p25' - 3: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_4 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_4_0p25 - 3: 1:0:d=2021032312:HGT:1 mb:25 hour fcst: - 3: 2:17997:d=2021032312:TMP:1 mb:25 hour fcst: - 3: 3:33857:d=2021032312:RH:1 mb:25 hour fcst: - 3: 4:43627:d=2021032312:SPFH:1 mb:25 hour fcst: - 3: 5:62108:d=2021032312:VVEL:1 mb:25 hour fcst: - 3: 6:88821:d=2021032312:DZDT:1 mb:25 hour fcst: - 3: 7:115011:d=2021032312:UGRD:1 mb:25 hour fcst: - 3: 8:133609:d=2021032312:VGRD:1 mb:25 hour fcst: - 3: 9:151281:d=2021032312:ABSV:1 mb:25 hour fcst: - 3: 10:167579:d=2021032312:O3MR:1 mb:25 hour fcst: - 3: 11:188711:d=2021032312:HGT:2 mb:25 hour fcst: - 3: 12:208914:d=2021032312:TMP:2 mb:25 hour fcst: - 3: 13:225566:d=2021032312:RH:2 mb:25 hour fcst: - 3: 14:234248:d=2021032312:SPFH:2 mb:25 hour fcst: - 3: 15:253116:d=2021032312:VVEL:2 mb:25 hour fcst: - 3: 16:281762:d=2021032312:DZDT:2 mb:25 hour fcst: - 3: 17:308266:d=2021032312:UGRD:2 mb:25 hour fcst: - 3: 18:319829:d=2021032312:VGRD:2 mb:25 hour fcst: - 3: 19:338257:d=2021032312:ABSV:2 mb:25 hour fcst: - 3: 20:355276:d=2021032312:O3MR:2 mb:25 hour fcst: - 3: 21:378051:d=2021032312:HGT:3 mb:25 hour fcst: - 3: 22:395851:d=2021032312:TMP:3 mb:25 hour fcst: - 3: 23:412339:d=2021032312:RH:3 mb:25 hour fcst: - 3: 24:418329:d=2021032312:SPFH:3 mb:25 hour fcst: - 3: 25:439457:d=2021032312:VVEL:3 mb:25 hour fcst: - 3: 26:469498:d=2021032312:DZDT:3 mb:25 hour fcst: - 3: 27:495703:d=2021032312:UGRD:3 mb:25 hour fcst: - 3: 28:507069:d=2021032312:VGRD:3 mb:25 hour fcst: - 3: 29:525061:d=2021032312:ABSV:3 mb:25 hour fcst: - 3: 30:541979:d=2021032312:O3MR:3 mb:25 hour fcst: - 3: 31:565281:d=2021032312:HGT:5 mb:25 hour fcst: - 3: + interp_atmos_master.sh[47]export err=0 - 3: + interp_atmos_master.sh[47]err=0 - 3: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 3: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 3: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_4_0p25 - 3: + product_functions.sh[5]local filename=pgb2file_f025_4_0p25 - 3: + product_functions.sh[6]wgrib2 pgb2file_f025_4_0p25 -not_if :RH: -grib pgb2file_f025_4_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_4_0p25.new - 3: 1:0:d=2021032312:HGT:1 mb:25 hour fcst: - 3: 2:407129:d=2021032312:TMP:1 mb:25 hour fcst: - 3: 3:754609:d=2021032312:RH:1 mb:25 hour fcst: - 3: 4:1011544:d=2021032312:SPFH:1 mb:25 hour fcst: - 3: 5:1504452:d=2021032312:VVEL:1 mb:25 hour fcst: - 3: 6:2252523:d=2021032312:DZDT:1 mb:25 hour fcst: - 3: 7:2977126:d=2021032312:UGRD:1 mb:25 hour fcst: - 3: 8:3409131:d=2021032312:VGRD:1 mb:25 hour fcst: - 3: 9:3815772:d=2021032312:ABSV:1 mb:25 hour fcst: - 3: 10:4158432:d=2021032312:O3MR:1 mb:25 hour fcst: - 3: 11:4693640:d=2021032312:HGT:2 mb:25 hour fcst: - 3: 12:5180276:d=2021032312:TMP:2 mb:25 hour fcst: - 3: 13:5550485:d=2021032312:RH:2 mb:25 hour fcst: - 3: 14:5798432:d=2021032312:SPFH:2 mb:25 hour fcst: - 3: 15:6279862:d=2021032312:VVEL:2 mb:25 hour fcst: - 3: 16:7076049:d=2021032312:DZDT:2 mb:25 hour fcst: - 3: 17:7824260:d=2021032312:UGRD:2 mb:25 hour fcst: - 3: 18:8281593:d=2021032312:VGRD:2 mb:25 hour fcst: - 3: 19:8711257:d=2021032312:ABSV:2 mb:25 hour fcst: - 3: 20:9083520:d=2021032312:O3MR:2 mb:25 hour fcst: - 3: 21:9668653:d=2021032312:HGT:3 mb:25 hour fcst: - 3: 22:10068826:d=2021032312:TMP:3 mb:25 hour fcst: - 3: 23:10431662:d=2021032312:RH:3 mb:25 hour fcst: - 3: 24:10600408:d=2021032312:SPFH:3 mb:25 hour fcst: - 3: 25:11144899:d=2021032312:VVEL:3 mb:25 hour fcst: - 3: 26:11987448:d=2021032312:DZDT:3 mb:25 hour fcst: - 3: 27:12724083:d=2021032312:UGRD:3 mb:25 hour fcst: - 3: 28:13175032:d=2021032312:VGRD:3 mb:25 hour fcst: - 3: 29:13585829:d=2021032312:ABSV:3 mb:25 hour fcst: - 3: 30:13950702:d=2021032312:O3MR:3 mb:25 hour fcst: - 3: 31:14552394:d=2021032312:HGT:5 mb:25 hour fcst: - 3: + product_functions.sh[10]rc=0 - 3: + product_functions.sh[11](( rc == 0 )) - 3: + product_functions.sh[11]mv pgb2file_f025_4_0p25.new pgb2file_f025_4_0p25 - 3: + product_functions.sh[12]return 0 - 3: + interp_atmos_master.sh[56]export err=0 - 3: + interp_atmos_master.sh[56]err=0 - 3: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 3: ++ interp_atmos_master.sh[62]wc -l - 3: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_4_0p25 -match 'LAND|ICEC' - 3: + interp_atmos_master.sh[62]var_count=0 - 3: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 3: + interp_atmos_master.sh[73]exit 0 -23: + bash[8]'[' -z '' ']' -23: + bash[9]case "$-" in -23: + bash[12]__lmod_vx=x -23: + bash[16]'[' -n x ']' -23: + bash[16]set +x -23: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -23: Shell debugging restarted -23: + bash[224]unset __lmod_vx -23: + interp_atmos_master.sh[7]input_file=tmpfile_f025_24 -23: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_24 -23: + interp_atmos_master.sh[9]grid_string=0p25 -23: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -23: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -23: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -23: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -23: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -23: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -23: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -23: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -23: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -23: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -23: + interp_atmos_master.sh[31]IFS=: -23: + interp_atmos_master.sh[31]read -ra grids -23: + interp_atmos_master.sh[33]output_grids= -23: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -23: + interp_atmos_master.sh[35]gridopt=grid0p25 -23: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_24_0p25' -23: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_24 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_24_0p25 -23: 1:0:d=2021032312:CAPE:180-0 mb above ground:25 hour fcst: -23: 2:13978:d=2021032312:CIN:180-0 mb above ground:25 hour fcst: -23: 3:28018:d=2021032312:HPBL:surface:25 hour fcst: -23: 4:59049:d=2021032312:RH:0.33-1 sigma layer:25 hour fcst: -23: 5:77421:d=2021032312:RH:0.44-1 sigma layer:25 hour fcst: -23: 6:95418:d=2021032312:RH:0.72-0.94 sigma layer:25 hour fcst: -23: 7:113859:d=2021032312:RH:0.44-0.72 sigma layer:25 hour fcst: -23: 8:133194:d=2021032312:TMP:0.995 sigma level:25 hour fcst: -23: 9:154030:d=2021032312:POT:0.995 sigma level:25 hour fcst: -23: 10:174858:d=2021032312:RH:0.995 sigma level:25 hour fcst: -23: 11:192647:d=2021032312:UGRD:0.995 sigma level:25 hour fcst: -23: 12:213839:d=2021032312:VGRD:0.995 sigma level:25 hour fcst: -23: 13:235365:d=2021032312:VVEL:0.995 sigma level:25 hour fcst: -23: 14:256083:d=2021032312:CAPE:90-0 mb above ground:25 hour fcst: -23: 15:268867:d=2021032312:CIN:90-0 mb above ground:25 hour fcst: -23: 16:282955:d=2021032312:CAPE:255-0 mb above ground:25 hour fcst: -23: 17:296302:d=2021032312:CIN:255-0 mb above ground:25 hour fcst: -23: 18:307697:d=2021032312:PLPL:255-0 mb above ground:25 hour fcst: -23: 19:331649:d=2021032312:LAND:surface:25 hour fcst: -23: 20:333590:d=2021032312:ICEC:surface:25 hour fcst: -23: 21:336759:d=2021032312:ALBDO:surface:24-25 hour ave fcst: -23: 22:351243:d=2021032312:ICETMP:surface:25 hour fcst: -23: 23:357922:d=2021032312:UGRD:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 24:368649:d=2021032312:VGRD:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 25:379365:d=2021032312:TMP:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 26:393710:d=2021032312:HGT:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 27:411069:d=2021032312:PRES:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 28:429438:d=2021032312:VWSH:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 29:441357:d=2021032312:UGRD:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 30:452235:d=2021032312:VGRD:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 31:467016:d=2021032312:TMP:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 32:481428:d=2021032312:HGT:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 33:498989:d=2021032312:PRES:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 34:517293:d=2021032312:VWSH:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: + interp_atmos_master.sh[47]export err=0 -23: + interp_atmos_master.sh[47]err=0 -23: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -23: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -23: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_24_0p25 -23: + product_functions.sh[5]local filename=pgb2file_f025_24_0p25 -23: + product_functions.sh[6]wgrib2 pgb2file_f025_24_0p25 -not_if :RH: -grib pgb2file_f025_24_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_24_0p25.new -23: 1:0:d=2021032312:CAPE:180-0 mb above ground:25 hour fcst: -23: 2:303999:d=2021032312:CIN:180-0 mb above ground:25 hour fcst: -23: 3:599203:d=2021032312:HPBL:surface:25 hour fcst: -23: 4:1464723:d=2021032312:RH:0.33-1 sigma layer:25 hour fcst: -23: 5:1875796:d=2021032312:RH:0.44-1 sigma layer:25 hour fcst: -23: 6:2276804:d=2021032312:RH:0.72-0.94 sigma layer:25 hour fcst: -23: 7:2694231:d=2021032312:RH:0.44-0.72 sigma layer:25 hour fcst: -23: 8:3132874:d=2021032312:TMP:0.995 sigma level:25 hour fcst: -23: 9:3618335:d=2021032312:POT:0.995 sigma level:25 hour fcst: -23: 10:4107654:d=2021032312:RH:0.995 sigma level:25 hour fcst: -23: 11:4493690:d=2021032312:UGRD:0.995 sigma level:25 hour fcst: -23: 12:5006724:d=2021032312:VGRD:0.995 sigma level:25 hour fcst: -23: 13:5535650:d=2021032312:VVEL:0.995 sigma level:25 hour fcst: -23: 14:6031543:d=2021032312:CAPE:90-0 mb above ground:25 hour fcst: -23: 15:6307053:d=2021032312:CIN:90-0 mb above ground:25 hour fcst: -23: 16:6599565:d=2021032312:CAPE:255-0 mb above ground:25 hour fcst: -23: 17:6899566:d=2021032312:CIN:255-0 mb above ground:25 hour fcst: -23: 18:7118760:d=2021032312:PLPL:255-0 mb above ground:25 hour fcst: -23: 19:7677986:d=2021032312:LAND:surface:25 hour fcst: -23: 20:7697906:d=2021032312:ICEC:surface:25 hour fcst: -23: 21:7752991:d=2021032312:ALBDO:surface:24-25 hour ave fcst: -23: 22:8081346:d=2021032312:ICETMP:surface:25 hour fcst: -23: 23:8303437:d=2021032312:UGRD:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 24:8617741:d=2021032312:VGRD:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 25:8929707:d=2021032312:TMP:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 26:9354999:d=2021032312:HGT:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 27:9906128:d=2021032312:PRES:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 28:10489263:d=2021032312:VWSH:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 29:10839662:d=2021032312:UGRD:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 30:11299259:d=2021032312:VGRD:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 31:11753115:d=2021032312:TMP:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 32:12185934:d=2021032312:HGT:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 33:12739640:d=2021032312:PRES:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 34:13316546:d=2021032312:VWSH:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: + product_functions.sh[10]rc=0 -23: + product_functions.sh[11](( rc == 0 )) -23: + product_functions.sh[11]mv pgb2file_f025_24_0p25.new pgb2file_f025_24_0p25 -23: + product_functions.sh[12]return 0 -23: + interp_atmos_master.sh[56]export err=0 -23: + interp_atmos_master.sh[56]err=0 -23: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -23: ++ interp_atmos_master.sh[62]wc -l -23: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_24_0p25 -match 'LAND|ICEC' -23: + interp_atmos_master.sh[62]var_count=2 -23: + interp_atmos_master.sh[63][[ 2 -eq 2 ]] -23: + interp_atmos_master.sh[64]mod_icec pgb2file_f025_24_0p25 -23: + product_functions.sh[17]local filename=pgb2file_f025_24_0p25 -23: + product_functions.sh[18]wgrib2 pgb2file_f025_24_0p25 -if LAND -rpn sto_1 -fi -if ICEC -rpn 'rcl_1:0:==:*' -fi -set_grib_type same -set_scaling same same -grib_out pgb2file_f025_24_0p25.new -23: 1:0:d=2021032312:CAPE:180-0 mb above ground:25 hour fcst: -23: 2:303999:d=2021032312:CIN:180-0 mb above ground:25 hour fcst: -23: 3:599203:d=2021032312:HPBL:surface:25 hour fcst: -23: 4:1464723:d=2021032312:RH:0.33-1 sigma layer:25 hour fcst: -23: 5:1875796:d=2021032312:RH:0.44-1 sigma layer:25 hour fcst: -23: 6:2276804:d=2021032312:RH:0.72-0.94 sigma layer:25 hour fcst: -23: 7:2694231:d=2021032312:RH:0.44-0.72 sigma layer:25 hour fcst: -23: 8:3132874:d=2021032312:TMP:0.995 sigma level:25 hour fcst: -23: 9:3618335:d=2021032312:POT:0.995 sigma level:25 hour fcst: -23: 10:4107654:d=2021032312:RH:0.995 sigma level:25 hour fcst: -23: 11:4493690:d=2021032312:UGRD:0.995 sigma level:25 hour fcst: -23: 12:5006724:d=2021032312:VGRD:0.995 sigma level:25 hour fcst: -23: 13:5535650:d=2021032312:VVEL:0.995 sigma level:25 hour fcst: -23: 14:6031543:d=2021032312:CAPE:90-0 mb above ground:25 hour fcst: -23: 15:6307053:d=2021032312:CIN:90-0 mb above ground:25 hour fcst: -23: 16:6599565:d=2021032312:CAPE:255-0 mb above ground:25 hour fcst: -23: 17:6899566:d=2021032312:CIN:255-0 mb above ground:25 hour fcst: -23: 18:7118760:d=2021032312:PLPL:255-0 mb above ground:25 hour fcst: -23: 19:7677986:d=2021032312:LAND:surface:25 hour fcst: -23: 20:7697906:d=2021032312:ICEC:surface:25 hour fcst: -23: 21:7752991:d=2021032312:ALBDO:surface:24-25 hour ave fcst: -23: 22:8081346:d=2021032312:ICETMP:surface:25 hour fcst: -23: 23:8303437:d=2021032312:UGRD:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 24:8617741:d=2021032312:VGRD:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 25:8929707:d=2021032312:TMP:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 26:9354999:d=2021032312:HGT:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 27:9906128:d=2021032312:PRES:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 28:10489263:d=2021032312:VWSH:PV=2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 29:10839662:d=2021032312:UGRD:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 30:11299259:d=2021032312:VGRD:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 31:11753115:d=2021032312:TMP:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 32:12185934:d=2021032312:HGT:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 33:12739640:d=2021032312:PRES:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 34:13316546:d=2021032312:VWSH:PV=-2e-06 (Km^2/kg/s) surface:25 hour fcst: -23: + product_functions.sh[24]rc=0 -23: + product_functions.sh[25](( rc == 0 )) -23: + product_functions.sh[25]mv pgb2file_f025_24_0p25.new pgb2file_f025_24_0p25 -23: + product_functions.sh[26]return 0 -23: + interp_atmos_master.sh[65]export err=0 -23: + interp_atmos_master.sh[65]err=0 -23: + interp_atmos_master.sh[66][[ 0 -ne 0 ]] -23: + interp_atmos_master.sh[73]exit 0 - 4: + bash[8]'[' -z '' ']' - 4: + bash[9]case "$-" in - 4: + bash[12]__lmod_vx=x - 4: + bash[16]'[' -n x ']' - 4: + bash[16]set +x - 4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 4: Shell debugging restarted - 4: + bash[224]unset __lmod_vx - 4: + interp_atmos_master.sh[7]input_file=tmpfile_f025_5 - 4: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_5 - 4: + interp_atmos_master.sh[9]grid_string=0p25 - 4: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 4: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 4: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 4: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 4: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 4: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 4: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 4: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 4: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 4: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 4: + interp_atmos_master.sh[31]IFS=: - 4: + interp_atmos_master.sh[31]read -ra grids - 4: + interp_atmos_master.sh[33]output_grids= - 4: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 4: + interp_atmos_master.sh[35]gridopt=grid0p25 - 4: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_5_0p25' - 4: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_5 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_5_0p25 - 4: 1:0:d=2021032312:TMP:5 mb:25 hour fcst: - 4: 2:16691:d=2021032312:RH:5 mb:25 hour fcst: - 4: 3:25881:d=2021032312:SPFH:5 mb:25 hour fcst: - 4: 4:46594:d=2021032312:VVEL:5 mb:25 hour fcst: - 4: 5:75989:d=2021032312:DZDT:5 mb:25 hour fcst: - 4: 6:102039:d=2021032312:UGRD:5 mb:25 hour fcst: - 4: 7:113636:d=2021032312:VGRD:5 mb:25 hour fcst: - 4: 8:132244:d=2021032312:ABSV:5 mb:25 hour fcst: - 4: 9:149929:d=2021032312:O3MR:5 mb:25 hour fcst: - 4: 10:172700:d=2021032312:HGT:7 mb:25 hour fcst: - 4: 11:190365:d=2021032312:TMP:7 mb:25 hour fcst: - 4: 12:206449:d=2021032312:RH:7 mb:25 hour fcst: - 4: 13:212499:d=2021032312:SPFH:7 mb:25 hour fcst: - 4: 14:231754:d=2021032312:VVEL:7 mb:25 hour fcst: - 4: 15:256376:d=2021032312:DZDT:7 mb:25 hour fcst: - 4: 16:282012:d=2021032312:UGRD:7 mb:25 hour fcst: - 4: 17:293429:d=2021032312:VGRD:7 mb:25 hour fcst: - 4: 18:312182:d=2021032312:ABSV:7 mb:25 hour fcst: - 4: 19:329917:d=2021032312:O3MR:7 mb:25 hour fcst: - 4: 20:353584:d=2021032312:HGT:10 mb:25 hour fcst: - 4: 21:373421:d=2021032312:TMP:10 mb:25 hour fcst: - 4: 22:389619:d=2021032312:RH:10 mb:25 hour fcst: - 4: 23:397290:d=2021032312:SPFH:10 mb:25 hour fcst: - 4: 24:415611:d=2021032312:VVEL:10 mb:25 hour fcst: - 4: 25:441149:d=2021032312:DZDT:10 mb:25 hour fcst: - 4: 26:466420:d=2021032312:UGRD:10 mb:25 hour fcst: - 4: 27:477748:d=2021032312:VGRD:10 mb:25 hour fcst: - 4: 28:496317:d=2021032312:ABSV:10 mb:25 hour fcst: - 4: 29:513986:d=2021032312:O3MR:10 mb:25 hour fcst: - 4: 30:532720:d=2021032312:HGT:15 mb:25 hour fcst: - 4: 31:552250:d=2021032312:TMP:15 mb:25 hour fcst: - 4: + interp_atmos_master.sh[47]export err=0 - 4: + interp_atmos_master.sh[47]err=0 - 4: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 4: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 4: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_5_0p25 - 4: + product_functions.sh[5]local filename=pgb2file_f025_5_0p25 - 4: + product_functions.sh[6]wgrib2 pgb2file_f025_5_0p25 -not_if :RH: -grib pgb2file_f025_5_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_5_0p25.new - 4: 1:0:d=2021032312:TMP:5 mb:25 hour fcst: - 4: 2:366800:d=2021032312:RH:5 mb:25 hour fcst: - 4: 3:618365:d=2021032312:SPFH:5 mb:25 hour fcst: - 4: 4:1141086:d=2021032312:VVEL:5 mb:25 hour fcst: - 4: 5:1965838:d=2021032312:DZDT:5 mb:25 hour fcst: - 4: 6:2689159:d=2021032312:UGRD:5 mb:25 hour fcst: - 4: 7:3140349:d=2021032312:VGRD:5 mb:25 hour fcst: - 4: 8:3567617:d=2021032312:ABSV:5 mb:25 hour fcst: - 4: 9:3953780:d=2021032312:O3MR:5 mb:25 hour fcst: - 4: 10:4519938:d=2021032312:HGT:7 mb:25 hour fcst: - 4: 11:4916459:d=2021032312:TMP:7 mb:25 hour fcst: - 4: 12:5266449:d=2021032312:RH:7 mb:25 hour fcst: - 4: 13:5441344:d=2021032312:SPFH:7 mb:25 hour fcst: - 4: 14:5922456:d=2021032312:VVEL:7 mb:25 hour fcst: - 4: 15:6586965:d=2021032312:DZDT:7 mb:25 hour fcst: - 4: 16:7294489:d=2021032312:UGRD:7 mb:25 hour fcst: - 4: 17:7739212:d=2021032312:VGRD:7 mb:25 hour fcst: - 4: 18:8171492:d=2021032312:ABSV:7 mb:25 hour fcst: - 4: 19:8565691:d=2021032312:O3MR:7 mb:25 hour fcst: - 4: 20:9169629:d=2021032312:HGT:10 mb:25 hour fcst: - 4: 21:9639946:d=2021032312:TMP:10 mb:25 hour fcst: - 4: 22:9992212:d=2021032312:RH:10 mb:25 hour fcst: - 4: 23:10217199:d=2021032312:SPFH:10 mb:25 hour fcst: - 4: 24:10662365:d=2021032312:VVEL:10 mb:25 hour fcst: - 4: 25:11357520:d=2021032312:DZDT:10 mb:25 hour fcst: - 4: 26:12045060:d=2021032312:UGRD:10 mb:25 hour fcst: - 4: 27:12487294:d=2021032312:VGRD:10 mb:25 hour fcst: - 4: 28:12914869:d=2021032312:ABSV:10 mb:25 hour fcst: - 4: 29:13307110:d=2021032312:O3MR:10 mb:25 hour fcst: - 4: 30:13734671:d=2021032312:HGT:15 mb:25 hour fcst: - 4: 31:14196363:d=2021032312:TMP:15 mb:25 hour fcst: - 4: + product_functions.sh[10]rc=0 - 4: + product_functions.sh[11](( rc == 0 )) - 4: + product_functions.sh[11]mv pgb2file_f025_5_0p25.new pgb2file_f025_5_0p25 - 4: + product_functions.sh[12]return 0 - 4: + interp_atmos_master.sh[56]export err=0 - 4: + interp_atmos_master.sh[56]err=0 - 4: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 4: ++ interp_atmos_master.sh[62]wc -l - 4: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_5_0p25 -match 'LAND|ICEC' - 4: + interp_atmos_master.sh[62]var_count=0 - 4: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 4: + interp_atmos_master.sh[73]exit 0 -22: + bash[8]'[' -z '' ']' -22: + bash[9]case "$-" in -22: + bash[12]__lmod_vx=x -22: + bash[16]'[' -n x ']' -22: + bash[16]set +x -22: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -22: Shell debugging restarted -22: + bash[224]unset __lmod_vx -22: + interp_atmos_master.sh[7]input_file=tmpfile_f025_23 -22: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_23 -22: + interp_atmos_master.sh[9]grid_string=0p25 -22: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -22: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -22: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -22: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -22: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -22: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -22: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -22: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -22: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -22: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -22: + interp_atmos_master.sh[31]IFS=: -22: + interp_atmos_master.sh[31]read -ra grids -22: + interp_atmos_master.sh[33]output_grids= -22: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -22: + interp_atmos_master.sh[35]gridopt=grid0p25 -22: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_23_0p25' -22: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_23 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_23_0p25 -22: 1:0:d=2021032312:UGRD:40 m above ground:25 hour fcst: -22: 2:20997:d=2021032312:VGRD:40 m above ground:25 hour fcst: -22: 3:42392:d=2021032312:UGRD:50 m above ground:25 hour fcst: -22: 4:63511:d=2021032312:VGRD:50 m above ground:25 hour fcst: -22: 5:85019:d=2021032312:TMP:80 m above ground:25 hour fcst: -22: 6:105795:d=2021032312:SPFH:80 m above ground:25 hour fcst: -22: 7:134492:d=2021032312:PRES:80 m above ground:25 hour fcst: -22: 8:160074:d=2021032312:UGRD:80 m above ground:25 hour fcst: -22: 9:181345:d=2021032312:VGRD:80 m above ground:25 hour fcst: -22: 10:203049:d=2021032312:TMP:100 m above ground:25 hour fcst: -22: 11:223801:d=2021032312:UGRD:100 m above ground:25 hour fcst: -22: 12:245143:d=2021032312:VGRD:100 m above ground:25 hour fcst: -22: 13:266951:d=2021032312:TMP:1829 m above mean sea level:25 hour fcst: -22: 14:288597:d=2021032312:UGRD:1829 m above mean sea level:25 hour fcst: -22: 15:310843:d=2021032312:VGRD:1829 m above mean sea level:25 hour fcst: -22: 16:333743:d=2021032312:TMP:2743 m above mean sea level:25 hour fcst: -22: 17:355390:d=2021032312:UGRD:2743 m above mean sea level:25 hour fcst: -22: 18:378051:d=2021032312:VGRD:2743 m above mean sea level:25 hour fcst: -22: 19:401365:d=2021032312:TMP:3658 m above mean sea level:25 hour fcst: -22: 20:422655:d=2021032312:UGRD:3658 m above mean sea level:25 hour fcst: -22: 21:445700:d=2021032312:VGRD:3658 m above mean sea level:25 hour fcst: -22: 22:469486:d=2021032312:HGT:0C isotherm:25 hour fcst: -22: 23:491130:d=2021032312:RH:0C isotherm:25 hour fcst: -22: 24:510368:d=2021032312:HGT:highest tropospheric freezing level:25 hour fcst: -22: 25:531856:d=2021032312:RH:highest tropospheric freezing level:25 hour fcst: -22: 26:550866:d=2021032312:TMP:30-0 mb above ground:25 hour fcst: -22: 27:571684:d=2021032312:RH:30-0 mb above ground:25 hour fcst: -22: 28:589879:d=2021032312:SPFH:30-0 mb above ground:25 hour fcst: -22: 29:618704:d=2021032312:UGRD:30-0 mb above ground:25 hour fcst: -22: 30:639866:d=2021032312:VGRD:30-0 mb above ground:25 hour fcst: -22: 31:661649:d=2021032312:4LFTX:surface:25 hour fcst: -22: + interp_atmos_master.sh[47]export err=0 -22: + interp_atmos_master.sh[47]err=0 -22: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -22: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -22: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_23_0p25 -22: + product_functions.sh[5]local filename=pgb2file_f025_23_0p25 -22: + product_functions.sh[6]wgrib2 pgb2file_f025_23_0p25 -not_if :RH: -grib pgb2file_f025_23_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_23_0p25.new -22: 1:0:d=2021032312:UGRD:40 m above ground:25 hour fcst: -22: 2:507327:d=2021032312:VGRD:40 m above ground:25 hour fcst: -22: 3:1031983:d=2021032312:UGRD:50 m above ground:25 hour fcst: -22: 4:1541865:d=2021032312:VGRD:50 m above ground:25 hour fcst: -22: 5:2069888:d=2021032312:TMP:80 m above ground:25 hour fcst: -22: 6:2553313:d=2021032312:SPFH:80 m above ground:25 hour fcst: -22: 7:3335904:d=2021032312:PRES:80 m above ground:25 hour fcst: -22: 8:3978202:d=2021032312:UGRD:80 m above ground:25 hour fcst: -22: 9:4492923:d=2021032312:VGRD:80 m above ground:25 hour fcst: -22: 10:5026692:d=2021032312:TMP:100 m above ground:25 hour fcst: -22: 11:5509562:d=2021032312:UGRD:100 m above ground:25 hour fcst: -22: 12:6026477:d=2021032312:VGRD:100 m above ground:25 hour fcst: -22: 13:6562534:d=2021032312:TMP:1829 m above mean sea level:25 hour fcst: -22: 14:7162789:d=2021032312:UGRD:1829 m above mean sea level:25 hour fcst: -22: 15:7789898:d=2021032312:VGRD:1829 m above mean sea level:25 hour fcst: -22: 16:8448384:d=2021032312:TMP:2743 m above mean sea level:25 hour fcst: -22: 17:9033331:d=2021032312:UGRD:2743 m above mean sea level:25 hour fcst: -22: 18:9662310:d=2021032312:VGRD:2743 m above mean sea level:25 hour fcst: -22: 19:10327593:d=2021032312:TMP:3658 m above mean sea level:25 hour fcst: -22: 20:10903492:d=2021032312:UGRD:3658 m above mean sea level:25 hour fcst: -22: 21:11547626:d=2021032312:VGRD:3658 m above mean sea level:25 hour fcst: -22: 22:12221406:d=2021032312:HGT:0C isotherm:25 hour fcst: -22: 23:12812489:d=2021032312:RH:0C isotherm:25 hour fcst: -22: 24:13246008:d=2021032312:HGT:highest tropospheric freezing level:25 hour fcst: -22: 25:13825842:d=2021032312:RH:highest tropospheric freezing level:25 hour fcst: -22: 26:14264023:d=2021032312:TMP:30-0 mb above ground:25 hour fcst: -22: 27:14747925:d=2021032312:RH:30-0 mb above ground:25 hour fcst: -22: 28:15142824:d=2021032312:SPFH:30-0 mb above ground:25 hour fcst: -22: 29:15927961:d=2021032312:UGRD:30-0 mb above ground:25 hour fcst: -22: 30:16443275:d=2021032312:VGRD:30-0 mb above ground:25 hour fcst: -22: 31:16978987:d=2021032312:4LFTX:surface:25 hour fcst: -22: + product_functions.sh[10]rc=0 -22: + product_functions.sh[11](( rc == 0 )) -22: + product_functions.sh[11]mv pgb2file_f025_23_0p25.new pgb2file_f025_23_0p25 -22: + product_functions.sh[12]return 0 -22: + interp_atmos_master.sh[56]export err=0 -22: + interp_atmos_master.sh[56]err=0 -22: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -22: ++ interp_atmos_master.sh[62]wc -l -22: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_23_0p25 -match 'LAND|ICEC' -22: + interp_atmos_master.sh[62]var_count=0 -22: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -22: + interp_atmos_master.sh[73]exit 0 - 1: + bash[8]'[' -z '' ']' - 1: + bash[9]case "$-" in - 1: + bash[12]__lmod_vx=x - 1: + bash[16]'[' -n x ']' - 1: + bash[16]set +x - 1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 1: Shell debugging restarted - 1: + bash[224]unset __lmod_vx - 1: + interp_atmos_master.sh[7]input_file=tmpfile_f025_2 - 1: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_2 - 1: + interp_atmos_master.sh[9]grid_string=0p25 - 1: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 1: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 1: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 1: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 1: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 1: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 1: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 1: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 1: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 1: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 1: + interp_atmos_master.sh[31]IFS=: - 1: + interp_atmos_master.sh[31]read -ra grids - 1: + interp_atmos_master.sh[33]output_grids= - 1: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 1: + interp_atmos_master.sh[35]gridopt=grid0p25 - 1: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_2_0p25' - 1: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_2 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_2_0p25 - 1: 1:0:d=2021032312:ABSV:0.02 mb:25 hour fcst: - 1: 2:16817:d=2021032312:O3MR:0.02 mb:25 hour fcst: - 1: 3:32472:d=2021032312:HGT:0.04 mb:25 hour fcst: - 1: 4:56698:d=2021032312:TMP:0.04 mb:25 hour fcst: - 1: 5:73952:d=2021032312:RH:0.04 mb:25 hour fcst: - 1: 6:79930:d=2021032312:SPFH:0.04 mb:25 hour fcst: - 1: 7:105566:d=2021032312:VVEL:0.04 mb:25 hour fcst: - 1: 8:133088:d=2021032312:DZDT:0.04 mb:25 hour fcst: - 1: 9:157653:d=2021032312:UGRD:0.04 mb:25 hour fcst: - 1: 10:177584:d=2021032312:VGRD:0.04 mb:25 hour fcst: - 1: 11:197362:d=2021032312:ABSV:0.04 mb:25 hour fcst: - 1: 12:215783:d=2021032312:O3MR:0.04 mb:25 hour fcst: - 1: 13:239359:d=2021032312:HGT:0.07 mb:25 hour fcst: - 1: 14:263400:d=2021032312:TMP:0.07 mb:25 hour fcst: - 1: 15:280374:d=2021032312:RH:0.07 mb:25 hour fcst: - 1: 16:289947:d=2021032312:SPFH:0.07 mb:25 hour fcst: - 1: 17:313675:d=2021032312:VVEL:0.07 mb:25 hour fcst: - 1: 18:342395:d=2021032312:DZDT:0.07 mb:25 hour fcst: - 1: 19:367676:d=2021032312:UGRD:0.07 mb:25 hour fcst: - 1: 20:388085:d=2021032312:VGRD:0.07 mb:25 hour fcst: - 1: 21:408335:d=2021032312:ABSV:0.07 mb:25 hour fcst: - 1: 22:427092:d=2021032312:O3MR:0.07 mb:25 hour fcst: - 1: 23:446591:d=2021032312:HGT:0.1 mb:25 hour fcst: - 1: 24:468135:d=2021032312:TMP:0.1 mb:25 hour fcst: - 1: 25:484788:d=2021032312:RH:0.1 mb:25 hour fcst: - 1: 26:492877:d=2021032312:SPFH:0.1 mb:25 hour fcst: - 1: 27:514200:d=2021032312:VVEL:0.1 mb:25 hour fcst: - 1: 28:543780:d=2021032312:DZDT:0.1 mb:25 hour fcst: - 1: 29:569545:d=2021032312:UGRD:0.1 mb:25 hour fcst: - 1: 30:589626:d=2021032312:VGRD:0.1 mb:25 hour fcst: - 1: 31:609630:d=2021032312:ABSV:0.1 mb:25 hour fcst: - 1: + interp_atmos_master.sh[47]export err=0 - 1: + interp_atmos_master.sh[47]err=0 - 1: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 1: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 1: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_2_0p25 - 1: + product_functions.sh[5]local filename=pgb2file_f025_2_0p25 - 1: + product_functions.sh[6]wgrib2 pgb2file_f025_2_0p25 -not_if :RH: -grib pgb2file_f025_2_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_2_0p25.new - 1: 1:0:d=2021032312:ABSV:0.02 mb:25 hour fcst: - 1: 2:365290:d=2021032312:O3MR:0.02 mb:25 hour fcst: - 1: 3:693040:d=2021032312:HGT:0.04 mb:25 hour fcst: - 1: 4:1347550:d=2021032312:TMP:0.04 mb:25 hour fcst: - 1: 5:1742699:d=2021032312:RH:0.04 mb:25 hour fcst: - 1: 6:1887521:d=2021032312:SPFH:0.04 mb:25 hour fcst: - 1: 7:2608206:d=2021032312:VVEL:0.04 mb:25 hour fcst: - 1: 8:3370634:d=2021032312:DZDT:0.04 mb:25 hour fcst: - 1: 9:4032103:d=2021032312:UGRD:0.04 mb:25 hour fcst: - 1: 10:4523534:d=2021032312:VGRD:0.04 mb:25 hour fcst: - 1: 11:5002379:d=2021032312:ABSV:0.04 mb:25 hour fcst: - 1: 12:5418337:d=2021032312:O3MR:0.04 mb:25 hour fcst: - 1: 13:6044110:d=2021032312:HGT:0.07 mb:25 hour fcst: - 1: 14:6689394:d=2021032312:TMP:0.07 mb:25 hour fcst: - 1: 15:7070510:d=2021032312:RH:0.07 mb:25 hour fcst: - 1: 16:7295815:d=2021032312:SPFH:0.07 mb:25 hour fcst: - 1: 17:7892301:d=2021032312:VVEL:0.07 mb:25 hour fcst: - 1: 18:8684108:d=2021032312:DZDT:0.07 mb:25 hour fcst: - 1: 19:9377475:d=2021032312:UGRD:0.07 mb:25 hour fcst: - 1: 20:9875382:d=2021032312:VGRD:0.07 mb:25 hour fcst: - 1: 21:10370313:d=2021032312:ABSV:0.07 mb:25 hour fcst: - 1: 22:10802374:d=2021032312:O3MR:0.07 mb:25 hour fcst: - 1: 23:11277109:d=2021032312:HGT:0.1 mb:25 hour fcst: - 1: 24:11814712:d=2021032312:TMP:0.1 mb:25 hour fcst: - 1: 25:12185844:d=2021032312:RH:0.1 mb:25 hour fcst: - 1: 26:12388651:d=2021032312:SPFH:0.1 mb:25 hour fcst: - 1: 27:12900109:d=2021032312:VVEL:0.1 mb:25 hour fcst: - 1: 28:13731188:d=2021032312:DZDT:0.1 mb:25 hour fcst: - 1: 29:14441400:d=2021032312:UGRD:0.1 mb:25 hour fcst: - 1: 30:14926756:d=2021032312:VGRD:0.1 mb:25 hour fcst: - 1: 31:15410629:d=2021032312:ABSV:0.1 mb:25 hour fcst: - 1: + product_functions.sh[10]rc=0 - 1: + product_functions.sh[11](( rc == 0 )) - 1: + product_functions.sh[11]mv pgb2file_f025_2_0p25.new pgb2file_f025_2_0p25 - 1: + product_functions.sh[12]return 0 - 1: + interp_atmos_master.sh[56]export err=0 - 1: + interp_atmos_master.sh[56]err=0 - 1: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 1: ++ interp_atmos_master.sh[62]wc -l - 1: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_2_0p25 -match 'LAND|ICEC' - 1: + interp_atmos_master.sh[62]var_count=0 - 1: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 1: + interp_atmos_master.sh[73]exit 0 - 7: + bash[8]'[' -z '' ']' - 7: + bash[9]case "$-" in - 7: + bash[12]__lmod_vx=x - 7: + bash[16]'[' -n x ']' - 7: + bash[16]set +x - 7: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 7: Shell debugging restarted - 7: + bash[224]unset __lmod_vx - 7: + interp_atmos_master.sh[7]input_file=tmpfile_f025_8 - 7: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_8 - 7: + interp_atmos_master.sh[9]grid_string=0p25 - 7: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 7: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 7: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 7: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 7: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 7: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 7: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 7: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 7: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 7: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 7: + interp_atmos_master.sh[31]IFS=: - 7: + interp_atmos_master.sh[31]read -ra grids - 7: + interp_atmos_master.sh[33]output_grids= - 7: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 7: + interp_atmos_master.sh[35]gridopt=grid0p25 - 7: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_8_0p25' - 7: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_8 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_8_0p25 - 7: 1:0:d=2021032312:ABSV:70 mb:25 hour fcst: - 7: 2:17761:d=2021032312:O3MR:70 mb:25 hour fcst: - 7: 3:43617:d=2021032312:HGT:100 mb:25 hour fcst: - 7: 4:63223:d=2021032312:TMP:100 mb:25 hour fcst: - 7: 5:80173:d=2021032312:RH:100 mb:25 hour fcst: - 7: 6:90306:d=2021032312:TCDC:100 mb:25 hour fcst: - 7: 7:90485:d=2021032312:SPFH:100 mb:25 hour fcst: - 7: 8:116102:d=2021032312:VVEL:100 mb:25 hour fcst: - 7: 9:141098:d=2021032312:DZDT:100 mb:25 hour fcst: - 7: 10:171467:d=2021032312:UGRD:100 mb:25 hour fcst: - 7: 11:190846:d=2021032312:VGRD:100 mb:25 hour fcst: - 7: 12:210230:d=2021032312:ABSV:100 mb:25 hour fcst: - 7: 13:228425:d=2021032312:CLMR:100 mb:25 hour fcst: - 7: 14:228604:d=2021032312:ICMR:100 mb:25 hour fcst: - 7: 15:229309:d=2021032312:RWMR:100 mb:25 hour fcst: - 7: 16:229488:d=2021032312:SNMR:100 mb:25 hour fcst: - 7: 17:231585:d=2021032312:GRLE:100 mb:25 hour fcst: - 7: 18:231764:d=2021032312:O3MR:100 mb:25 hour fcst: - 7: 19:255777:d=2021032312:HGT:150 mb:25 hour fcst: - 7: 20:276120:d=2021032312:TMP:150 mb:25 hour fcst: - 7: 21:293097:d=2021032312:RH:150 mb:25 hour fcst: - 7: 22:304517:d=2021032312:TCDC:150 mb:25 hour fcst: - 7: 23:308505:d=2021032312:SPFH:150 mb:25 hour fcst: - 7: 24:332842:d=2021032312:VVEL:150 mb:25 hour fcst: - 7: 25:360125:d=2021032312:DZDT:150 mb:25 hour fcst: - 7: 26:385725:d=2021032312:UGRD:150 mb:25 hour fcst: - 7: 27:405665:d=2021032312:VGRD:150 mb:25 hour fcst: - 7: 28:425698:d=2021032312:ABSV:150 mb:25 hour fcst: - 7: 29:444566:d=2021032312:CLMR:150 mb:25 hour fcst: - 7: 30:444745:d=2021032312:ICMR:150 mb:25 hour fcst: - 7: 31:450000:d=2021032312:RWMR:150 mb:25 hour fcst: - 7: + interp_atmos_master.sh[47]export err=0 - 7: + interp_atmos_master.sh[47]err=0 - 7: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 7: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 7: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_8_0p25 - 7: + product_functions.sh[5]local filename=pgb2file_f025_8_0p25 - 7: + product_functions.sh[6]wgrib2 pgb2file_f025_8_0p25 -not_if :RH: -grib pgb2file_f025_8_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_8_0p25.new - 7: 1:0:d=2021032312:ABSV:70 mb:25 hour fcst: - 7: 2:399012:d=2021032312:O3MR:70 mb:25 hour fcst: - 7: 3:1110219:d=2021032312:HGT:100 mb:25 hour fcst: - 7: 4:1578240:d=2021032312:TMP:100 mb:25 hour fcst: - 7: 5:1955746:d=2021032312:RH:100 mb:25 hour fcst: - 7: 6:2191231:d=2021032312:TCDC:100 mb:25 hour fcst: - 7: 7:2191410:d=2021032312:SPFH:100 mb:25 hour fcst: - 7: 8:2850425:d=2021032312:VVEL:100 mb:25 hour fcst: - 7: 9:3521089:d=2021032312:DZDT:100 mb:25 hour fcst: - 7: 10:4419754:d=2021032312:UGRD:100 mb:25 hour fcst: - 7: 11:4886169:d=2021032312:VGRD:100 mb:25 hour fcst: - 7: 12:5350782:d=2021032312:ABSV:100 mb:25 hour fcst: - 7: 13:5757834:d=2021032312:CLMR:100 mb:25 hour fcst: - 7: 14:5758013:d=2021032312:ICMR:100 mb:25 hour fcst: - 7: 15:5767586:d=2021032312:RWMR:100 mb:25 hour fcst: - 7: 16:5767765:d=2021032312:SNMR:100 mb:25 hour fcst: - 7: 17:5810501:d=2021032312:GRLE:100 mb:25 hour fcst: - 7: 18:5810680:d=2021032312:O3MR:100 mb:25 hour fcst: - 7: 19:6426979:d=2021032312:HGT:150 mb:25 hour fcst: - 7: 20:6922533:d=2021032312:TMP:150 mb:25 hour fcst: - 7: 21:7303003:d=2021032312:RH:150 mb:25 hour fcst: - 7: 22:7571393:d=2021032312:TCDC:150 mb:25 hour fcst: - 7: 23:7654723:d=2021032312:SPFH:150 mb:25 hour fcst: - 7: 24:8286077:d=2021032312:VVEL:150 mb:25 hour fcst: - 7: 25:9040387:d=2021032312:DZDT:150 mb:25 hour fcst: - 7: 26:9736864:d=2021032312:UGRD:150 mb:25 hour fcst: - 7: 27:10229763:d=2021032312:VGRD:150 mb:25 hour fcst: - 7: 28:10730484:d=2021032312:ABSV:150 mb:25 hour fcst: - 7: 29:11161369:d=2021032312:CLMR:150 mb:25 hour fcst: - 7: 30:11161548:d=2021032312:ICMR:150 mb:25 hour fcst: - 7: 31:11285150:d=2021032312:RWMR:150 mb:25 hour fcst: - 7: + product_functions.sh[10]rc=0 - 7: + product_functions.sh[11](( rc == 0 )) - 7: + product_functions.sh[11]mv pgb2file_f025_8_0p25.new pgb2file_f025_8_0p25 - 7: + product_functions.sh[12]return 0 - 7: + interp_atmos_master.sh[56]export err=0 - 7: + interp_atmos_master.sh[56]err=0 - 7: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 7: ++ interp_atmos_master.sh[62]wc -l - 7: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_8_0p25 -match 'LAND|ICEC' - 7: + interp_atmos_master.sh[62]var_count=0 - 7: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 7: + interp_atmos_master.sh[73]exit 0 -21: + bash[8]'[' -z '' ']' -21: + bash[9]case "$-" in -21: + bash[12]__lmod_vx=x -21: + bash[16]'[' -n x ']' -21: + bash[16]set +x -21: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -21: Shell debugging restarted -21: + bash[224]unset __lmod_vx -21: + interp_atmos_master.sh[7]input_file=tmpfile_f025_22 -21: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_22 -21: + interp_atmos_master.sh[9]grid_string=0p25 -21: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -21: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -21: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -21: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -21: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -21: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -21: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -21: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -21: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -21: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -21: + interp_atmos_master.sh[31]IFS=: -21: + interp_atmos_master.sh[31]read -ra grids -21: + interp_atmos_master.sh[33]output_grids= -21: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -21: + interp_atmos_master.sh[35]gridopt=grid0p25 -21: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_22_0p25' -21: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_22 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_22_0p25 -21: 1:0:d=2021032312:DLWRF:surface:24-25 hour ave fcst: -21: 2:21711:d=2021032312:USWRF:surface:24-25 hour ave fcst: -21: 3:39003:d=2021032312:ULWRF:surface:24-25 hour ave fcst: -21: 4:57957:d=2021032312:USWRF:top of atmosphere:24-25 hour ave fcst: -21: 5:76041:d=2021032312:ULWRF:top of atmosphere:24-25 hour ave fcst: -21: 6:97187:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Total Aerosol:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 7:124470:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Dust Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 8:142589:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Sea Salt Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 9:169959:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Sulphate Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 10:195638:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Particulate Organic Matter Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 11:221265:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Black Carbon Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 12:243538:d=2021032312:HLCY:3000-0 m above ground:25 hour fcst: -21: 13:259402:d=2021032312:HLCY:1000-0 m above ground:25 hour fcst: -21: 14:281658:d=2021032312:USTM:6000-0 m above ground:25 hour fcst: -21: 15:304749:d=2021032312:VSTM:6000-0 m above ground:25 hour fcst: -21: 16:327581:d=2021032312:PRES:tropopause:25 hour fcst: -21: 17:357380:d=2021032312:ICAHT:tropopause:25 hour fcst: -21: 18:388209:d=2021032312:HGT:tropopause:25 hour fcst: -21: 19:416426:d=2021032312:TMP:tropopause:25 hour fcst: -21: 20:438237:d=2021032312:UGRD:tropopause:25 hour fcst: -21: 21:454226:d=2021032312:VGRD:tropopause:25 hour fcst: -21: 22:470120:d=2021032312:VWSH:tropopause:25 hour fcst: -21: 23:488628:d=2021032312:PRES:max wind:25 hour fcst: -21: 24:523116:d=2021032312:ICAHT:max wind:25 hour fcst: -21: 25:558318:d=2021032312:HGT:max wind:25 hour fcst: -21: 26:593669:d=2021032312:UGRD:max wind:25 hour fcst: -21: 27:610123:d=2021032312:VGRD:max wind:25 hour fcst: -21: 28:627571:d=2021032312:TMP:max wind:25 hour fcst: -21: 29:654320:d=2021032312:UGRD:20 m above ground:25 hour fcst: -21: 30:674870:d=2021032312:VGRD:20 m above ground:25 hour fcst: -21: 31:695830:d=2021032312:UGRD:30 m above ground:25 hour fcst: -21: 32:716720:d=2021032312:VGRD:30 m above ground:25 hour fcst: -21: + interp_atmos_master.sh[47]export err=0 -21: + interp_atmos_master.sh[47]err=0 -21: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -21: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -21: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_22_0p25 -21: + product_functions.sh[5]local filename=pgb2file_f025_22_0p25 -21: + product_functions.sh[6]wgrib2 pgb2file_f025_22_0p25 -not_if :RH: -grib pgb2file_f025_22_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_22_0p25.new -21: 1:0:d=2021032312:DLWRF:surface:24-25 hour ave fcst: -21: 2:528333:d=2021032312:USWRF:surface:24-25 hour ave fcst: -21: 3:964568:d=2021032312:ULWRF:surface:24-25 hour ave fcst: -21: 4:1394272:d=2021032312:USWRF:top of atmosphere:24-25 hour ave fcst: -21: 5:1857033:d=2021032312:ULWRF:top of atmosphere:24-25 hour ave fcst: -21: 6:2370877:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Total Aerosol:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 7:3115419:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Dust Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 8:3561803:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Sea Salt Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 9:4297439:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Sulphate Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 10:4975578:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Particulate Organic Matter Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 11:5634876:d=2021032312:AOTK:entire atmosphere:25 hour fcst:aerosol=Black Carbon Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 12:6188500:d=2021032312:HLCY:3000-0 m above ground:25 hour fcst: -21: 13:6526724:d=2021032312:HLCY:1000-0 m above ground:25 hour fcst: -21: 14:7064681:d=2021032312:USTM:6000-0 m above ground:25 hour fcst: -21: 15:7635891:d=2021032312:VSTM:6000-0 m above ground:25 hour fcst: -21: 16:8201878:d=2021032312:PRES:tropopause:25 hour fcst: -21: 17:9002282:d=2021032312:ICAHT:tropopause:25 hour fcst: -21: 18:9852897:d=2021032312:HGT:tropopause:25 hour fcst: -21: 19:10588064:d=2021032312:TMP:tropopause:25 hour fcst: -21: 20:11103846:d=2021032312:UGRD:tropopause:25 hour fcst: -21: 21:11446381:d=2021032312:VGRD:tropopause:25 hour fcst: -21: 22:11787387:d=2021032312:VWSH:tropopause:25 hour fcst: -21: 23:12201353:d=2021032312:PRES:max wind:25 hour fcst: -21: 24:13091486:d=2021032312:ICAHT:max wind:25 hour fcst: -21: 25:13993579:d=2021032312:HGT:max wind:25 hour fcst: -21: 26:14898744:d=2021032312:UGRD:max wind:25 hour fcst: -21: 27:15254955:d=2021032312:VGRD:max wind:25 hour fcst: -21: 28:15629535:d=2021032312:TMP:max wind:25 hour fcst: -21: 29:16292792:d=2021032312:UGRD:20 m above ground:25 hour fcst: -21: 30:16782605:d=2021032312:VGRD:20 m above ground:25 hour fcst: -21: 31:17296531:d=2021032312:UGRD:30 m above ground:25 hour fcst: -21: 32:17800147:d=2021032312:VGRD:30 m above ground:25 hour fcst: -21: + product_functions.sh[10]rc=0 -21: + product_functions.sh[11](( rc == 0 )) -21: + product_functions.sh[11]mv pgb2file_f025_22_0p25.new pgb2file_f025_22_0p25 -21: + product_functions.sh[12]return 0 -21: + interp_atmos_master.sh[56]export err=0 -21: + interp_atmos_master.sh[56]err=0 -21: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -21: ++ interp_atmos_master.sh[62]wc -l -21: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_22_0p25 -match 'LAND|ICEC' -21: + interp_atmos_master.sh[62]var_count=0 -21: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -21: + interp_atmos_master.sh[73]exit 0 -12: + bash[8]'[' -z '' ']' -12: + bash[9]case "$-" in -12: + bash[12]__lmod_vx=x -12: + bash[16]'[' -n x ']' -12: + bash[16]set +x -12: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -12: Shell debugging restarted -12: + bash[224]unset __lmod_vx -12: + interp_atmos_master.sh[7]input_file=tmpfile_f025_13 -12: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_13 -12: + interp_atmos_master.sh[9]grid_string=0p25 -12: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -12: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -12: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -12: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -12: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -12: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -12: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -12: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -12: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -12: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -12: + interp_atmos_master.sh[31]IFS=: -12: + interp_atmos_master.sh[31]read -ra grids -12: + interp_atmos_master.sh[33]output_grids= -12: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -12: + interp_atmos_master.sh[35]gridopt=grid0p25 -12: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_13_0p25' -12: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_13 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_13_0p25 -12: 1:0:d=2021032312:ABSV:550 mb:25 hour fcst: -12: 2:20950:d=2021032312:CLMR:550 mb:25 hour fcst: -12: 3:24023:d=2021032312:ICMR:550 mb:25 hour fcst: -12: 4:40651:d=2021032312:RWMR:550 mb:25 hour fcst: -12: 5:42277:d=2021032312:SNMR:550 mb:25 hour fcst: -12: 6:56852:d=2021032312:GRLE:550 mb:25 hour fcst: -12: 7:59040:d=2021032312:O3MR:550 mb:25 hour fcst: -12: 8:81899:d=2021032312:HGT:600 mb:25 hour fcst: -12: 9:106309:d=2021032312:TMP:600 mb:25 hour fcst: -12: 10:125360:d=2021032312:RH:600 mb:25 hour fcst: -12: 11:145433:d=2021032312:TCDC:600 mb:25 hour fcst: -12: 12:157427:d=2021032312:SPFH:600 mb:25 hour fcst: -12: 13:187886:d=2021032312:VVEL:600 mb:25 hour fcst: -12: 14:213293:d=2021032312:DZDT:600 mb:25 hour fcst: -12: 15:240721:d=2021032312:UGRD:600 mb:25 hour fcst: -12: 16:261800:d=2021032312:VGRD:600 mb:25 hour fcst: -12: 17:283403:d=2021032312:ABSV:600 mb:25 hour fcst: -12: 18:304142:d=2021032312:CLMR:600 mb:25 hour fcst: -12: 19:309110:d=2021032312:ICMR:600 mb:25 hour fcst: -12: 20:322957:d=2021032312:RWMR:600 mb:25 hour fcst: -12: 21:326677:d=2021032312:SNMR:600 mb:25 hour fcst: -12: 22:342457:d=2021032312:GRLE:600 mb:25 hour fcst: -12: 23:344921:d=2021032312:O3MR:600 mb:25 hour fcst: -12: 24:372868:d=2021032312:HGT:650 mb:25 hour fcst: -12: 25:397011:d=2021032312:TMP:650 mb:25 hour fcst: -12: 26:416319:d=2021032312:RH:650 mb:25 hour fcst: -12: 27:436561:d=2021032312:TCDC:650 mb:25 hour fcst: -12: 28:448324:d=2021032312:SPFH:650 mb:25 hour fcst: -12: 29:477204:d=2021032312:VVEL:650 mb:25 hour fcst: -12: 30:502563:d=2021032312:DZDT:650 mb:25 hour fcst: -12: 31:529968:d=2021032312:UGRD:650 mb:25 hour fcst: -12: 32:550894:d=2021032312:VGRD:650 mb:25 hour fcst: -12: + interp_atmos_master.sh[47]export err=0 -12: + interp_atmos_master.sh[47]err=0 -12: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -12: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -12: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_13_0p25 -12: + product_functions.sh[5]local filename=pgb2file_f025_13_0p25 -12: + product_functions.sh[6]wgrib2 pgb2file_f025_13_0p25 -not_if :RH: -grib pgb2file_f025_13_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_13_0p25.new -12: 1:0:d=2021032312:ABSV:550 mb:25 hour fcst: -12: 2:495145:d=2021032312:CLMR:550 mb:25 hour fcst: -12: 3:561902:d=2021032312:ICMR:550 mb:25 hour fcst: -12: 4:928845:d=2021032312:RWMR:550 mb:25 hour fcst: -12: 5:955563:d=2021032312:SNMR:550 mb:25 hour fcst: -12: 6:1268353:d=2021032312:GRLE:550 mb:25 hour fcst: -12: 7:1306043:d=2021032312:O3MR:550 mb:25 hour fcst: -12: 8:1879121:d=2021032312:HGT:600 mb:25 hour fcst: -12: 9:2511583:d=2021032312:TMP:600 mb:25 hour fcst: -12: 10:2937349:d=2021032312:RH:600 mb:25 hour fcst: -12: 11:3402795:d=2021032312:TCDC:600 mb:25 hour fcst: -12: 12:3628829:d=2021032312:SPFH:600 mb:25 hour fcst: -12: 13:4454762:d=2021032312:VVEL:600 mb:25 hour fcst: -12: 14:5127255:d=2021032312:DZDT:600 mb:25 hour fcst: -12: 15:5885967:d=2021032312:UGRD:600 mb:25 hour fcst: -12: 16:6393152:d=2021032312:VGRD:600 mb:25 hour fcst: -12: 17:6929126:d=2021032312:ABSV:600 mb:25 hour fcst: -12: 18:7412323:d=2021032312:CLMR:600 mb:25 hour fcst: -12: 19:7514347:d=2021032312:ICMR:600 mb:25 hour fcst: -12: 20:7819517:d=2021032312:RWMR:600 mb:25 hour fcst: -12: 21:7896638:d=2021032312:SNMR:600 mb:25 hour fcst: -12: 22:8239941:d=2021032312:GRLE:600 mb:25 hour fcst: -12: 23:8283950:d=2021032312:O3MR:600 mb:25 hour fcst: -12: 24:9050247:d=2021032312:HGT:650 mb:25 hour fcst: -12: 25:9679760:d=2021032312:TMP:650 mb:25 hour fcst: -12: 26:10115733:d=2021032312:RH:650 mb:25 hour fcst: -12: 27:10586872:d=2021032312:TCDC:650 mb:25 hour fcst: -12: 28:10811461:d=2021032312:SPFH:650 mb:25 hour fcst: -12: 29:11585307:d=2021032312:VVEL:650 mb:25 hour fcst: -12: 30:12261928:d=2021032312:DZDT:650 mb:25 hour fcst: -12: 31:13017057:d=2021032312:UGRD:650 mb:25 hour fcst: -12: 32:13521699:d=2021032312:VGRD:650 mb:25 hour fcst: -12: + product_functions.sh[10]rc=0 -12: + product_functions.sh[11](( rc == 0 )) -12: + product_functions.sh[11]mv pgb2file_f025_13_0p25.new pgb2file_f025_13_0p25 -12: + product_functions.sh[12]return 0 -12: + interp_atmos_master.sh[56]export err=0 -12: + interp_atmos_master.sh[56]err=0 -12: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -12: ++ interp_atmos_master.sh[62]wc -l -12: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_13_0p25 -match 'LAND|ICEC' -12: + interp_atmos_master.sh[62]var_count=0 -12: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -12: + interp_atmos_master.sh[73]exit 0 -19: + bash[8]'[' -z '' ']' -19: + bash[9]case "$-" in -19: + bash[12]__lmod_vx=x -19: + bash[16]'[' -n x ']' -19: + bash[16]set +x -19: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -19: Shell debugging restarted -19: + bash[224]unset __lmod_vx -19: + interp_atmos_master.sh[7]input_file=tmpfile_f025_20 -19: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_20 -19: + interp_atmos_master.sh[9]grid_string=0p25 -19: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -19: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -19: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -19: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -19: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -19: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -19: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -19: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -19: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -19: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -19: + interp_atmos_master.sh[31]IFS=: -19: + interp_atmos_master.sh[31]read -ra grids -19: + interp_atmos_master.sh[33]output_grids= -19: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -19: + interp_atmos_master.sh[35]gridopt=grid0p25 -19: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_20_0p25' -19: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_20 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_20_0p25 -19: 1:0:d=2021032312:CPRAT:surface:24-25 hour ave fcst: -19: 2:23804:d=2021032312:PRATE:surface:24-25 hour ave fcst: -19: 3:49119:d=2021032312:APCP:surface:24-25 hour acc fcst: -19: 4:55304:d=2021032312:APCP:surface:0-25 hour acc fcst: -19: 5:68578:d=2021032312:ACPCP:surface:24-25 hour acc fcst: -19: 6:73569:d=2021032312:ACPCP:surface:0-25 hour acc fcst: -19: 7:84615:d=2021032312:FROZR:surface:0-25 hour acc fcst: -19: 8:90966:d=2021032312:FRZR:surface:0-25 hour acc fcst: -19: 9:94108:d=2021032312:TSNOWP:surface:0-25 hour acc fcst: -19: 10:101716:d=2021032312:WATR:surface:24-25 hour acc fcst: -19: 11:114459:d=2021032312:FROZR:surface:24-25 hour acc fcst: -19: 12:117399:d=2021032312:FRZR:surface:24-25 hour acc fcst: -19: 13:118548:d=2021032312:TSNOWP:surface:24-25 hour acc fcst: -19: 14:124110:d=2021032312:CSNOW:surface:25 hour fcst: -19: 15:125910:d=2021032312:CICEP:surface:25 hour fcst: -19: 16:126089:d=2021032312:CFRZR:surface:25 hour fcst: -19: 17:126309:d=2021032312:CRAIN:surface:25 hour fcst: -19: 18:129815:d=2021032312:LHTFL:surface:24-25 hour ave fcst: -19: 19:151621:d=2021032312:SHTFL:surface:24-25 hour ave fcst: -19: 20:172617:d=2021032312:GFLUX:surface:24-25 hour ave fcst: -19: 21:184689:d=2021032312:UFLX:surface:24-25 hour ave fcst: -19: 22:201090:d=2021032312:VFLX:surface:24-25 hour ave fcst: -19: 23:217606:d=2021032312:SFCR:surface:25 hour fcst: -19: 24:236029:d=2021032312:FRICV:surface:25 hour fcst: -19: 25:257766:d=2021032312:U-GWD:surface:24-25 hour ave fcst: -19: 26:272638:d=2021032312:V-GWD:surface:24-25 hour ave fcst: -19: 27:287578:d=2021032312:VEG:surface:25 hour fcst: -19: 28:295612:d=2021032312:SOTYP:surface:25 hour fcst: -19: 29:306661:d=2021032312:WILT:surface:25 hour fcst: -19: 30:316571:d=2021032312:FLDCP:surface:25 hour fcst: -19: 31:326777:d=2021032312:SUNSD:surface:25 hour fcst: -19: + interp_atmos_master.sh[47]export err=0 -19: + interp_atmos_master.sh[47]err=0 -19: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -19: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -19: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_20_0p25 -19: + product_functions.sh[5]local filename=pgb2file_f025_20_0p25 -19: + product_functions.sh[6]wgrib2 pgb2file_f025_20_0p25 -not_if :RH: -grib pgb2file_f025_20_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_20_0p25.new -19: 1:0:d=2021032312:CPRAT:surface:24-25 hour ave fcst: -19: 2:621389:d=2021032312:PRATE:surface:24-25 hour ave fcst: -19: 3:1267274:d=2021032312:APCP:surface:24-25 hour acc fcst: -19: 4:1377233:d=2021032312:APCP:surface:0-25 hour acc fcst: -19: 5:1638561:d=2021032312:ACPCP:surface:24-25 hour acc fcst: -19: 6:1721124:d=2021032312:ACPCP:surface:0-25 hour acc fcst: -19: 7:1938603:d=2021032312:FROZR:surface:0-25 hour acc fcst: -19: 8:2061742:d=2021032312:FRZR:surface:0-25 hour acc fcst: -19: 9:2116904:d=2021032312:TSNOWP:surface:0-25 hour acc fcst: -19: 10:2259917:d=2021032312:WATR:surface:24-25 hour acc fcst: -19: 11:2634188:d=2021032312:FROZR:surface:24-25 hour acc fcst: -19: 12:2687169:d=2021032312:FRZR:surface:24-25 hour acc fcst: -19: 13:2700975:d=2021032312:TSNOWP:surface:24-25 hour acc fcst: -19: 14:2798146:d=2021032312:CSNOW:surface:25 hour fcst: -19: 15:2818482:d=2021032312:CICEP:surface:25 hour fcst: -19: 16:2818661:d=2021032312:CFRZR:surface:25 hour fcst: -19: 17:2818940:d=2021032312:CRAIN:surface:25 hour fcst: -19: 18:2884324:d=2021032312:LHTFL:surface:24-25 hour ave fcst: -19: 19:3436032:d=2021032312:SHTFL:surface:24-25 hour ave fcst: -19: 20:3936570:d=2021032312:GFLUX:surface:24-25 hour ave fcst: -19: 21:4309118:d=2021032312:UFLX:surface:24-25 hour ave fcst: -19: 22:4674686:d=2021032312:VFLX:surface:24-25 hour ave fcst: -19: 23:5039587:d=2021032312:SFCR:surface:25 hour fcst: -19: 24:5430789:d=2021032312:FRICV:surface:25 hour fcst: -19: 25:5932467:d=2021032312:U-GWD:surface:24-25 hour ave fcst: -19: 26:6267433:d=2021032312:V-GWD:surface:24-25 hour ave fcst: -19: 27:6601699:d=2021032312:VEG:surface:25 hour fcst: -19: 28:6874882:d=2021032312:SOTYP:surface:25 hour fcst: -19: 29:7112749:d=2021032312:WILT:surface:25 hour fcst: -19: 30:7425230:d=2021032312:FLDCP:surface:25 hour fcst: -19: 31:7746991:d=2021032312:SUNSD:surface:25 hour fcst: -19: + product_functions.sh[10]rc=0 -19: + product_functions.sh[11](( rc == 0 )) -19: + product_functions.sh[11]mv pgb2file_f025_20_0p25.new pgb2file_f025_20_0p25 -19: + product_functions.sh[12]return 0 -19: + interp_atmos_master.sh[56]export err=0 -19: + interp_atmos_master.sh[56]err=0 -19: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -19: ++ interp_atmos_master.sh[62]wc -l -19: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_20_0p25 -match 'LAND|ICEC' -19: + interp_atmos_master.sh[62]var_count=0 -19: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -19: + interp_atmos_master.sh[73]exit 0 - 8: + bash[8]'[' -z '' ']' - 8: + bash[9]case "$-" in - 8: + bash[12]__lmod_vx=x - 8: + bash[16]'[' -n x ']' - 8: + bash[16]set +x - 8: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 8: Shell debugging restarted - 8: + bash[224]unset __lmod_vx - 8: + interp_atmos_master.sh[7]input_file=tmpfile_f025_9 - 8: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_9 - 8: + interp_atmos_master.sh[9]grid_string=0p25 - 8: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 8: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 8: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 8: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 8: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 8: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 8: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 8: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 8: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 8: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 8: + interp_atmos_master.sh[31]IFS=: - 8: + interp_atmos_master.sh[31]read -ra grids - 8: + interp_atmos_master.sh[33]output_grids= - 8: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 8: + interp_atmos_master.sh[35]gridopt=grid0p25 - 8: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_9_0p25' - 8: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_9 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_9_0p25 - 8: 1:0:d=2021032312:SNMR:150 mb:25 hour fcst: - 8: 2:3468:d=2021032312:GRLE:150 mb:25 hour fcst: - 8: 3:3994:d=2021032312:O3MR:150 mb:25 hour fcst: - 8: 4:25875:d=2021032312:HGT:200 mb:25 hour fcst: - 8: 5:46925:d=2021032312:TMP:200 mb:25 hour fcst: - 8: 6:65051:d=2021032312:RH:200 mb:25 hour fcst: - 8: 7:80111:d=2021032312:TCDC:200 mb:25 hour fcst: - 8: 8:86556:d=2021032312:SPFH:200 mb:25 hour fcst: - 8: 9:112310:d=2021032312:VVEL:200 mb:25 hour fcst: - 8: 10:139169:d=2021032312:DZDT:200 mb:25 hour fcst: - 8: 11:165461:d=2021032312:UGRD:200 mb:25 hour fcst: - 8: 12:178831:d=2021032312:VGRD:200 mb:25 hour fcst: - 8: 13:192587:d=2021032312:ABSV:200 mb:25 hour fcst: - 8: 14:212626:d=2021032312:CLMR:200 mb:25 hour fcst: - 8: 15:212805:d=2021032312:ICMR:200 mb:25 hour fcst: - 8: 16:221315:d=2021032312:RWMR:200 mb:25 hour fcst: - 8: 17:221494:d=2021032312:SNMR:200 mb:25 hour fcst: - 8: 18:226821:d=2021032312:GRLE:200 mb:25 hour fcst: - 8: 19:227179:d=2021032312:O3MR:200 mb:25 hour fcst: - 8: 20:247785:d=2021032312:HGT:250 mb:25 hour fcst: - 8: 21:269568:d=2021032312:TMP:250 mb:25 hour fcst: - 8: 22:288417:d=2021032312:RH:250 mb:25 hour fcst: - 8: 23:306244:d=2021032312:TCDC:250 mb:25 hour fcst: - 8: 24:316869:d=2021032312:SPFH:250 mb:25 hour fcst: - 8: 25:344912:d=2021032312:VVEL:250 mb:25 hour fcst: - 8: 26:373290:d=2021032312:DZDT:250 mb:25 hour fcst: - 8: 27:400437:d=2021032312:UGRD:250 mb:25 hour fcst: - 8: 28:414567:d=2021032312:VGRD:250 mb:25 hour fcst: - 8: 29:429422:d=2021032312:ABSV:250 mb:25 hour fcst: - 8: 30:450995:d=2021032312:CLMR:250 mb:25 hour fcst: - 8: 31:451174:d=2021032312:ICMR:250 mb:25 hour fcst: - 8: + interp_atmos_master.sh[47]export err=0 - 8: + interp_atmos_master.sh[47]err=0 - 8: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 8: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 8: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_9_0p25 - 8: + product_functions.sh[5]local filename=pgb2file_f025_9_0p25 - 8: + product_functions.sh[6]wgrib2 pgb2file_f025_9_0p25 -not_if :RH: -grib pgb2file_f025_9_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_9_0p25.new - 8: 1:0:d=2021032312:SNMR:150 mb:25 hour fcst: - 8: 2:57171:d=2021032312:GRLE:150 mb:25 hour fcst: - 8: 3:64503:d=2021032312:O3MR:150 mb:25 hour fcst: - 8: 4:594790:d=2021032312:HGT:200 mb:25 hour fcst: - 8: 5:1120605:d=2021032312:TMP:200 mb:25 hour fcst: - 8: 6:1532272:d=2021032312:RH:200 mb:25 hour fcst: - 8: 7:1881130:d=2021032312:TCDC:200 mb:25 hour fcst: - 8: 8:2015478:d=2021032312:SPFH:200 mb:25 hour fcst: - 8: 9:2689657:d=2021032312:VVEL:200 mb:25 hour fcst: - 8: 10:3412387:d=2021032312:DZDT:200 mb:25 hour fcst: - 8: 11:4149531:d=2021032312:UGRD:200 mb:25 hour fcst: - 8: 12:4453235:d=2021032312:VGRD:200 mb:25 hour fcst: - 8: 13:4762519:d=2021032312:ABSV:200 mb:25 hour fcst: - 8: 14:5234480:d=2021032312:CLMR:200 mb:25 hour fcst: - 8: 15:5234659:d=2021032312:ICMR:200 mb:25 hour fcst: - 8: 16:5415411:d=2021032312:RWMR:200 mb:25 hour fcst: - 8: 17:5415590:d=2021032312:SNMR:200 mb:25 hour fcst: - 8: 18:5517107:d=2021032312:GRLE:200 mb:25 hour fcst: - 8: 19:5520247:d=2021032312:O3MR:200 mb:25 hour fcst: - 8: 20:6008990:d=2021032312:HGT:250 mb:25 hour fcst: - 8: 21:6557464:d=2021032312:TMP:250 mb:25 hour fcst: - 8: 22:6980459:d=2021032312:RH:250 mb:25 hour fcst: - 8: 23:7388192:d=2021032312:TCDC:250 mb:25 hour fcst: - 8: 24:7596312:d=2021032312:SPFH:250 mb:25 hour fcst: - 8: 25:8351760:d=2021032312:VVEL:250 mb:25 hour fcst: - 8: 26:9134630:d=2021032312:DZDT:250 mb:25 hour fcst: - 8: 27:9898582:d=2021032312:UGRD:250 mb:25 hour fcst: - 8: 28:10212472:d=2021032312:VGRD:250 mb:25 hour fcst: - 8: 29:10541021:d=2021032312:ABSV:250 mb:25 hour fcst: - 8: 30:11067855:d=2021032312:CLMR:250 mb:25 hour fcst: - 8: 31:11068034:d=2021032312:ICMR:250 mb:25 hour fcst: - 8: + product_functions.sh[10]rc=0 - 8: + product_functions.sh[11](( rc == 0 )) - 8: + product_functions.sh[11]mv pgb2file_f025_9_0p25.new pgb2file_f025_9_0p25 - 8: + product_functions.sh[12]return 0 - 8: + interp_atmos_master.sh[56]export err=0 - 8: + interp_atmos_master.sh[56]err=0 - 8: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 8: ++ interp_atmos_master.sh[62]wc -l - 8: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_9_0p25 -match 'LAND|ICEC' - 8: + interp_atmos_master.sh[62]var_count=0 - 8: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 8: + interp_atmos_master.sh[73]exit 0 -16: + bash[8]'[' -z '' ']' -16: + bash[9]case "$-" in -16: + bash[12]__lmod_vx=x -16: + bash[16]'[' -n x ']' -16: + bash[16]set +x -16: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -16: Shell debugging restarted -16: + bash[224]unset __lmod_vx -16: + interp_atmos_master.sh[7]input_file=tmpfile_f025_17 -16: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_17 -16: + interp_atmos_master.sh[9]grid_string=0p25 -16: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -16: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -16: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -16: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -16: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -16: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -16: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -16: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -16: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -16: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -16: + interp_atmos_master.sh[31]IFS=: -16: + interp_atmos_master.sh[31]read -ra grids -16: + interp_atmos_master.sh[33]output_grids= -16: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -16: + interp_atmos_master.sh[35]gridopt=grid0p25 -16: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_17_0p25' -16: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_17 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_17_0p25 -16: 1:0:d=2021032312:ABSV:925 mb:25 hour fcst: -16: 2:21276:d=2021032312:CLMR:925 mb:25 hour fcst: -16: 3:29865:d=2021032312:ICMR:925 mb:25 hour fcst: -16: 4:37480:d=2021032312:RWMR:925 mb:25 hour fcst: -16: 5:49779:d=2021032312:SNMR:925 mb:25 hour fcst: -16: 6:59388:d=2021032312:GRLE:925 mb:25 hour fcst: -16: 7:62150:d=2021032312:O3MR:925 mb:25 hour fcst: -16: 8:88099:d=2021032312:HGT:950 mb:25 hour fcst: -16: 9:113410:d=2021032312:TMP:950 mb:25 hour fcst: -16: 10:134492:d=2021032312:RH:950 mb:25 hour fcst: -16: 11:153965:d=2021032312:TCDC:950 mb:25 hour fcst: -16: 12:163913:d=2021032312:SPFH:950 mb:25 hour fcst: -16: 13:191081:d=2021032312:VVEL:950 mb:25 hour fcst: -16: 14:214904:d=2021032312:DZDT:950 mb:25 hour fcst: -16: 15:239768:d=2021032312:UGRD:950 mb:25 hour fcst: -16: 16:261060:d=2021032312:VGRD:950 mb:25 hour fcst: -16: 17:283003:d=2021032312:ABSV:950 mb:25 hour fcst: -16: 18:304080:d=2021032312:CLMR:950 mb:25 hour fcst: -16: 19:310616:d=2021032312:ICMR:950 mb:25 hour fcst: -16: 20:317719:d=2021032312:RWMR:950 mb:25 hour fcst: -16: 21:330068:d=2021032312:SNMR:950 mb:25 hour fcst: -16: 22:338877:d=2021032312:GRLE:950 mb:25 hour fcst: -16: 23:341729:d=2021032312:O3MR:950 mb:25 hour fcst: -16: 24:367644:d=2021032312:HINDEX:surface:25 hour fcst: -16: 25:372741:d=2021032312:HGT:975 mb:25 hour fcst: -16: 26:398201:d=2021032312:TMP:975 mb:25 hour fcst: -16: 27:419108:d=2021032312:RH:975 mb:25 hour fcst: -16: 28:438227:d=2021032312:TCDC:975 mb:25 hour fcst: -16: 29:446366:d=2021032312:SPFH:975 mb:25 hour fcst: -16: 30:473440:d=2021032312:VVEL:975 mb:25 hour fcst: -16: 31:496136:d=2021032312:DZDT:975 mb:25 hour fcst: -16: + interp_atmos_master.sh[47]export err=0 -16: + interp_atmos_master.sh[47]err=0 -16: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -16: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -16: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_17_0p25 -16: + product_functions.sh[5]local filename=pgb2file_f025_17_0p25 -16: + product_functions.sh[6]wgrib2 pgb2file_f025_17_0p25 -not_if :RH: -grib pgb2file_f025_17_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_17_0p25.new -16: 1:0:d=2021032312:ABSV:925 mb:25 hour fcst: -16: 2:504971:d=2021032312:CLMR:925 mb:25 hour fcst: -16: 3:696771:d=2021032312:ICMR:925 mb:25 hour fcst: -16: 4:872937:d=2021032312:RWMR:925 mb:25 hour fcst: -16: 5:1123999:d=2021032312:SNMR:925 mb:25 hour fcst: -16: 6:1340785:d=2021032312:GRLE:925 mb:25 hour fcst: -16: 7:1392705:d=2021032312:O3MR:925 mb:25 hour fcst: -16: 8:2097967:d=2021032312:HGT:950 mb:25 hour fcst: -16: 9:2770785:d=2021032312:TMP:950 mb:25 hour fcst: -16: 10:3266324:d=2021032312:RH:950 mb:25 hour fcst: -16: 11:3709726:d=2021032312:TCDC:950 mb:25 hour fcst: -16: 12:3908172:d=2021032312:SPFH:950 mb:25 hour fcst: -16: 13:4647007:d=2021032312:VVEL:950 mb:25 hour fcst: -16: 14:5258432:d=2021032312:DZDT:950 mb:25 hour fcst: -16: 15:5907827:d=2021032312:UGRD:950 mb:25 hour fcst: -16: 16:6427787:d=2021032312:VGRD:950 mb:25 hour fcst: -16: 17:6973362:d=2021032312:ABSV:950 mb:25 hour fcst: -16: 18:7477010:d=2021032312:CLMR:950 mb:25 hour fcst: -16: 19:7621123:d=2021032312:ICMR:950 mb:25 hour fcst: -16: 20:7788207:d=2021032312:RWMR:950 mb:25 hour fcst: -16: 21:8042408:d=2021032312:SNMR:950 mb:25 hour fcst: -16: 22:8243368:d=2021032312:GRLE:950 mb:25 hour fcst: -16: 23:8298008:d=2021032312:O3MR:950 mb:25 hour fcst: -16: 24:9000370:d=2021032312:HINDEX:surface:25 hour fcst: -16: 25:9181573:d=2021032312:HGT:975 mb:25 hour fcst: -16: 26:9861275:d=2021032312:TMP:975 mb:25 hour fcst: -16: 27:10351469:d=2021032312:RH:975 mb:25 hour fcst: -16: 28:10780025:d=2021032312:TCDC:975 mb:25 hour fcst: -16: 29:10936918:d=2021032312:SPFH:975 mb:25 hour fcst: -16: 30:11670016:d=2021032312:VVEL:975 mb:25 hour fcst: -16: 31:12242024:d=2021032312:DZDT:975 mb:25 hour fcst: -16: + product_functions.sh[10]rc=0 -16: + product_functions.sh[11](( rc == 0 )) -16: + product_functions.sh[11]mv pgb2file_f025_17_0p25.new pgb2file_f025_17_0p25 -16: + product_functions.sh[12]return 0 -16: + interp_atmos_master.sh[56]export err=0 -16: + interp_atmos_master.sh[56]err=0 -16: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -16: ++ interp_atmos_master.sh[62]wc -l -16: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_17_0p25 -match 'LAND|ICEC' -16: + interp_atmos_master.sh[62]var_count=0 -16: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -16: + interp_atmos_master.sh[73]exit 0 -18: + bash[8]'[' -z '' ']' -18: + bash[9]case "$-" in -18: + bash[12]__lmod_vx=x -18: + bash[16]'[' -n x ']' -18: + bash[16]set +x -18: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -18: Shell debugging restarted -18: + bash[224]unset __lmod_vx -18: + interp_atmos_master.sh[7]input_file=tmpfile_f025_19 -18: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_19 -18: + interp_atmos_master.sh[9]grid_string=0p25 -18: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -18: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -18: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -18: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -18: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -18: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -18: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -18: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -18: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -18: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -18: + interp_atmos_master.sh[31]IFS=: -18: + interp_atmos_master.sh[31]read -ra grids -18: + interp_atmos_master.sh[33]output_grids= -18: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -18: + interp_atmos_master.sh[35]gridopt=grid0p25 -18: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_19_0p25' -18: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_19 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_19_0p25 -18: 1:0:d=2021032312:TMP:surface:25 hour fcst: -18: 2:16643:d=2021032312:TSOIL:0-0.1 m below ground:25 hour fcst: -18: 3:28567:d=2021032312:SOILW:0-0.1 m below ground:25 hour fcst: -18: 4:37481:d=2021032312:SOILL:0-0.1 m below ground:25 hour fcst: -18: 5:46287:d=2021032312:TSOIL:0.1-0.4 m below ground:25 hour fcst: -18: 6:61763:d=2021032312:SOILW:0.1-0.4 m below ground:25 hour fcst: -18: 7:70779:d=2021032312:SOILL:0.1-0.4 m below ground:25 hour fcst: -18: 8:79673:d=2021032312:TSOIL:0.4-1 m below ground:25 hour fcst: -18: 9:94119:d=2021032312:SOILW:0.4-1 m below ground:25 hour fcst: -18: 10:103256:d=2021032312:SOILL:0.4-1 m below ground:25 hour fcst: -18: 11:112175:d=2021032312:TSOIL:1-2 m below ground:25 hour fcst: -18: 12:126461:d=2021032312:SOILW:1-2 m below ground:25 hour fcst: -18: 13:135524:d=2021032312:SOILL:1-2 m below ground:25 hour fcst: -18: 14:144602:d=2021032312:CNWAT:surface:25 hour fcst: -18: 15:150146:d=2021032312:WEASD:surface:25 hour fcst: -18: 16:162959:d=2021032312:SNOD:surface:25 hour fcst: -18: 17:177321:d=2021032312:PEVPR:surface:25 hour fcst: -18: 18:193089:d=2021032312:ICETK:surface:25 hour fcst: -18: 19:196506:d=2021032312:TMP:2 m above ground:25 hour fcst: -18: 20:217307:d=2021032312:SPFH:2 m above ground:25 hour fcst: -18: 21:246225:d=2021032312:DPT:2 m above ground:25 hour fcst: -18: 22:267491:d=2021032312:RH:2 m above ground:25 hour fcst: -18: 23:286023:d=2021032312:APTMP:2 m above ground:25 hour fcst: -18: 24:300421:d=2021032312:TMAX:2 m above ground:24-25 hour max fcst: -18: 25:321180:d=2021032312:TMIN:2 m above ground:24-25 hour min fcst: -18: 26:342000:d=2021032312:UGRD:10 m above ground:25 hour fcst: -18: 27:362511:d=2021032312:VGRD:10 m above ground:25 hour fcst: -18: 28:383409:d=2021032312:ICEG:10 m above mean sea level:25 hour fcst: -18: 29:384044:d=2021032312:CPOFP:surface:25 hour fcst: -18: 30:404417:d=2021032312:CPRAT:surface:25 hour fcst: -18: 31:425918:d=2021032312:PRATE:surface:25 hour fcst: -18: + interp_atmos_master.sh[47]export err=0 -18: + interp_atmos_master.sh[47]err=0 -18: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -18: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -18: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_19_0p25 -18: + product_functions.sh[5]local filename=pgb2file_f025_19_0p25 -18: + product_functions.sh[6]wgrib2 pgb2file_f025_19_0p25 -not_if :RH: -grib pgb2file_f025_19_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_19_0p25.new -18: 1:0:d=2021032312:TMP:surface:25 hour fcst: -18: 2:343057:d=2021032312:TSOIL:0-0.1 m below ground:25 hour fcst: -18: 3:677858:d=2021032312:SOILW:0-0.1 m below ground:25 hour fcst: -18: 4:964664:d=2021032312:SOILL:0-0.1 m below ground:25 hour fcst: -18: 5:1253440:d=2021032312:TSOIL:0.1-0.4 m below ground:25 hour fcst: -18: 6:1712379:d=2021032312:SOILW:0.1-0.4 m below ground:25 hour fcst: -18: 7:2003896:d=2021032312:SOILL:0.1-0.4 m below ground:25 hour fcst: -18: 8:2297967:d=2021032312:TSOIL:0.4-1 m below ground:25 hour fcst: -18: 9:2735551:d=2021032312:SOILW:0.4-1 m below ground:25 hour fcst: -18: 10:3031968:d=2021032312:SOILL:0.4-1 m below ground:25 hour fcst: -18: 11:3327562:d=2021032312:TSOIL:1-2 m below ground:25 hour fcst: -18: 12:3757266:d=2021032312:SOILW:1-2 m below ground:25 hour fcst: -18: 13:4048214:d=2021032312:SOILL:1-2 m below ground:25 hour fcst: -18: 14:4344233:d=2021032312:CNWAT:surface:25 hour fcst: -18: 15:4530369:d=2021032312:WEASD:surface:25 hour fcst: -18: 16:4899056:d=2021032312:SNOD:surface:25 hour fcst: -18: 17:5292513:d=2021032312:PEVPR:surface:25 hour fcst: -18: 18:5775776:d=2021032312:ICETK:surface:25 hour fcst: -18: 19:5829214:d=2021032312:TMP:2 m above ground:25 hour fcst: -18: 20:6311978:d=2021032312:SPFH:2 m above ground:25 hour fcst: -18: 21:7097709:d=2021032312:DPT:2 m above ground:25 hour fcst: -18: 22:7603910:d=2021032312:RH:2 m above ground:25 hour fcst: -18: 23:8013788:d=2021032312:APTMP:2 m above ground:25 hour fcst: -18: 24:8325501:d=2021032312:TMAX:2 m above ground:24-25 hour max fcst: -18: 25:8806980:d=2021032312:TMIN:2 m above ground:24-25 hour min fcst: -18: 26:9290004:d=2021032312:UGRD:10 m above ground:25 hour fcst: -18: 27:9780539:d=2021032312:VGRD:10 m above ground:25 hour fcst: -18: 28:10291979:d=2021032312:ICEG:10 m above mean sea level:25 hour fcst: -18: 29:10302843:d=2021032312:CPOFP:surface:25 hour fcst: -18: 30:10724248:d=2021032312:CPRAT:surface:25 hour fcst: -18: 31:11261138:d=2021032312:PRATE:surface:25 hour fcst: -18: + product_functions.sh[10]rc=0 -18: + product_functions.sh[11](( rc == 0 )) -18: + product_functions.sh[11]mv pgb2file_f025_19_0p25.new pgb2file_f025_19_0p25 -18: + product_functions.sh[12]return 0 -18: + interp_atmos_master.sh[56]export err=0 -18: + interp_atmos_master.sh[56]err=0 -18: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -18: ++ interp_atmos_master.sh[62]wc -l -18: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_19_0p25 -match 'LAND|ICEC' -18: + interp_atmos_master.sh[62]var_count=0 -18: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -18: + interp_atmos_master.sh[73]exit 0 -13: + bash[8]'[' -z '' ']' -13: + bash[9]case "$-" in -13: + bash[12]__lmod_vx=x -13: + bash[16]'[' -n x ']' -13: + bash[16]set +x -13: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -13: Shell debugging restarted -13: + bash[224]unset __lmod_vx -13: + interp_atmos_master.sh[7]input_file=tmpfile_f025_14 -13: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_14 -13: + interp_atmos_master.sh[9]grid_string=0p25 -13: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -13: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -13: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -13: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -13: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -13: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -13: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -13: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -13: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -13: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -13: + interp_atmos_master.sh[31]IFS=: -13: + interp_atmos_master.sh[31]read -ra grids -13: + interp_atmos_master.sh[33]output_grids= -13: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -13: + interp_atmos_master.sh[35]gridopt=grid0p25 -13: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_14_0p25' -13: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_14 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_14_0p25 -13: 1:0:d=2021032312:ABSV:650 mb:25 hour fcst: -13: 2:20855:d=2021032312:CLMR:650 mb:25 hour fcst: -13: 3:26084:d=2021032312:ICMR:650 mb:25 hour fcst: -13: 4:38501:d=2021032312:RWMR:650 mb:25 hour fcst: -13: 5:43154:d=2021032312:SNMR:650 mb:25 hour fcst: -13: 6:58080:d=2021032312:GRLE:650 mb:25 hour fcst: -13: 7:60381:d=2021032312:O3MR:650 mb:25 hour fcst: -13: 8:87894:d=2021032312:HGT:700 mb:25 hour fcst: -13: 9:112048:d=2021032312:TMP:700 mb:25 hour fcst: -13: 10:131831:d=2021032312:RH:700 mb:25 hour fcst: -13: 11:152160:d=2021032312:TCDC:700 mb:25 hour fcst: -13: 12:163751:d=2021032312:SPFH:700 mb:25 hour fcst: -13: 13:192358:d=2021032312:VVEL:700 mb:25 hour fcst: -13: 14:217928:d=2021032312:DZDT:700 mb:25 hour fcst: -13: 15:245301:d=2021032312:UGRD:700 mb:25 hour fcst: -13: 16:266157:d=2021032312:VGRD:700 mb:25 hour fcst: -13: 17:287763:d=2021032312:ABSV:700 mb:25 hour fcst: -13: 18:308798:d=2021032312:CLMR:700 mb:25 hour fcst: -13: 19:315809:d=2021032312:ICMR:700 mb:25 hour fcst: -13: 20:327004:d=2021032312:RWMR:700 mb:25 hour fcst: -13: 21:332617:d=2021032312:SNMR:700 mb:25 hour fcst: -13: 22:346745:d=2021032312:GRLE:700 mb:25 hour fcst: -13: 23:348636:d=2021032312:O3MR:700 mb:25 hour fcst: -13: 24:375958:d=2021032312:HGT:750 mb:25 hour fcst: -13: 25:400392:d=2021032312:TMP:750 mb:25 hour fcst: -13: 26:420300:d=2021032312:RH:750 mb:25 hour fcst: -13: 27:440711:d=2021032312:TCDC:750 mb:25 hour fcst: -13: 28:453105:d=2021032312:SPFH:750 mb:25 hour fcst: -13: 29:482255:d=2021032312:VVEL:750 mb:25 hour fcst: -13: 30:508017:d=2021032312:DZDT:750 mb:25 hour fcst: -13: 31:535287:d=2021032312:UGRD:750 mb:25 hour fcst: -13: 32:556303:d=2021032312:VGRD:750 mb:25 hour fcst: -13: + interp_atmos_master.sh[47]export err=0 -13: + interp_atmos_master.sh[47]err=0 -13: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -13: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -13: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_14_0p25 -13: + product_functions.sh[5]local filename=pgb2file_f025_14_0p25 -13: + product_functions.sh[6]wgrib2 pgb2file_f025_14_0p25 -not_if :RH: -grib pgb2file_f025_14_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_14_0p25.new -13: 1:0:d=2021032312:ABSV:650 mb:25 hour fcst: -13: 2:491758:d=2021032312:CLMR:650 mb:25 hour fcst: -13: 3:596656:d=2021032312:ICMR:650 mb:25 hour fcst: -13: 4:872669:d=2021032312:RWMR:650 mb:25 hour fcst: -13: 5:966947:d=2021032312:SNMR:650 mb:25 hour fcst: -13: 6:1293777:d=2021032312:GRLE:650 mb:25 hour fcst: -13: 7:1333366:d=2021032312:O3MR:650 mb:25 hour fcst: -13: 8:2088219:d=2021032312:HGT:700 mb:25 hour fcst: -13: 9:2716209:d=2021032312:TMP:700 mb:25 hour fcst: -13: 10:3167183:d=2021032312:RH:700 mb:25 hour fcst: -13: 11:3641253:d=2021032312:TCDC:700 mb:25 hour fcst: -13: 12:3872693:d=2021032312:SPFH:700 mb:25 hour fcst: -13: 13:4645745:d=2021032312:VVEL:700 mb:25 hour fcst: -13: 14:5327812:d=2021032312:DZDT:700 mb:25 hour fcst: -13: 15:6081968:d=2021032312:UGRD:700 mb:25 hour fcst: -13: 16:6585416:d=2021032312:VGRD:700 mb:25 hour fcst: -13: 17:7117329:d=2021032312:ABSV:700 mb:25 hour fcst: -13: 18:7608863:d=2021032312:CLMR:700 mb:25 hour fcst: -13: 19:7756157:d=2021032312:ICMR:700 mb:25 hour fcst: -13: 20:8011368:d=2021032312:RWMR:700 mb:25 hour fcst: -13: 21:8120358:d=2021032312:SNMR:700 mb:25 hour fcst: -13: 22:8431307:d=2021032312:GRLE:700 mb:25 hour fcst: -13: 23:8462486:d=2021032312:O3MR:700 mb:25 hour fcst: -13: 24:9195134:d=2021032312:HGT:750 mb:25 hour fcst: -13: 25:9825438:d=2021032312:TMP:750 mb:25 hour fcst: -13: 26:10289306:d=2021032312:RH:750 mb:25 hour fcst: -13: 27:10768846:d=2021032312:TCDC:750 mb:25 hour fcst: -13: 28:11010869:d=2021032312:SPFH:750 mb:25 hour fcst: -13: 29:11800388:d=2021032312:VVEL:750 mb:25 hour fcst: -13: 30:12485091:d=2021032312:DZDT:750 mb:25 hour fcst: -13: 31:13235689:d=2021032312:UGRD:750 mb:25 hour fcst: -13: 32:13741051:d=2021032312:VGRD:750 mb:25 hour fcst: -13: + product_functions.sh[10]rc=0 -13: + product_functions.sh[11](( rc == 0 )) -13: + product_functions.sh[11]mv pgb2file_f025_14_0p25.new pgb2file_f025_14_0p25 -13: + product_functions.sh[12]return 0 -13: + interp_atmos_master.sh[56]export err=0 -13: + interp_atmos_master.sh[56]err=0 -13: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -13: ++ interp_atmos_master.sh[62]wc -l -13: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_14_0p25 -match 'LAND|ICEC' -13: + interp_atmos_master.sh[62]var_count=0 -13: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -13: + interp_atmos_master.sh[73]exit 0 -10: + bash[8]'[' -z '' ']' -10: + bash[9]case "$-" in -10: + bash[12]__lmod_vx=x -10: + bash[16]'[' -n x ']' -10: + bash[16]set +x -10: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -10: Shell debugging restarted -10: + bash[224]unset __lmod_vx -10: + interp_atmos_master.sh[7]input_file=tmpfile_f025_11 -10: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_11 -10: + interp_atmos_master.sh[9]grid_string=0p25 -10: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -10: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -10: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -10: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -10: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -10: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -10: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -10: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -10: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -10: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -10: + interp_atmos_master.sh[31]IFS=: -10: + interp_atmos_master.sh[31]read -ra grids -10: + interp_atmos_master.sh[33]output_grids= -10: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -10: + interp_atmos_master.sh[35]gridopt=grid0p25 -10: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_11_0p25' -10: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_11 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_11_0p25 -10: 1:0:d=2021032312:ICMR:350 mb:25 hour fcst: -10: 2:16988:d=2021032312:RWMR:350 mb:25 hour fcst: -10: 3:17167:d=2021032312:SNMR:350 mb:25 hour fcst: -10: 4:29192:d=2021032312:GRLE:350 mb:25 hour fcst: -10: 5:29716:d=2021032312:O3MR:350 mb:25 hour fcst: -10: 6:55274:d=2021032312:HGT:400 mb:25 hour fcst: -10: 7:77162:d=2021032312:TMP:400 mb:25 hour fcst: -10: 8:95483:d=2021032312:RH:400 mb:25 hour fcst: -10: 9:114673:d=2021032312:TCDC:400 mb:25 hour fcst: -10: 10:127891:d=2021032312:SPFH:400 mb:25 hour fcst: -10: 11:155664:d=2021032312:VVEL:400 mb:25 hour fcst: -10: 12:180637:d=2021032312:DZDT:400 mb:25 hour fcst: -10: 13:208727:d=2021032312:UGRD:400 mb:25 hour fcst: -10: 14:230705:d=2021032312:VGRD:400 mb:25 hour fcst: -10: 15:253425:d=2021032312:ABSV:400 mb:25 hour fcst: -10: 16:275369:d=2021032312:CLMR:400 mb:25 hour fcst: -10: 17:275752:d=2021032312:ICMR:400 mb:25 hour fcst: -10: 18:292731:d=2021032312:RWMR:400 mb:25 hour fcst: -10: 19:293027:d=2021032312:SNMR:400 mb:25 hour fcst: -10: 20:306943:d=2021032312:GRLE:400 mb:25 hour fcst: -10: 21:307828:d=2021032312:O3MR:400 mb:25 hour fcst: -10: 22:332394:d=2021032312:HGT:450 mb:25 hour fcst: -10: 23:356346:d=2021032312:TMP:450 mb:25 hour fcst: -10: 24:374911:d=2021032312:RH:450 mb:25 hour fcst: -10: 25:394298:d=2021032312:TCDC:450 mb:25 hour fcst: -10: 26:407082:d=2021032312:SPFH:450 mb:25 hour fcst: -10: 27:436403:d=2021032312:VVEL:450 mb:25 hour fcst: -10: 28:461578:d=2021032312:DZDT:450 mb:25 hour fcst: -10: 29:489606:d=2021032312:UGRD:450 mb:25 hour fcst: -10: 30:511285:d=2021032312:VGRD:450 mb:25 hour fcst: -10: 31:533600:d=2021032312:ABSV:450 mb:25 hour fcst: -10: + interp_atmos_master.sh[47]export err=0 -10: + interp_atmos_master.sh[47]err=0 -10: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -10: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -10: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_11_0p25 -10: + product_functions.sh[5]local filename=pgb2file_f025_11_0p25 -10: + product_functions.sh[6]wgrib2 pgb2file_f025_11_0p25 -not_if :RH: -grib pgb2file_f025_11_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_11_0p25.new -10: 1:0:d=2021032312:ICMR:350 mb:25 hour fcst: -10: 2:373481:d=2021032312:RWMR:350 mb:25 hour fcst: -10: 3:373660:d=2021032312:SNMR:350 mb:25 hour fcst: -10: 4:616340:d=2021032312:GRLE:350 mb:25 hour fcst: -10: 5:623921:d=2021032312:O3MR:350 mb:25 hour fcst: -10: 6:1293747:d=2021032312:HGT:400 mb:25 hour fcst: -10: 7:1829954:d=2021032312:TMP:400 mb:25 hour fcst: -10: 8:2240078:d=2021032312:RH:400 mb:25 hour fcst: -10: 9:2676657:d=2021032312:TCDC:400 mb:25 hour fcst: -10: 10:2919057:d=2021032312:SPFH:400 mb:25 hour fcst: -10: 11:3660164:d=2021032312:VVEL:400 mb:25 hour fcst: -10: 12:4321809:d=2021032312:DZDT:400 mb:25 hour fcst: -10: 13:5113496:d=2021032312:UGRD:400 mb:25 hour fcst: -10: 14:5658065:d=2021032312:VGRD:400 mb:25 hour fcst: -10: 15:6233054:d=2021032312:ABSV:400 mb:25 hour fcst: -10: 16:6768868:d=2021032312:CLMR:400 mb:25 hour fcst: -10: 17:6772910:d=2021032312:ICMR:400 mb:25 hour fcst: -10: 18:7149905:d=2021032312:RWMR:400 mb:25 hour fcst: -10: 19:7151999:d=2021032312:SNMR:400 mb:25 hour fcst: -10: 20:7447245:d=2021032312:GRLE:400 mb:25 hour fcst: -10: 21:7457885:d=2021032312:O3MR:400 mb:25 hour fcst: -10: 22:8091150:d=2021032312:HGT:450 mb:25 hour fcst: -10: 23:8713570:d=2021032312:TMP:450 mb:25 hour fcst: -10: 24:9127778:d=2021032312:RH:450 mb:25 hour fcst: -10: 25:9569404:d=2021032312:TCDC:450 mb:25 hour fcst: -10: 26:9808720:d=2021032312:SPFH:450 mb:25 hour fcst: -10: 27:10607539:d=2021032312:VVEL:450 mb:25 hour fcst: -10: 28:11273216:d=2021032312:DZDT:450 mb:25 hour fcst: -10: 29:12058775:d=2021032312:UGRD:450 mb:25 hour fcst: -10: 30:12592940:d=2021032312:VGRD:450 mb:25 hour fcst: -10: 31:13156842:d=2021032312:ABSV:450 mb:25 hour fcst: -10: + product_functions.sh[10]rc=0 -10: + product_functions.sh[11](( rc == 0 )) -10: + product_functions.sh[11]mv pgb2file_f025_11_0p25.new pgb2file_f025_11_0p25 -10: + product_functions.sh[12]return 0 -10: + interp_atmos_master.sh[56]export err=0 -10: + interp_atmos_master.sh[56]err=0 -10: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -10: ++ interp_atmos_master.sh[62]wc -l -10: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_11_0p25 -match 'LAND|ICEC' -10: + interp_atmos_master.sh[62]var_count=0 -10: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -10: + interp_atmos_master.sh[73]exit 0 -20: + bash[8]'[' -z '' ']' -20: + bash[9]case "$-" in -20: + bash[12]__lmod_vx=x -20: + bash[16]'[' -n x ']' -20: + bash[16]set +x -20: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -20: Shell debugging restarted -20: + bash[224]unset __lmod_vx -20: + interp_atmos_master.sh[7]input_file=tmpfile_f025_21 -20: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_21 -20: + interp_atmos_master.sh[9]grid_string=0p25 -20: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -20: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -20: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -20: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -20: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -20: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -20: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -20: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -20: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -20: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -20: + interp_atmos_master.sh[31]IFS=: -20: + interp_atmos_master.sh[31]read -ra grids -20: + interp_atmos_master.sh[33]output_grids= -20: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -20: + interp_atmos_master.sh[35]gridopt=grid0p25 -20: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_21_0p25' -20: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_21 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_21_0p25 -20: 1:0:d=2021032312:LFTX:surface:25 hour fcst: -20: 2:13827:d=2021032312:CAPE:surface:25 hour fcst: -20: 3:29007:d=2021032312:CIN:surface:25 hour fcst: -20: 4:41871:d=2021032312:PWAT:entire atmosphere (considered as a single layer):25 hour fcst: -20: 5:68812:d=2021032312:CWAT:entire atmosphere (considered as a single layer):25 hour fcst: -20: 6:95326:d=2021032312:RH:entire atmosphere (considered as a single layer):25 hour fcst: -20: 7:110444:d=2021032312:TOZNE:entire atmosphere (considered as a single layer):25 hour fcst: -20: 8:124783:d=2021032312:LCDC:low cloud layer:25 hour fcst: -20: 9:142663:d=2021032312:LCDC:low cloud layer:24-25 hour ave fcst: -20: 10:160480:d=2021032312:MCDC:middle cloud layer:25 hour fcst: -20: 11:176305:d=2021032312:MCDC:middle cloud layer:24-25 hour ave fcst: -20: 12:191770:d=2021032312:HCDC:high cloud layer:25 hour fcst: -20: 13:206573:d=2021032312:HCDC:high cloud layer:24-25 hour ave fcst: -20: 14:222416:d=2021032312:TCDC:entire atmosphere (considered as a single layer):25 hour fcst: -20: 15:239620:d=2021032312:TCDC:entire atmosphere (considered as a single layer):24-25 hour ave fcst: -20: 16:256850:d=2021032312:HGT:cloud ceiling:25 hour fcst: -20: 17:288943:d=2021032312:PRES:convective cloud bottom level:25 hour fcst: -20: 18:302601:d=2021032312:PRES:low cloud bottom level:24-25 hour ave fcst: -20: 19:323689:d=2021032312:PRES:middle cloud bottom level:24-25 hour ave fcst: -20: 20:339979:d=2021032312:PRES:high cloud bottom level:24-25 hour ave fcst: -20: 21:356749:d=2021032312:PRES:convective cloud top level:25 hour fcst: -20: 22:372112:d=2021032312:PRES:low cloud top level:24-25 hour ave fcst: -20: 23:393200:d=2021032312:PRES:middle cloud top level:24-25 hour ave fcst: -20: 24:409311:d=2021032312:PRES:high cloud top level:24-25 hour ave fcst: -20: 25:426606:d=2021032312:TMP:low cloud top level:24-25 hour ave fcst: -20: 26:442489:d=2021032312:TMP:middle cloud top level:24-25 hour ave fcst: -20: 27:455534:d=2021032312:TMP:high cloud top level:24-25 hour ave fcst: -20: 28:469902:d=2021032312:TCDC:convective cloud layer:25 hour fcst: -20: 29:490537:d=2021032312:TCDC:boundary layer cloud layer:24-25 hour ave fcst: -20: 30:505893:d=2021032312:CWORK:entire atmosphere (considered as a single layer):24-25 hour ave fcst: -20: 31:518442:d=2021032312:DSWRF:surface:24-25 hour ave fcst: -20: + interp_atmos_master.sh[47]export err=0 -20: + interp_atmos_master.sh[47]err=0 -20: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -20: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -20: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_21_0p25 -20: + product_functions.sh[5]local filename=pgb2file_f025_21_0p25 -20: + product_functions.sh[6]wgrib2 pgb2file_f025_21_0p25 -not_if :RH: -grib pgb2file_f025_21_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_21_0p25.new -20: 1:0:d=2021032312:LFTX:surface:25 hour fcst: -20: 2:298349:d=2021032312:CAPE:surface:25 hour fcst: -20: 3:629950:d=2021032312:CIN:surface:25 hour fcst: -20: 4:873536:d=2021032312:PWAT:entire atmosphere (considered as a single layer):25 hour fcst: -20: 5:1604564:d=2021032312:CWAT:entire atmosphere (considered as a single layer):25 hour fcst: -20: 6:2270233:d=2021032312:RH:entire atmosphere (considered as a single layer):25 hour fcst: -20: 7:2607133:d=2021032312:TOZNE:entire atmosphere (considered as a single layer):25 hour fcst: -20: 8:2915712:d=2021032312:LCDC:low cloud layer:25 hour fcst: -20: 9:3250742:d=2021032312:LCDC:low cloud layer:24-25 hour ave fcst: -20: 10:3584480:d=2021032312:MCDC:middle cloud layer:25 hour fcst: -20: 11:3855270:d=2021032312:MCDC:middle cloud layer:24-25 hour ave fcst: -20: 12:4128433:d=2021032312:HCDC:high cloud layer:25 hour fcst: -20: 13:4376055:d=2021032312:HCDC:high cloud layer:24-25 hour ave fcst: -20: 14:4636592:d=2021032312:TCDC:entire atmosphere (considered as a single layer):25 hour fcst: -20: 15:4948964:d=2021032312:TCDC:entire atmosphere (considered as a single layer):24-25 hour ave fcst: -20: 16:5262024:d=2021032312:HGT:cloud ceiling:25 hour fcst: -20: 17:5936421:d=2021032312:PRES:convective cloud bottom level:25 hour fcst: -20: 18:6428956:d=2021032312:PRES:low cloud bottom level:24-25 hour ave fcst: -20: 19:7101975:d=2021032312:PRES:middle cloud bottom level:24-25 hour ave fcst: -20: 20:7616358:d=2021032312:PRES:high cloud bottom level:24-25 hour ave fcst: -20: 21:8150014:d=2021032312:PRES:convective cloud top level:25 hour fcst: -20: 22:8694235:d=2021032312:PRES:low cloud top level:24-25 hour ave fcst: -20: 23:9348205:d=2021032312:PRES:middle cloud top level:24-25 hour ave fcst: -20: 24:9829262:d=2021032312:PRES:high cloud top level:24-25 hour ave fcst: -20: 25:10380333:d=2021032312:TMP:low cloud top level:24-25 hour ave fcst: -20: 26:10875195:d=2021032312:TMP:middle cloud top level:24-25 hour ave fcst: -20: 27:11278731:d=2021032312:TMP:high cloud top level:24-25 hour ave fcst: -20: 28:11751867:d=2021032312:TCDC:convective cloud layer:25 hour fcst: -20: 29:12227443:d=2021032312:TCDC:boundary layer cloud layer:24-25 hour ave fcst: -20: 30:12512052:d=2021032312:CWORK:entire atmosphere (considered as a single layer):24-25 hour ave fcst: -20: 31:12803358:d=2021032312:DSWRF:surface:24-25 hour ave fcst: -20: + product_functions.sh[10]rc=0 -20: + product_functions.sh[11](( rc == 0 )) -20: + product_functions.sh[11]mv pgb2file_f025_21_0p25.new pgb2file_f025_21_0p25 -20: + product_functions.sh[12]return 0 -20: + interp_atmos_master.sh[56]export err=0 -20: + interp_atmos_master.sh[56]err=0 -20: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -20: ++ interp_atmos_master.sh[62]wc -l -20: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_21_0p25 -match 'LAND|ICEC' -20: + interp_atmos_master.sh[62]var_count=0 -20: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -20: + interp_atmos_master.sh[73]exit 0 -14: + bash[8]'[' -z '' ']' -14: + bash[9]case "$-" in -14: + bash[12]__lmod_vx=x -14: + bash[16]'[' -n x ']' -14: + bash[16]set +x -14: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -14: Shell debugging restarted -14: + bash[224]unset __lmod_vx -14: + interp_atmos_master.sh[7]input_file=tmpfile_f025_15 -14: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_15 -14: + interp_atmos_master.sh[9]grid_string=0p25 -14: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -14: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -14: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -14: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -14: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -14: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -14: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -14: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -14: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -14: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -14: + interp_atmos_master.sh[31]IFS=: -14: + interp_atmos_master.sh[31]read -ra grids -14: + interp_atmos_master.sh[33]output_grids= -14: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -14: + interp_atmos_master.sh[35]gridopt=grid0p25 -14: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_15_0p25' -14: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_15 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_15_0p25 -14: 1:0:d=2021032312:ABSV:750 mb:25 hour fcst: -14: 2:21156:d=2021032312:CLMR:750 mb:25 hour fcst: -14: 3:30298:d=2021032312:ICMR:750 mb:25 hour fcst: -14: 4:41944:d=2021032312:RWMR:750 mb:25 hour fcst: -14: 5:47832:d=2021032312:SNMR:750 mb:25 hour fcst: -14: 6:61455:d=2021032312:GRLE:750 mb:25 hour fcst: -14: 7:63610:d=2021032312:O3MR:750 mb:25 hour fcst: -14: 8:90863:d=2021032312:HGT:800 mb:25 hour fcst: -14: 9:115393:d=2021032312:TMP:800 mb:25 hour fcst: -14: 10:135946:d=2021032312:RH:800 mb:25 hour fcst: -14: 11:156397:d=2021032312:TCDC:800 mb:25 hour fcst: -14: 12:169840:d=2021032312:SPFH:800 mb:25 hour fcst: -14: 13:199307:d=2021032312:VVEL:800 mb:25 hour fcst: -14: 14:224990:d=2021032312:DZDT:800 mb:25 hour fcst: -14: 15:252044:d=2021032312:UGRD:800 mb:25 hour fcst: -14: 16:273160:d=2021032312:VGRD:800 mb:25 hour fcst: -14: 17:294911:d=2021032312:ABSV:800 mb:25 hour fcst: -14: 18:316095:d=2021032312:CLMR:800 mb:25 hour fcst: -14: 19:327850:d=2021032312:ICMR:800 mb:25 hour fcst: -14: 20:338889:d=2021032312:RWMR:800 mb:25 hour fcst: -14: 21:347251:d=2021032312:SNMR:800 mb:25 hour fcst: -14: 22:359999:d=2021032312:GRLE:800 mb:25 hour fcst: -14: 23:362414:d=2021032312:O3MR:800 mb:25 hour fcst: -14: 24:389401:d=2021032312:HGT:850 mb:25 hour fcst: -14: 25:414151:d=2021032312:TMP:850 mb:25 hour fcst: -14: 26:435029:d=2021032312:RH:850 mb:25 hour fcst: -14: 27:455278:d=2021032312:TCDC:850 mb:25 hour fcst: -14: 28:469886:d=2021032312:SPFH:850 mb:25 hour fcst: -14: 29:499438:d=2021032312:VVEL:850 mb:25 hour fcst: -14: 30:524757:d=2021032312:DZDT:850 mb:25 hour fcst: -14: 31:551248:d=2021032312:UGRD:850 mb:25 hour fcst: -14: 32:572431:d=2021032312:VGRD:850 mb:25 hour fcst: -14: + interp_atmos_master.sh[47]export err=0 -14: + interp_atmos_master.sh[47]err=0 -14: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -14: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -14: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_15_0p25 -14: + product_functions.sh[5]local filename=pgb2file_f025_15_0p25 -14: + product_functions.sh[6]wgrib2 pgb2file_f025_15_0p25 -not_if :RH: -grib pgb2file_f025_15_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_15_0p25.new -14: 1:0:d=2021032312:ABSV:750 mb:25 hour fcst: -14: 2:502823:d=2021032312:CLMR:750 mb:25 hour fcst: -14: 3:701456:d=2021032312:ICMR:750 mb:25 hour fcst: -14: 4:982235:d=2021032312:RWMR:750 mb:25 hour fcst: -14: 5:1092351:d=2021032312:SNMR:750 mb:25 hour fcst: -14: 6:1388475:d=2021032312:GRLE:750 mb:25 hour fcst: -14: 7:1426867:d=2021032312:O3MR:750 mb:25 hour fcst: -14: 8:2158325:d=2021032312:HGT:800 mb:25 hour fcst: -14: 9:2792783:d=2021032312:TMP:800 mb:25 hour fcst: -14: 10:3271419:d=2021032312:RH:800 mb:25 hour fcst: -14: 11:3753231:d=2021032312:TCDC:800 mb:25 hour fcst: -14: 12:4022743:d=2021032312:SPFH:800 mb:25 hour fcst: -14: 13:4820806:d=2021032312:VVEL:800 mb:25 hour fcst: -14: 14:5506084:d=2021032312:DZDT:800 mb:25 hour fcst: -14: 15:6249471:d=2021032312:UGRD:800 mb:25 hour fcst: -14: 16:6757153:d=2021032312:VGRD:800 mb:25 hour fcst: -14: 17:7294737:d=2021032312:ABSV:800 mb:25 hour fcst: -14: 18:7806580:d=2021032312:CLMR:800 mb:25 hour fcst: -14: 19:8061909:d=2021032312:ICMR:800 mb:25 hour fcst: -14: 20:8317576:d=2021032312:RWMR:800 mb:25 hour fcst: -14: 21:8477152:d=2021032312:SNMR:800 mb:25 hour fcst: -14: 22:8761109:d=2021032312:GRLE:800 mb:25 hour fcst: -14: 23:8805898:d=2021032312:O3MR:800 mb:25 hour fcst: -14: 24:9529667:d=2021032312:HGT:850 mb:25 hour fcst: -14: 25:10169732:d=2021032312:TMP:850 mb:25 hour fcst: -14: 26:10661281:d=2021032312:RH:850 mb:25 hour fcst: -14: 27:11130261:d=2021032312:TCDC:850 mb:25 hour fcst: -14: 28:11428391:d=2021032312:SPFH:850 mb:25 hour fcst: -14: 29:12232648:d=2021032312:VVEL:850 mb:25 hour fcst: -14: 30:12911772:d=2021032312:DZDT:850 mb:25 hour fcst: -14: 31:13645401:d=2021032312:UGRD:850 mb:25 hour fcst: -14: 32:14156134:d=2021032312:VGRD:850 mb:25 hour fcst: -14: + product_functions.sh[10]rc=0 -14: + product_functions.sh[11](( rc == 0 )) -14: + product_functions.sh[11]mv pgb2file_f025_15_0p25.new pgb2file_f025_15_0p25 -14: + product_functions.sh[12]return 0 -14: + interp_atmos_master.sh[56]export err=0 -14: + interp_atmos_master.sh[56]err=0 -14: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -14: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_15_0p25 -match 'LAND|ICEC' -14: ++ interp_atmos_master.sh[62]wc -l -14: + interp_atmos_master.sh[62]var_count=0 -14: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -14: + interp_atmos_master.sh[73]exit 0 - 5: + bash[8]'[' -z '' ']' - 5: + bash[9]case "$-" in - 5: + bash[12]__lmod_vx=x - 5: + bash[16]'[' -n x ']' - 5: + bash[16]set +x - 5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 5: Shell debugging restarted - 5: + bash[224]unset __lmod_vx - 5: + interp_atmos_master.sh[7]input_file=tmpfile_f025_6 - 5: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_6 - 5: + interp_atmos_master.sh[9]grid_string=0p25 - 5: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 5: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 5: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 5: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 5: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 5: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 5: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 5: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 5: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 5: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 5: + interp_atmos_master.sh[31]IFS=: - 5: + interp_atmos_master.sh[31]read -ra grids - 5: + interp_atmos_master.sh[33]output_grids= - 5: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 5: + interp_atmos_master.sh[35]gridopt=grid0p25 - 5: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_6_0p25' - 5: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_6 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_6_0p25 - 5: 1:0:d=2021032312:RH:15 mb:25 hour fcst: - 5: 2:5286:d=2021032312:SPFH:15 mb:25 hour fcst: - 5: 3:20733:d=2021032312:VVEL:15 mb:25 hour fcst: - 5: 4:47389:d=2021032312:DZDT:15 mb:25 hour fcst: - 5: 5:77895:d=2021032312:UGRD:15 mb:25 hour fcst: - 5: 6:89093:d=2021032312:VGRD:15 mb:25 hour fcst: - 5: 7:107536:d=2021032312:ABSV:15 mb:25 hour fcst: - 5: 8:125086:d=2021032312:O3MR:15 mb:25 hour fcst: - 5: 9:144017:d=2021032312:HGT:20 mb:25 hour fcst: - 5: 10:163429:d=2021032312:TMP:20 mb:25 hour fcst: - 5: 11:179444:d=2021032312:RH:20 mb:25 hour fcst: - 5: 12:185613:d=2021032312:SPFH:20 mb:25 hour fcst: - 5: 13:199789:d=2021032312:VVEL:20 mb:25 hour fcst: - 5: 14:227288:d=2021032312:DZDT:20 mb:25 hour fcst: - 5: 15:257523:d=2021032312:UGRD:20 mb:25 hour fcst: - 5: 16:276133:d=2021032312:VGRD:20 mb:25 hour fcst: - 5: 17:294638:d=2021032312:ABSV:20 mb:25 hour fcst: - 5: 18:312197:d=2021032312:O3MR:20 mb:25 hour fcst: - 5: 19:337066:d=2021032312:HGT:30 mb:25 hour fcst: - 5: 20:356173:d=2021032312:TMP:30 mb:25 hour fcst: - 5: 21:372106:d=2021032312:RH:30 mb:25 hour fcst: - 5: 22:379715:d=2021032312:SPFH:30 mb:25 hour fcst: - 5: 23:395046:d=2021032312:VVEL:30 mb:25 hour fcst: - 5: 24:423378:d=2021032312:DZDT:30 mb:25 hour fcst: - 5: 25:453274:d=2021032312:UGRD:30 mb:25 hour fcst: - 5: 26:471979:d=2021032312:VGRD:30 mb:25 hour fcst: - 5: 27:490478:d=2021032312:ABSV:30 mb:25 hour fcst: - 5: 28:508182:d=2021032312:O3MR:30 mb:25 hour fcst: - 5: 29:534925:d=2021032312:HGT:40 mb:25 hour fcst: - 5: 30:553909:d=2021032312:TMP:40 mb:25 hour fcst: - 5: 31:570224:d=2021032312:RH:40 mb:25 hour fcst: - 5: + interp_atmos_master.sh[47]export err=0 - 5: + interp_atmos_master.sh[47]err=0 - 5: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 5: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 5: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_6_0p25 - 5: + product_functions.sh[5]local filename=pgb2file_f025_6_0p25 - 5: + product_functions.sh[6]wgrib2 pgb2file_f025_6_0p25 -not_if :RH: -grib pgb2file_f025_6_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_6_0p25.new - 5: 1:0:d=2021032312:RH:15 mb:25 hour fcst: - 5: 2:116150:d=2021032312:SPFH:15 mb:25 hour fcst: - 5: 3:489638:d=2021032312:VVEL:15 mb:25 hour fcst: - 5: 4:1229986:d=2021032312:DZDT:15 mb:25 hour fcst: - 5: 5:2149180:d=2021032312:UGRD:15 mb:25 hour fcst: - 5: 6:2587393:d=2021032312:VGRD:15 mb:25 hour fcst: - 5: 7:3014652:d=2021032312:ABSV:15 mb:25 hour fcst: - 5: 8:3403216:d=2021032312:O3MR:15 mb:25 hour fcst: - 5: 9:3849313:d=2021032312:HGT:20 mb:25 hour fcst: - 5: 10:4304139:d=2021032312:TMP:20 mb:25 hour fcst: - 5: 11:4655508:d=2021032312:RH:20 mb:25 hour fcst: - 5: 12:4818536:d=2021032312:SPFH:20 mb:25 hour fcst: - 5: 13:5152769:d=2021032312:VVEL:20 mb:25 hour fcst: - 5: 14:5912653:d=2021032312:DZDT:20 mb:25 hour fcst: - 5: 15:6819763:d=2021032312:UGRD:20 mb:25 hour fcst: - 5: 16:7252760:d=2021032312:VGRD:20 mb:25 hour fcst: - 5: 17:7680867:d=2021032312:ABSV:20 mb:25 hour fcst: - 5: 18:8071929:d=2021032312:O3MR:20 mb:25 hour fcst: - 5: 19:8735977:d=2021032312:HGT:30 mb:25 hour fcst: - 5: 20:9186707:d=2021032312:TMP:30 mb:25 hour fcst: - 5: 21:9539374:d=2021032312:RH:30 mb:25 hour fcst: - 5: 22:9753709:d=2021032312:SPFH:30 mb:25 hour fcst: - 5: 23:10121616:d=2021032312:VVEL:30 mb:25 hour fcst: - 5: 24:10927051:d=2021032312:DZDT:30 mb:25 hour fcst: - 5: 25:11808221:d=2021032312:UGRD:30 mb:25 hour fcst: - 5: 26:12243517:d=2021032312:VGRD:30 mb:25 hour fcst: - 5: 27:12675235:d=2021032312:ABSV:30 mb:25 hour fcst: - 5: 28:13071201:d=2021032312:O3MR:30 mb:25 hour fcst: - 5: 29:13803638:d=2021032312:HGT:40 mb:25 hour fcst: - 5: 30:14242306:d=2021032312:TMP:40 mb:25 hour fcst: - 5: 31:14605199:d=2021032312:RH:40 mb:25 hour fcst: - 5: + product_functions.sh[10]rc=0 - 5: + product_functions.sh[11](( rc == 0 )) - 5: + product_functions.sh[11]mv pgb2file_f025_6_0p25.new pgb2file_f025_6_0p25 - 5: + product_functions.sh[12]return 0 - 5: + interp_atmos_master.sh[56]export err=0 - 5: + interp_atmos_master.sh[56]err=0 - 5: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 5: ++ interp_atmos_master.sh[62]wc -l - 5: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_6_0p25 -match 'LAND|ICEC' - 5: + interp_atmos_master.sh[62]var_count=0 - 5: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 5: + interp_atmos_master.sh[73]exit 0 -11: + bash[8]'[' -z '' ']' -11: + bash[9]case "$-" in -11: + bash[12]__lmod_vx=x -11: + bash[16]'[' -n x ']' -11: + bash[16]set +x -11: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -11: Shell debugging restarted -11: + bash[224]unset __lmod_vx -11: + interp_atmos_master.sh[7]input_file=tmpfile_f025_12 -11: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f025_12 -11: + interp_atmos_master.sh[9]grid_string=0p25 -11: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -11: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -11: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -11: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -11: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -11: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -11: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -11: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -11: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -11: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -11: + interp_atmos_master.sh[31]IFS=: -11: + interp_atmos_master.sh[31]read -ra grids -11: + interp_atmos_master.sh[33]output_grids= -11: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -11: + interp_atmos_master.sh[35]gridopt=grid0p25 -11: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_12_0p25' -11: + interp_atmos_master.sh[40]wgrib2 tmpfile_f025_12 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f025_12_0p25 -11: 1:0:d=2021032312:CLMR:450 mb:25 hour fcst: -11: 2:948:d=2021032312:ICMR:450 mb:25 hour fcst: -11: 3:17768:d=2021032312:RWMR:450 mb:25 hour fcst: -11: 4:18322:d=2021032312:SNMR:450 mb:25 hour fcst: -11: 5:31960:d=2021032312:GRLE:450 mb:25 hour fcst: -11: 6:33045:d=2021032312:O3MR:450 mb:25 hour fcst: -11: 7:57094:d=2021032312:HGT:500 mb:25 hour fcst: -11: 8:80814:d=2021032312:TMP:500 mb:25 hour fcst: -11: 9:99486:d=2021032312:RH:500 mb:25 hour fcst: -11: 10:119028:d=2021032312:TCDC:500 mb:25 hour fcst: -11: 11:131370:d=2021032312:SPFH:500 mb:25 hour fcst: -11: 12:159673:d=2021032312:VVEL:500 mb:25 hour fcst: -11: 13:184926:d=2021032312:DZDT:500 mb:25 hour fcst: -11: 14:212649:d=2021032312:UGRD:500 mb:25 hour fcst: -11: 15:234118:d=2021032312:VGRD:500 mb:25 hour fcst: -11: 16:256154:d=2021032312:ABSV:500 mb:25 hour fcst: -11: 17:277311:d=2021032312:CLMR:500 mb:25 hour fcst: -11: 18:278954:d=2021032312:ICMR:500 mb:25 hour fcst: -11: 19:294883:d=2021032312:RWMR:500 mb:25 hour fcst: -11: 20:295801:d=2021032312:SNMR:500 mb:25 hour fcst: -11: 21:310092:d=2021032312:GRLE:500 mb:25 hour fcst: -11: 22:311877:d=2021032312:O3MR:500 mb:25 hour fcst: -11: 23:335280:d=2021032312:HGT:550 mb:25 hour fcst: -11: 24:358852:d=2021032312:TMP:550 mb:25 hour fcst: -11: 25:377641:d=2021032312:RH:550 mb:25 hour fcst: -11: 26:397436:d=2021032312:TCDC:550 mb:25 hour fcst: -11: 27:409511:d=2021032312:SPFH:550 mb:25 hour fcst: -11: 28:439158:d=2021032312:VVEL:550 mb:25 hour fcst: -11: 29:464480:d=2021032312:DZDT:550 mb:25 hour fcst: -11: 30:491971:d=2021032312:UGRD:550 mb:25 hour fcst: -11: 31:512971:d=2021032312:VGRD:550 mb:25 hour fcst: -11: + interp_atmos_master.sh[47]export err=0 -11: + interp_atmos_master.sh[47]err=0 -11: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -11: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -11: + interp_atmos_master.sh[55]trim_rh pgb2file_f025_12_0p25 -11: + product_functions.sh[5]local filename=pgb2file_f025_12_0p25 -11: + product_functions.sh[6]wgrib2 pgb2file_f025_12_0p25 -not_if :RH: -grib pgb2file_f025_12_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f025_12_0p25.new -11: 1:0:d=2021032312:CLMR:450 mb:25 hour fcst: -11: 2:18184:d=2021032312:ICMR:450 mb:25 hour fcst: -11: 3:388942:d=2021032312:RWMR:450 mb:25 hour fcst: -11: 4:395563:d=2021032312:SNMR:450 mb:25 hour fcst: -11: 5:681762:d=2021032312:GRLE:450 mb:25 hour fcst: -11: 6:694856:d=2021032312:O3MR:450 mb:25 hour fcst: -11: 7:1310743:d=2021032312:HGT:500 mb:25 hour fcst: -11: 8:1923326:d=2021032312:TMP:500 mb:25 hour fcst: -11: 9:2352623:d=2021032312:RH:500 mb:25 hour fcst: -11: 10:2799351:d=2021032312:TCDC:500 mb:25 hour fcst: -11: 11:3033769:d=2021032312:SPFH:500 mb:25 hour fcst: -11: 12:3791253:d=2021032312:VVEL:500 mb:25 hour fcst: -11: 13:4460753:d=2021032312:DZDT:500 mb:25 hour fcst: -11: 14:5240001:d=2021032312:UGRD:500 mb:25 hour fcst: -11: 15:5762261:d=2021032312:VGRD:500 mb:25 hour fcst: -11: 16:6312609:d=2021032312:ABSV:500 mb:25 hour fcst: -11: 17:6816836:d=2021032312:CLMR:500 mb:25 hour fcst: -11: 18:6850429:d=2021032312:ICMR:500 mb:25 hour fcst: -11: 19:7204691:d=2021032312:RWMR:500 mb:25 hour fcst: -11: 20:7219012:d=2021032312:SNMR:500 mb:25 hour fcst: -11: 21:7526190:d=2021032312:GRLE:500 mb:25 hour fcst: -11: 22:7552879:d=2021032312:O3MR:500 mb:25 hour fcst: -11: 23:8152220:d=2021032312:HGT:550 mb:25 hour fcst: -11: 24:8751312:d=2021032312:TMP:550 mb:25 hour fcst: -11: 25:9178054:d=2021032312:RH:550 mb:25 hour fcst: -11: 26:9629807:d=2021032312:TCDC:550 mb:25 hour fcst: -11: 27:9860041:d=2021032312:SPFH:550 mb:25 hour fcst: -11: 28:10643018:d=2021032312:VVEL:550 mb:25 hour fcst: -11: 29:11313307:d=2021032312:DZDT:550 mb:25 hour fcst: -11: 30:12085190:d=2021032312:UGRD:550 mb:25 hour fcst: -11: 31:12598157:d=2021032312:VGRD:550 mb:25 hour fcst: -11: + product_functions.sh[10]rc=0 -11: + product_functions.sh[11](( rc == 0 )) -11: + product_functions.sh[11]mv pgb2file_f025_12_0p25.new pgb2file_f025_12_0p25 -11: + product_functions.sh[12]return 0 -11: + interp_atmos_master.sh[56]export err=0 -11: + interp_atmos_master.sh[56]err=0 -11: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -11: ++ interp_atmos_master.sh[62]wc -l -11: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f025_12_0p25 -match 'LAND|ICEC' -11: + interp_atmos_master.sh[62]var_count=0 -11: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -11: + interp_atmos_master.sh[73]exit 0 -+ run_mpmd.sh[113]exit 0 -+ run_mpmd.sh[1]postamble run_mpmd.sh 1753755993 0 -+ preamble.sh[62]set +x -End run_mpmd.sh at 02:26:55 with error code 0 (time elapsed: 00:00:22) -+ exglobal_atmos_products.sh[142]true -+ exglobal_atmos_products.sh[143]export err=0 -+ exglobal_atmos_products.sh[143]err=0 -+ exglobal_atmos_products.sh[144][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[149]mv mpmd.out mpmd_1.out -+ exglobal_atmos_products.sh[153]echo 'Concatenating processor-specific grib2 files into a single product file' -Concatenating processor-specific grib2 files into a single product file -+ exglobal_atmos_products.sh[154](( iproc = 1 )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_1_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_1_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_1 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_2_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_2_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_2 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_3_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_3_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_3 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_4_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_4_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_4 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_5_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_5_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_5 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_6_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_6_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_6 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_7_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_7_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_7 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_8_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_8_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_8 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_9_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_9_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_9 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_10_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_10_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_10 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_11_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_11_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_11 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_12_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_12_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_12 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_13_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_13_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_13 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_14_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_14_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_14 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_15_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_15_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_15 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_16_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_16_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_16 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_17_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_17_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_17 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_18_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_18_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_18 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_19_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_19_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_19 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_20_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_20_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_20 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_21_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_21_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_21 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_22_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_22_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_22 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_23_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_23_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_23 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f025_24_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f025_24_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f025_24 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[164]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[165]prod_dir=COMOUT_ATMOS_GRIB_0p25 -+ exglobal_atmos_products.sh[166]cpfs pgb2file_f025_0p25 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f025 -+ cpfs[3]'[' 2 -ne 2 ']' -+ cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f025 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f025 = ./ ']' -+ cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f025 ']' -+ cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f025 -+ cpfs[16]cp pgb2file_f025_0p25 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f025.cptmp -+ cpfs[18]'[' 0 -ne 0 ']' -+ cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f025.cptmp -+ cpfs[23]'[' 0 -ne 0 ']' -+ cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f025.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f025 -+ cpfs[28]'[' 0 -ne 0 ']' -+ exglobal_atmos_products.sh[167]wgrib2 -s pgb2file_f025_0p25 -+ exglobal_atmos_products.sh[170]echo 'Finished processing nset = 1' -Finished processing nset = 1 -+ exglobal_atmos_products.sh[154](( nset++ )) -+ exglobal_atmos_products.sh[154](( nset <= downset )) -+ exglobal_atmos_products.sh[74]echo 'Begin processing nset = 2' -Begin processing nset = 2 -+ exglobal_atmos_products.sh[77]nproc=24 -+ exglobal_atmos_products.sh[80][[ 2 == 1 ]] -+ exglobal_atmos_products.sh[82][[ 2 == 2 ]] -+ exglobal_atmos_products.sh[83]grp=b -+ exglobal_atmos_products.sh[87]tmpfile=tmpfileb_f025 -++ exglobal_atmos_products.sh[90]wgrib2 tmpfileb_f025 -++ exglobal_atmos_products.sh[90]wc -l -+ exglobal_atmos_products.sh[90]ncount=349 -+ exglobal_atmos_products.sh[91][[ 24 -gt 349 ]] -+ exglobal_atmos_products.sh[95]inv=14 -+ exglobal_atmos_products.sh[96]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601/poescript -+ exglobal_atmos_products.sh[98]last=0 -+ exglobal_atmos_products.sh[134](( iproc = 1 )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=1 -+ exglobal_atmos_products.sh[101]last=14 -+ exglobal_atmos_products.sh[102][[ 14 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 14 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 1 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 1:14 -grib tmpfileb_f025_1 -1:0:d=2021032312:HGT:1 mb:25 hour fcst: -2:17997:d=2021032312:TMP:1 mb:25 hour fcst: -3:33857:d=2021032312:RH:1 mb:25 hour fcst: -4:43627:d=2021032312:UGRD:1 mb:25 hour fcst: -5:62225:d=2021032312:VGRD:1 mb:25 hour fcst: -6:79897:d=2021032312:ABSV:1 mb:25 hour fcst: -7:96195:d=2021032312:O3MR:1 mb:25 hour fcst: -8:117327:d=2021032312:HGT:2 mb:25 hour fcst: -9:137530:d=2021032312:TMP:2 mb:25 hour fcst: -10:154182:d=2021032312:RH:2 mb:25 hour fcst: -11:162864:d=2021032312:UGRD:2 mb:25 hour fcst: -12:174427:d=2021032312:VGRD:2 mb:25 hour fcst: -13:192855:d=2021032312:ABSV:2 mb:25 hour fcst: -14:209874:d=2021032312:O3MR:2 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_1 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_1 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_1 pgb2bfile_f025_1 0p25' -+ exglobal_atmos_products.sh[133][[ 14 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=15 -+ exglobal_atmos_products.sh[101]last=28 -+ exglobal_atmos_products.sh[102][[ 28 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 28 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 2 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 15:28 -grib tmpfileb_f025_2 -15:232649:d=2021032312:HGT:3 mb:25 hour fcst: -16:250449:d=2021032312:TMP:3 mb:25 hour fcst: -17:266937:d=2021032312:RH:3 mb:25 hour fcst: -18:272927:d=2021032312:UGRD:3 mb:25 hour fcst: -19:284293:d=2021032312:VGRD:3 mb:25 hour fcst: -20:302285:d=2021032312:ABSV:3 mb:25 hour fcst: -21:319203:d=2021032312:O3MR:3 mb:25 hour fcst: -22:342505:d=2021032312:HGT:5 mb:25 hour fcst: -23:360162:d=2021032312:TMP:5 mb:25 hour fcst: -24:376853:d=2021032312:RH:5 mb:25 hour fcst: -25:386043:d=2021032312:UGRD:5 mb:25 hour fcst: -26:397640:d=2021032312:VGRD:5 mb:25 hour fcst: -27:416248:d=2021032312:ABSV:5 mb:25 hour fcst: -28:433933:d=2021032312:O3MR:5 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_2 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_2 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_2 pgb2bfile_f025_2 0p25' -+ exglobal_atmos_products.sh[133][[ 28 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=29 -+ exglobal_atmos_products.sh[101]last=42 -+ exglobal_atmos_products.sh[102][[ 42 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 42 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 3 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 29:42 -grib tmpfileb_f025_3 -29:456704:d=2021032312:HGT:7 mb:25 hour fcst: -30:474369:d=2021032312:TMP:7 mb:25 hour fcst: -31:490453:d=2021032312:RH:7 mb:25 hour fcst: -32:496503:d=2021032312:UGRD:7 mb:25 hour fcst: -33:507920:d=2021032312:VGRD:7 mb:25 hour fcst: -34:526673:d=2021032312:ABSV:7 mb:25 hour fcst: -35:544408:d=2021032312:O3MR:7 mb:25 hour fcst: -36:568075:d=2021032312:TCDC:70 mb:25 hour fcst: -37:568254:d=2021032312:CLMR:70 mb:25 hour fcst: -38:568433:d=2021032312:ICMR:70 mb:25 hour fcst: -39:568612:d=2021032312:RWMR:70 mb:25 hour fcst: -40:568791:d=2021032312:SNMR:70 mb:25 hour fcst: -41:568970:d=2021032312:GRLE:70 mb:25 hour fcst: -42:569149:d=2021032312:HGT:125 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_3 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_3 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_3 pgb2bfile_f025_3 0p25' -+ exglobal_atmos_products.sh[133][[ 42 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=43 -+ exglobal_atmos_products.sh[101]last=56 -+ exglobal_atmos_products.sh[102][[ 56 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 56 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 4 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 43:56 -grib tmpfileb_f025_4 -43:589224:d=2021032312:TMP:125 mb:25 hour fcst: -44:606233:d=2021032312:RH:125 mb:25 hour fcst: -45:616743:d=2021032312:TCDC:125 mb:25 hour fcst: -46:619180:d=2021032312:VVEL:125 mb:25 hour fcst: -47:645375:d=2021032312:DZDT:125 mb:25 hour fcst: -48:670467:d=2021032312:UGRD:125 mb:25 hour fcst: -49:690084:d=2021032312:VGRD:125 mb:25 hour fcst: -50:709713:d=2021032312:ABSV:125 mb:25 hour fcst: -51:728220:d=2021032312:CLMR:125 mb:25 hour fcst: -52:728399:d=2021032312:ICMR:125 mb:25 hour fcst: -53:732892:d=2021032312:RWMR:125 mb:25 hour fcst: -54:733071:d=2021032312:SNMR:125 mb:25 hour fcst: -55:736888:d=2021032312:GRLE:125 mb:25 hour fcst: -56:737299:d=2021032312:HGT:175 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_4 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_4 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_4 pgb2bfile_f025_4 0p25' -+ exglobal_atmos_products.sh[133][[ 56 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=57 -+ exglobal_atmos_products.sh[101]last=70 -+ exglobal_atmos_products.sh[102][[ 70 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 70 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 5 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 57:70 -grib tmpfileb_f025_5 -57:757929:d=2021032312:TMP:175 mb:25 hour fcst: -58:775411:d=2021032312:RH:175 mb:25 hour fcst: -59:788543:d=2021032312:TCDC:175 mb:25 hour fcst: -60:793251:d=2021032312:VVEL:175 mb:25 hour fcst: -61:821453:d=2021032312:DZDT:175 mb:25 hour fcst: -62:847513:d=2021032312:UGRD:175 mb:25 hour fcst: -63:860512:d=2021032312:VGRD:175 mb:25 hour fcst: -64:881254:d=2021032312:ABSV:175 mb:25 hour fcst: -65:900789:d=2021032312:CLMR:175 mb:25 hour fcst: -66:900968:d=2021032312:ICMR:175 mb:25 hour fcst: -67:907598:d=2021032312:RWMR:175 mb:25 hour fcst: -68:907777:d=2021032312:SNMR:175 mb:25 hour fcst: -69:911447:d=2021032312:GRLE:175 mb:25 hour fcst: -70:911926:d=2021032312:HGT:225 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_5 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_5 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_5 pgb2bfile_f025_5 0p25' -+ exglobal_atmos_products.sh[133][[ 70 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=71 -+ exglobal_atmos_products.sh[101]last=84 -+ exglobal_atmos_products.sh[102][[ 84 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 84 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 6 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 71:84 -grib tmpfileb_f025_6 -71:933359:d=2021032312:TMP:225 mb:25 hour fcst: -72:951958:d=2021032312:RH:225 mb:25 hour fcst: -73:968605:d=2021032312:TCDC:225 mb:25 hour fcst: -74:977418:d=2021032312:VVEL:225 mb:25 hour fcst: -75:1005029:d=2021032312:DZDT:225 mb:25 hour fcst: -76:1031779:d=2021032312:UGRD:225 mb:25 hour fcst: -77:1045741:d=2021032312:VGRD:225 mb:25 hour fcst: -78:1060114:d=2021032312:ABSV:225 mb:25 hour fcst: -79:1081044:d=2021032312:CLMR:225 mb:25 hour fcst: -80:1081223:d=2021032312:ICMR:225 mb:25 hour fcst: -81:1092731:d=2021032312:RWMR:225 mb:25 hour fcst: -82:1092910:d=2021032312:SNMR:225 mb:25 hour fcst: -83:1099724:d=2021032312:GRLE:225 mb:25 hour fcst: -84:1099997:d=2021032312:HGT:275 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_6 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_6 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_6 pgb2bfile_f025_6 0p25' -+ exglobal_atmos_products.sh[133][[ 84 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=85 -+ exglobal_atmos_products.sh[101]last=98 -+ exglobal_atmos_products.sh[102][[ 98 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 98 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 7 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 85:98 -grib tmpfileb_f025_7 -85:1122183:d=2021032312:TMP:275 mb:25 hour fcst: -86:1141020:d=2021032312:RH:275 mb:25 hour fcst: -87:1159459:d=2021032312:TCDC:275 mb:25 hour fcst: -88:1171209:d=2021032312:VVEL:275 mb:25 hour fcst: -89:1194521:d=2021032312:DZDT:275 mb:25 hour fcst: -90:1222154:d=2021032312:UGRD:275 mb:25 hour fcst: -91:1236469:d=2021032312:VGRD:275 mb:25 hour fcst: -92:1251641:d=2021032312:ABSV:275 mb:25 hour fcst: -93:1273695:d=2021032312:CLMR:275 mb:25 hour fcst: -94:1273874:d=2021032312:ICMR:275 mb:25 hour fcst: -95:1288394:d=2021032312:RWMR:275 mb:25 hour fcst: -96:1288573:d=2021032312:SNMR:275 mb:25 hour fcst: -97:1298522:d=2021032312:GRLE:275 mb:25 hour fcst: -98:1298832:d=2021032312:HGT:325 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_7 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_7 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_7 pgb2bfile_f025_7 0p25' -+ exglobal_atmos_products.sh[133][[ 98 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=99 -+ exglobal_atmos_products.sh[101]last=112 -+ exglobal_atmos_products.sh[102][[ 112 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 112 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 8 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 99:112 -grib tmpfileb_f025_8 -99:1321016:d=2021032312:TMP:325 mb:25 hour fcst: -100:1339370:d=2021032312:RH:325 mb:25 hour fcst: -101:1358331:d=2021032312:TCDC:325 mb:25 hour fcst: -102:1371488:d=2021032312:VVEL:325 mb:25 hour fcst: -103:1395772:d=2021032312:DZDT:325 mb:25 hour fcst: -104:1424115:d=2021032312:UGRD:325 mb:25 hour fcst: -105:1438792:d=2021032312:VGRD:325 mb:25 hour fcst: -106:1454123:d=2021032312:ABSV:325 mb:25 hour fcst: -107:1476437:d=2021032312:CLMR:325 mb:25 hour fcst: -108:1476616:d=2021032312:ICMR:325 mb:25 hour fcst: -109:1493380:d=2021032312:RWMR:325 mb:25 hour fcst: -110:1493559:d=2021032312:SNMR:325 mb:25 hour fcst: -111:1505529:d=2021032312:GRLE:325 mb:25 hour fcst: -112:1505963:d=2021032312:HGT:375 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_8 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_8 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_8 pgb2bfile_f025_8 0p25' -+ exglobal_atmos_products.sh[133][[ 112 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=113 -+ exglobal_atmos_products.sh[101]last=126 -+ exglobal_atmos_products.sh[102][[ 126 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 126 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 9 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 113:126 -grib tmpfileb_f025_9 -113:1527811:d=2021032312:TMP:375 mb:25 hour fcst: -114:1546143:d=2021032312:RH:375 mb:25 hour fcst: -115:1565262:d=2021032312:TCDC:375 mb:25 hour fcst: -116:1578425:d=2021032312:VVEL:375 mb:25 hour fcst: -117:1603256:d=2021032312:DZDT:375 mb:25 hour fcst: -118:1631327:d=2021032312:UGRD:375 mb:25 hour fcst: -119:1645774:d=2021032312:VGRD:375 mb:25 hour fcst: -120:1660882:d=2021032312:ABSV:375 mb:25 hour fcst: -121:1683048:d=2021032312:CLMR:375 mb:25 hour fcst: -122:1683227:d=2021032312:ICMR:375 mb:25 hour fcst: -123:1700327:d=2021032312:RWMR:375 mb:25 hour fcst: -124:1700506:d=2021032312:SNMR:375 mb:25 hour fcst: -125:1713674:d=2021032312:GRLE:375 mb:25 hour fcst: -126:1714307:d=2021032312:HGT:425 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_9 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_9 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_9 pgb2bfile_f025_9 0p25' -+ exglobal_atmos_products.sh[133][[ 126 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=127 -+ exglobal_atmos_products.sh[101]last=140 -+ exglobal_atmos_products.sh[102][[ 140 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 140 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 10 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 127:140 -grib tmpfileb_f025_10 -127:1738316:d=2021032312:TMP:425 mb:25 hour fcst: -128:1756837:d=2021032312:RH:425 mb:25 hour fcst: -129:1776139:d=2021032312:TCDC:425 mb:25 hour fcst: -130:1789012:d=2021032312:VVEL:425 mb:25 hour fcst: -131:1813907:d=2021032312:DZDT:425 mb:25 hour fcst: -132:1841967:d=2021032312:UGRD:425 mb:25 hour fcst: -133:1863804:d=2021032312:VGRD:425 mb:25 hour fcst: -134:1886261:d=2021032312:ABSV:425 mb:25 hour fcst: -135:1907875:d=2021032312:CLMR:425 mb:25 hour fcst: -136:1908513:d=2021032312:ICMR:425 mb:25 hour fcst: -137:1925428:d=2021032312:RWMR:425 mb:25 hour fcst: -138:1925827:d=2021032312:SNMR:425 mb:25 hour fcst: -139:1938931:d=2021032312:GRLE:425 mb:25 hour fcst: -140:1939972:d=2021032312:HGT:475 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_10 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_10 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_10 pgb2bfile_f025_10 0p25' -+ exglobal_atmos_products.sh[133][[ 140 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=141 -+ exglobal_atmos_products.sh[101]last=154 -+ exglobal_atmos_products.sh[102][[ 154 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 154 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 11 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 141:154 -grib tmpfileb_f025_11 -141:1963860:d=2021032312:TMP:475 mb:25 hour fcst: -142:1982505:d=2021032312:RH:475 mb:25 hour fcst: -143:2002011:d=2021032312:TCDC:475 mb:25 hour fcst: -144:2014423:d=2021032312:VVEL:475 mb:25 hour fcst: -145:2039440:d=2021032312:DZDT:475 mb:25 hour fcst: -146:2067267:d=2021032312:UGRD:475 mb:25 hour fcst: -147:2088833:d=2021032312:VGRD:475 mb:25 hour fcst: -148:2111032:d=2021032312:ABSV:475 mb:25 hour fcst: -149:2132297:d=2021032312:CLMR:475 mb:25 hour fcst: -150:2133556:d=2021032312:ICMR:475 mb:25 hour fcst: -151:2149842:d=2021032312:RWMR:475 mb:25 hour fcst: -152:2150556:d=2021032312:SNMR:475 mb:25 hour fcst: -153:2164581:d=2021032312:GRLE:475 mb:25 hour fcst: -154:2166094:d=2021032312:HGT:525 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_11 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_11 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_11 pgb2bfile_f025_11 0p25' -+ exglobal_atmos_products.sh[133][[ 154 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=155 -+ exglobal_atmos_products.sh[101]last=168 -+ exglobal_atmos_products.sh[102][[ 168 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 168 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 12 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 155:168 -grib tmpfileb_f025_12 -155:2189730:d=2021032312:TMP:525 mb:25 hour fcst: -156:2208551:d=2021032312:RH:525 mb:25 hour fcst: -157:2228228:d=2021032312:TCDC:525 mb:25 hour fcst: -158:2240446:d=2021032312:VVEL:525 mb:25 hour fcst: -159:2265537:d=2021032312:DZDT:525 mb:25 hour fcst: -160:2293134:d=2021032312:UGRD:525 mb:25 hour fcst: -161:2314275:d=2021032312:VGRD:525 mb:25 hour fcst: -162:2336171:d=2021032312:ABSV:525 mb:25 hour fcst: -163:2357183:d=2021032312:CLMR:525 mb:25 hour fcst: -164:2359269:d=2021032312:ICMR:525 mb:25 hour fcst: -165:2376296:d=2021032312:RWMR:525 mb:25 hour fcst: -166:2377398:d=2021032312:SNMR:525 mb:25 hour fcst: -167:2391837:d=2021032312:GRLE:525 mb:25 hour fcst: -168:2393873:d=2021032312:HGT:575 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_12 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_12 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_12 pgb2bfile_f025_12 0p25' -+ exglobal_atmos_products.sh[133][[ 168 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=169 -+ exglobal_atmos_products.sh[101]last=182 -+ exglobal_atmos_products.sh[102][[ 182 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 182 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 13 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 169:182 -grib tmpfileb_f025_13 -169:2417355:d=2021032312:TMP:575 mb:25 hour fcst: -170:2436233:d=2021032312:RH:575 mb:25 hour fcst: -171:2456151:d=2021032312:TCDC:575 mb:25 hour fcst: -172:2468385:d=2021032312:VVEL:575 mb:25 hour fcst: -173:2493609:d=2021032312:DZDT:575 mb:25 hour fcst: -174:2521028:d=2021032312:UGRD:575 mb:25 hour fcst: -175:2541928:d=2021032312:VGRD:575 mb:25 hour fcst: -176:2563605:d=2021032312:ABSV:575 mb:25 hour fcst: -177:2584461:d=2021032312:CLMR:575 mb:25 hour fcst: -178:2588642:d=2021032312:ICMR:575 mb:25 hour fcst: -179:2604695:d=2021032312:RWMR:575 mb:25 hour fcst: -180:2607077:d=2021032312:SNMR:575 mb:25 hour fcst: -181:2621662:d=2021032312:GRLE:575 mb:25 hour fcst: -182:2624107:d=2021032312:HGT:625 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_13 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_13 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_13 pgb2bfile_f025_13 0p25' -+ exglobal_atmos_products.sh[133][[ 182 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=183 -+ exglobal_atmos_products.sh[101]last=196 -+ exglobal_atmos_products.sh[102][[ 196 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 196 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 14 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 183:196 -grib tmpfileb_f025_14 -183:2648457:d=2021032312:TMP:625 mb:25 hour fcst: -184:2667591:d=2021032312:RH:625 mb:25 hour fcst: -185:2687714:d=2021032312:TCDC:625 mb:25 hour fcst: -186:2699533:d=2021032312:VVEL:625 mb:25 hour fcst: -187:2725021:d=2021032312:DZDT:625 mb:25 hour fcst: -188:2752442:d=2021032312:UGRD:625 mb:25 hour fcst: -189:2773312:d=2021032312:VGRD:625 mb:25 hour fcst: -190:2794946:d=2021032312:ABSV:625 mb:25 hour fcst: -191:2815737:d=2021032312:CLMR:625 mb:25 hour fcst: -192:2820669:d=2021032312:ICMR:625 mb:25 hour fcst: -193:2833761:d=2021032312:RWMR:625 mb:25 hour fcst: -194:2837992:d=2021032312:SNMR:625 mb:25 hour fcst: -195:2853325:d=2021032312:GRLE:625 mb:25 hour fcst: -196:2855786:d=2021032312:HGT:675 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_14 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_14 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_14 pgb2bfile_f025_14 0p25' -+ exglobal_atmos_products.sh[133][[ 196 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=197 -+ exglobal_atmos_products.sh[101]last=210 -+ exglobal_atmos_products.sh[102][[ 210 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 210 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 15 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 197:210 -grib tmpfileb_f025_15 -197:2879933:d=2021032312:TMP:675 mb:25 hour fcst: -198:2899378:d=2021032312:RH:675 mb:25 hour fcst: -199:2919621:d=2021032312:TCDC:675 mb:25 hour fcst: -200:2931284:d=2021032312:VVEL:675 mb:25 hour fcst: -201:2956976:d=2021032312:DZDT:675 mb:25 hour fcst: -202:2984367:d=2021032312:UGRD:675 mb:25 hour fcst: -203:3005419:d=2021032312:VGRD:675 mb:25 hour fcst: -204:3026926:d=2021032312:ABSV:675 mb:25 hour fcst: -205:3047970:d=2021032312:CLMR:675 mb:25 hour fcst: -206:3053508:d=2021032312:ICMR:675 mb:25 hour fcst: -207:3065370:d=2021032312:RWMR:675 mb:25 hour fcst: -208:3070418:d=2021032312:SNMR:675 mb:25 hour fcst: -209:3084995:d=2021032312:GRLE:675 mb:25 hour fcst: -210:3087140:d=2021032312:HGT:725 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_15 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_15 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_15 pgb2bfile_f025_15 0p25' -+ exglobal_atmos_products.sh[133][[ 210 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=211 -+ exglobal_atmos_products.sh[101]last=224 -+ exglobal_atmos_products.sh[102][[ 224 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 224 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 16 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 211:224 -grib tmpfileb_f025_16 -211:3111322:d=2021032312:TMP:725 mb:25 hour fcst: -212:3131053:d=2021032312:RH:725 mb:25 hour fcst: -213:3151361:d=2021032312:TCDC:725 mb:25 hour fcst: -214:3163237:d=2021032312:VVEL:725 mb:25 hour fcst: -215:3188983:d=2021032312:DZDT:725 mb:25 hour fcst: -216:3216322:d=2021032312:UGRD:725 mb:25 hour fcst: -217:3237167:d=2021032312:VGRD:725 mb:25 hour fcst: -218:3258821:d=2021032312:ABSV:725 mb:25 hour fcst: -219:3279922:d=2021032312:CLMR:725 mb:25 hour fcst: -220:3287772:d=2021032312:ICMR:725 mb:25 hour fcst: -221:3299358:d=2021032312:RWMR:725 mb:25 hour fcst: -222:3305058:d=2021032312:SNMR:725 mb:25 hour fcst: -223:3318867:d=2021032312:GRLE:725 mb:25 hour fcst: -224:3320787:d=2021032312:HGT:775 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_16 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_16 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_16 pgb2bfile_f025_16 0p25' -+ exglobal_atmos_products.sh[133][[ 224 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=225 -+ exglobal_atmos_products.sh[101]last=238 -+ exglobal_atmos_products.sh[102][[ 238 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 238 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 17 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 225:238 -grib tmpfileb_f025_17 -225:3345235:d=2021032312:TMP:775 mb:25 hour fcst: -226:3365369:d=2021032312:RH:775 mb:25 hour fcst: -227:3385838:d=2021032312:TCDC:775 mb:25 hour fcst: -228:3398597:d=2021032312:VVEL:775 mb:25 hour fcst: -229:3424340:d=2021032312:DZDT:775 mb:25 hour fcst: -230:3451331:d=2021032312:UGRD:775 mb:25 hour fcst: -231:3472186:d=2021032312:VGRD:775 mb:25 hour fcst: -232:3493910:d=2021032312:ABSV:775 mb:25 hour fcst: -233:3515099:d=2021032312:CLMR:775 mb:25 hour fcst: -234:3525461:d=2021032312:ICMR:775 mb:25 hour fcst: -235:3536521:d=2021032312:RWMR:775 mb:25 hour fcst: -236:3543236:d=2021032312:SNMR:775 mb:25 hour fcst: -237:3556368:d=2021032312:GRLE:775 mb:25 hour fcst: -238:3558647:d=2021032312:HGT:825 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_17 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_17 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_17 pgb2bfile_f025_17 0p25' -+ exglobal_atmos_products.sh[133][[ 238 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=239 -+ exglobal_atmos_products.sh[101]last=252 -+ exglobal_atmos_products.sh[102][[ 252 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 252 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 18 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 239:252 -grib tmpfileb_f025_18 -239:3583262:d=2021032312:TMP:825 mb:25 hour fcst: -240:3604073:d=2021032312:RH:825 mb:25 hour fcst: -241:3624614:d=2021032312:TCDC:825 mb:25 hour fcst: -242:3638967:d=2021032312:VVEL:825 mb:25 hour fcst: -243:3664577:d=2021032312:DZDT:825 mb:25 hour fcst: -244:3691440:d=2021032312:UGRD:825 mb:25 hour fcst: -245:3712620:d=2021032312:VGRD:825 mb:25 hour fcst: -246:3734468:d=2021032312:ABSV:825 mb:25 hour fcst: -247:3755715:d=2021032312:CLMR:825 mb:25 hour fcst: -248:3768720:d=2021032312:ICMR:825 mb:25 hour fcst: -249:3779134:d=2021032312:RWMR:825 mb:25 hour fcst: -250:3788535:d=2021032312:SNMR:825 mb:25 hour fcst: -251:3800897:d=2021032312:GRLE:825 mb:25 hour fcst: -252:3803364:d=2021032312:HGT:875 mb:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_18 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_18 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_18 pgb2bfile_f025_18 0p25' -+ exglobal_atmos_products.sh[133][[ 252 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=253 -+ exglobal_atmos_products.sh[101]last=266 -+ exglobal_atmos_products.sh[102][[ 266 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 266 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 19 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 253:266 -grib tmpfileb_f025_19 -253:3828232:d=2021032312:TMP:875 mb:25 hour fcst: -254:3849053:d=2021032312:RH:875 mb:25 hour fcst: -255:3869033:d=2021032312:TCDC:875 mb:25 hour fcst: -256:3883370:d=2021032312:VVEL:875 mb:25 hour fcst: -257:3908705:d=2021032312:DZDT:875 mb:25 hour fcst: -258:3935115:d=2021032312:UGRD:875 mb:25 hour fcst: -259:3956325:d=2021032312:VGRD:875 mb:25 hour fcst: -260:3978176:d=2021032312:ABSV:875 mb:25 hour fcst: -261:3999413:d=2021032312:CLMR:875 mb:25 hour fcst: -262:4011716:d=2021032312:ICMR:875 mb:25 hour fcst: -263:4021626:d=2021032312:RWMR:875 mb:25 hour fcst: -264:4033149:d=2021032312:SNMR:875 mb:25 hour fcst: -265:4044116:d=2021032312:GRLE:875 mb:25 hour fcst: -266:4047179:d=2021032312:SOILL:0-0.1 m below ground:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_19 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_19 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_19 pgb2bfile_f025_19 0p25' -+ exglobal_atmos_products.sh[133][[ 266 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=267 -+ exglobal_atmos_products.sh[101]last=280 -+ exglobal_atmos_products.sh[102][[ 280 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 280 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 20 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 267:280 -grib tmpfileb_f025_20 -267:4055985:d=2021032312:SOILL:0.1-0.4 m below ground:25 hour fcst: -268:4064879:d=2021032312:SOILL:0.4-1 m below ground:25 hour fcst: -269:4073798:d=2021032312:SOILL:1-2 m below ground:25 hour fcst: -270:4082876:d=2021032312:CNWAT:surface:25 hour fcst: -271:4088420:d=2021032312:ICETK:surface:25 hour fcst: -272:4091837:d=2021032312:DUVB:surface:24-25 hour ave fcst: -273:4108553:d=2021032312:CDUVB:surface:24-25 hour ave fcst: -274:4121719:d=2021032312:TMP:305 m above mean sea level:25 hour fcst: -275:4139920:d=2021032312:UGRD:305 m above mean sea level:25 hour fcst: -276:4159075:d=2021032312:VGRD:305 m above mean sea level:25 hour fcst: -277:4178712:d=2021032312:TMP:457 m above mean sea level:25 hour fcst: -278:4198237:d=2021032312:UGRD:457 m above mean sea level:25 hour fcst: -279:4218444:d=2021032312:VGRD:457 m above mean sea level:25 hour fcst: -280:4238995:d=2021032312:TMP:610 m above mean sea level:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_20 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_20 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_20 pgb2bfile_f025_20 0p25' -+ exglobal_atmos_products.sh[133][[ 280 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=281 -+ exglobal_atmos_products.sh[101]last=294 -+ exglobal_atmos_products.sh[102][[ 294 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 294 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 21 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 281:294 -grib tmpfileb_f025_21 -281:4259277:d=2021032312:UGRD:610 m above mean sea level:25 hour fcst: -282:4280113:d=2021032312:VGRD:610 m above mean sea level:25 hour fcst: -283:4301263:d=2021032312:TMP:914 m above mean sea level:25 hour fcst: -284:4322113:d=2021032312:UGRD:914 m above mean sea level:25 hour fcst: -285:4343506:d=2021032312:VGRD:914 m above mean sea level:25 hour fcst: -286:4365341:d=2021032312:TMP:4572 m above mean sea level:25 hour fcst: -287:4386491:d=2021032312:UGRD:4572 m above mean sea level:25 hour fcst: -288:4409712:d=2021032312:VGRD:4572 m above mean sea level:25 hour fcst: -289:4433669:d=2021032312:TMP:60-30 mb above ground:25 hour fcst: -290:4454475:d=2021032312:RH:60-30 mb above ground:25 hour fcst: -291:4473147:d=2021032312:SPFH:60-30 mb above ground:25 hour fcst: -292:4501888:d=2021032312:UGRD:60-30 mb above ground:25 hour fcst: -293:4523471:d=2021032312:VGRD:60-30 mb above ground:25 hour fcst: -294:4545605:d=2021032312:TMP:90-60 mb above ground:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_21 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_21 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_21 pgb2bfile_f025_21 0p25' -+ exglobal_atmos_products.sh[133][[ 294 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=295 -+ exglobal_atmos_products.sh[101]last=308 -+ exglobal_atmos_products.sh[102][[ 308 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 308 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 22 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 295:308 -grib tmpfileb_f025_22 -295:4566494:d=2021032312:RH:90-60 mb above ground:25 hour fcst: -296:4585573:d=2021032312:SPFH:90-60 mb above ground:25 hour fcst: -297:4614529:d=2021032312:UGRD:90-60 mb above ground:25 hour fcst: -298:4635932:d=2021032312:VGRD:90-60 mb above ground:25 hour fcst: -299:4657893:d=2021032312:TMP:120-90 mb above ground:25 hour fcst: -300:4678835:d=2021032312:RH:120-90 mb above ground:25 hour fcst: -301:4698183:d=2021032312:SPFH:120-90 mb above ground:25 hour fcst: -302:4727117:d=2021032312:UGRD:120-90 mb above ground:25 hour fcst: -303:4748295:d=2021032312:VGRD:120-90 mb above ground:25 hour fcst: -304:4770085:d=2021032312:TMP:150-120 mb above ground:25 hour fcst: -305:4790991:d=2021032312:RH:150-120 mb above ground:25 hour fcst: -306:4810695:d=2021032312:SPFH:150-120 mb above ground:25 hour fcst: -307:4839811:d=2021032312:UGRD:150-120 mb above ground:25 hour fcst: -308:4860833:d=2021032312:VGRD:150-120 mb above ground:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_22 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_22 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_22 pgb2bfile_f025_22 0p25' -+ exglobal_atmos_products.sh[133][[ 308 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=309 -+ exglobal_atmos_products.sh[101]last=322 -+ exglobal_atmos_products.sh[102][[ 322 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 322 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 23 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 309:322 -grib tmpfileb_f025_23 -309:4882518:d=2021032312:TMP:180-150 mb above ground:25 hour fcst: -310:4903682:d=2021032312:RH:180-150 mb above ground:25 hour fcst: -311:4923670:d=2021032312:SPFH:180-150 mb above ground:25 hour fcst: -312:4952929:d=2021032312:UGRD:180-150 mb above ground:25 hour fcst: -313:4974011:d=2021032312:VGRD:180-150 mb above ground:25 hour fcst: -314:4995656:d=2021032312:UGRD:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -315:5006254:d=2021032312:VGRD:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -316:5016819:d=2021032312:TMP:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -317:5031427:d=2021032312:HGT:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -318:5048159:d=2021032312:PRES:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -319:5065000:d=2021032312:VWSH:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -320:5075964:d=2021032312:UGRD:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -321:5086932:d=2021032312:VGRD:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -322:5101443:d=2021032312:TMP:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_23 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_23 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_23 pgb2bfile_f025_23 0p25' -+ exglobal_atmos_products.sh[133][[ 322 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=323 -+ exglobal_atmos_products.sh[101]last=336 -+ exglobal_atmos_products.sh[102][[ 336 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 336 tmpfileb_f025 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 24 -eq 24 ]] -+ exglobal_atmos_products.sh[118]last=349 -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f025 -for 323:349 -grib tmpfileb_f025_24 -323:5116806:d=2021032312:HGT:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -324:5134451:d=2021032312:PRES:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -325:5152144:d=2021032312:VWSH:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -326:5163556:d=2021032312:UGRD:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -327:5174881:d=2021032312:VGRD:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -328:5186193:d=2021032312:TMP:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -329:5202130:d=2021032312:HGT:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -330:5220698:d=2021032312:PRES:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -331:5239376:d=2021032312:VWSH:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -332:5251590:d=2021032312:UGRD:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -333:5266958:d=2021032312:VGRD:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -334:5278283:d=2021032312:TMP:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -335:5294077:d=2021032312:HGT:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -336:5312474:d=2021032312:PRES:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -337:5330935:d=2021032312:VWSH:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -338:5343266:d=2021032312:UGRD:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -339:5354312:d=2021032312:VGRD:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -340:5365298:d=2021032312:TMP:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -341:5380280:d=2021032312:HGT:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -342:5398209:d=2021032312:PRES:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -343:5416021:d=2021032312:VWSH:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -344:5428143:d=2021032312:UGRD:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -345:5439190:d=2021032312:VGRD:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -346:5450144:d=2021032312:TMP:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -347:5465206:d=2021032312:HGT:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -348:5483128:d=2021032312:PRES:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -349:5500879:d=2021032312:VWSH:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f025_24 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f025_24 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_24 pgb2bfile_f025_24 0p25' -+ exglobal_atmos_products.sh[133][[ 349 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( pproc = iproc+1 )) -+ exglobal_atmos_products.sh[134](( pproc < nproc )) -+ exglobal_atmos_products.sh[137]break -+ exglobal_atmos_products.sh[142]/work2/noaa/global/mterry/global-workflow_forked/ush/run_mpmd.sh /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601/poescript -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ run_mpmd.sh[31]source /work2/noaa/global/mterry/global-workflow_forked/ush/preamble.sh -++ preamble.sh[20]set +x -Begin run_mpmd.sh at Tue Jul 29 02:27:10 UTC 2025 -+ run_mpmd.sh[33]cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601/poescript -+ run_mpmd.sh[36][[ YES != \Y\E\S ]] -+ run_mpmd.sh[46]export OMP_NUM_THREADS=1 -+ run_mpmd.sh[46]OMP_NUM_THREADS=1 -++ run_mpmd.sh[49]wc -l -+ run_mpmd.sh[49]nprocs=24 -+ run_mpmd.sh[52]mpmd_cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601/mpmd_cmdfile -+ run_mpmd.sh[53][[ -s /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601/mpmd_cmdfile ]] -+ run_mpmd.sh[55]cat - INFO: Executing MPMD job, STDOUT redirected for each process separately - INFO: On failure, logs for each job will be available in /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601/mpmd.proc_num.out - INFO: The proc_num corresponds to the line in '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601/mpmd_cmdfile' -+ run_mpmd.sh[61][[ srun -l --export=ALL --hint=nomultithread =~ ^srun.* ]] -+ run_mpmd.sh[65]nm=0 -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '0 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_1 pgb2bfile_f025_1 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '1 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_2 pgb2bfile_f025_2 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '2 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_3 pgb2bfile_f025_3 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '3 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_4 pgb2bfile_f025_4 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '4 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_5 pgb2bfile_f025_5 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '5 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_6 pgb2bfile_f025_6 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '6 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_7 pgb2bfile_f025_7 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '7 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_8 pgb2bfile_f025_8 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '8 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_9 pgb2bfile_f025_9 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '9 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_10 pgb2bfile_f025_10 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '10 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_11 pgb2bfile_f025_11 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '11 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_12 pgb2bfile_f025_12 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '12 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_13 pgb2bfile_f025_13 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '13 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_14 pgb2bfile_f025_14 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '14 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_15 pgb2bfile_f025_15 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '15 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_16 pgb2bfile_f025_16 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '16 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_17 pgb2bfile_f025_17 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '17 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_18 pgb2bfile_f025_18 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '18 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_19 pgb2bfile_f025_19 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '19 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_20 pgb2bfile_f025_20 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '20 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_21 pgb2bfile_f025_21 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '21 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_22 pgb2bfile_f025_22 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '22 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_23 pgb2bfile_f025_23 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '23 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f025_24 pgb2bfile_f025_24 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[72]set +e -+ run_mpmd.sh[74]srun -l --export=ALL --hint=nomultithread --multi-prog --output=mpmd.%j.%t.out -n 24 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601/mpmd_cmdfile -+ run_mpmd.sh[75]err=0 -+ run_mpmd.sh[76]set_strict -+ preamble.sh[35][[ YES == \Y\E\S ]] -+ preamble.sh[37]set -eu -+ run_mpmd.sh[103][[ 0 -eq 0 ]] -+ run_mpmd.sh[104]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601/mpmd_cmdfile -++ run_mpmd.sh[105]find . -name 'mpmd.*.out' -+ run_mpmd.sh[105]out_files='./mpmd.5951061.17.out -./mpmd.5951061.0.out -./mpmd.5951061.6.out -./mpmd.5951061.15.out -./mpmd.5951061.9.out -./mpmd.5951061.2.out -./mpmd.5951061.3.out -./mpmd.5951061.23.out -./mpmd.5951061.4.out -./mpmd.5951061.22.out -./mpmd.5951061.1.out -./mpmd.5951061.7.out -./mpmd.5951061.21.out -./mpmd.5951061.12.out -./mpmd.5951061.19.out -./mpmd.5951061.8.out -./mpmd.5951061.16.out -./mpmd.5951061.18.out -./mpmd.5951061.13.out -./mpmd.5951061.10.out -./mpmd.5951061.20.out -./mpmd.5951061.14.out -./mpmd.5951061.5.out -./mpmd.5951061.11.out' -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.17.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.17.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.0.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.0.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.6.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.6.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.15.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.15.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.9.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.9.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.2.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.2.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.3.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.3.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.23.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.23.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.4.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.4.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.22.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.22.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.1.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.1.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.7.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.7.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.21.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.21.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.12.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.12.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.19.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.19.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.8.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.8.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.16.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.16.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.18.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.18.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.13.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.13.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.10.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.10.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.20.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.20.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.14.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.14.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.5.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.5.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.11.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.11.out -+ run_mpmd.sh[110]cat mpmd.out -17: + bash[8]'[' -z '' ']' -17: + bash[9]case "$-" in -17: + bash[12]__lmod_vx=x -17: + bash[16]'[' -n x ']' -17: + bash[16]set +x -17: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -17: Shell debugging restarted -17: + bash[224]unset __lmod_vx -17: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_18 -17: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_18 -17: + interp_atmos_master.sh[9]grid_string=0p25 -17: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -17: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -17: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -17: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -17: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -17: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -17: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -17: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -17: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -17: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -17: + interp_atmos_master.sh[31]IFS=: -17: + interp_atmos_master.sh[31]read -ra grids -17: + interp_atmos_master.sh[33]output_grids= -17: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -17: + interp_atmos_master.sh[35]gridopt=grid0p25 -17: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_18_0p25' -17: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_18 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_18_0p25 -17: 1:0:d=2021032312:TMP:825 mb:25 hour fcst: -17: 2:20811:d=2021032312:RH:825 mb:25 hour fcst: -17: 3:41352:d=2021032312:TCDC:825 mb:25 hour fcst: -17: 4:55705:d=2021032312:VVEL:825 mb:25 hour fcst: -17: 5:81315:d=2021032312:DZDT:825 mb:25 hour fcst: -17: 6:108178:d=2021032312:UGRD:825 mb:25 hour fcst: -17: 7:129358:d=2021032312:VGRD:825 mb:25 hour fcst: -17: 8:151206:d=2021032312:ABSV:825 mb:25 hour fcst: -17: 9:172453:d=2021032312:CLMR:825 mb:25 hour fcst: -17: 10:185458:d=2021032312:ICMR:825 mb:25 hour fcst: -17: 11:195872:d=2021032312:RWMR:825 mb:25 hour fcst: -17: 12:205273:d=2021032312:SNMR:825 mb:25 hour fcst: -17: 13:217635:d=2021032312:GRLE:825 mb:25 hour fcst: -17: 14:220102:d=2021032312:HGT:875 mb:25 hour fcst: -17: + interp_atmos_master.sh[47]export err=0 -17: + interp_atmos_master.sh[47]err=0 -17: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -17: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -17: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_18_0p25 -17: + product_functions.sh[5]local filename=pgb2bfile_f025_18_0p25 -17: + product_functions.sh[6]wgrib2 pgb2bfile_f025_18_0p25 -not_if :RH: -grib pgb2bfile_f025_18_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_18_0p25.new -17: 1:0:d=2021032312:TMP:825 mb:25 hour fcst: -17: 2:488834:d=2021032312:RH:825 mb:25 hour fcst: -17: 3:970204:d=2021032312:TCDC:825 mb:25 hour fcst: -17: 4:1258523:d=2021032312:VVEL:825 mb:25 hour fcst: -17: 5:1941419:d=2021032312:DZDT:825 mb:25 hour fcst: -17: 6:2680584:d=2021032312:UGRD:825 mb:25 hour fcst: -17: 7:3189950:d=2021032312:VGRD:825 mb:25 hour fcst: -17: 8:3729835:d=2021032312:ABSV:825 mb:25 hour fcst: -17: 9:4237948:d=2021032312:CLMR:825 mb:25 hour fcst: -17: 10:4517763:d=2021032312:ICMR:825 mb:25 hour fcst: -17: 11:4767027:d=2021032312:RWMR:825 mb:25 hour fcst: -17: 12:4948021:d=2021032312:SNMR:825 mb:25 hour fcst: -17: 13:5223309:d=2021032312:GRLE:825 mb:25 hour fcst: -17: 14:5268781:d=2021032312:HGT:875 mb:25 hour fcst: -17: + product_functions.sh[10]rc=0 -17: + product_functions.sh[11](( rc == 0 )) -17: + product_functions.sh[11]mv pgb2bfile_f025_18_0p25.new pgb2bfile_f025_18_0p25 -17: + product_functions.sh[12]return 0 -17: + interp_atmos_master.sh[56]export err=0 -17: + interp_atmos_master.sh[56]err=0 -17: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -17: ++ interp_atmos_master.sh[62]wc -l -17: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_18_0p25 -match 'LAND|ICEC' -17: + interp_atmos_master.sh[62]var_count=0 -17: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -17: + interp_atmos_master.sh[73]exit 0 - 0: + bash[8]'[' -z '' ']' - 0: + bash[9]case "$-" in - 0: + bash[12]__lmod_vx=x - 0: + bash[16]'[' -n x ']' - 0: + bash[16]set +x - 0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 0: Shell debugging restarted - 0: + bash[224]unset __lmod_vx - 0: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_1 - 0: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_1 - 0: + interp_atmos_master.sh[9]grid_string=0p25 - 0: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 0: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 0: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 0: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 0: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 0: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 0: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 0: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 0: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 0: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 0: + interp_atmos_master.sh[31]IFS=: - 0: + interp_atmos_master.sh[31]read -ra grids - 0: + interp_atmos_master.sh[33]output_grids= - 0: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 0: + interp_atmos_master.sh[35]gridopt=grid0p25 - 0: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_1_0p25' - 0: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_1 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_1_0p25 - 0: 1:0:d=2021032312:HGT:1 mb:25 hour fcst: - 0: 2:17997:d=2021032312:TMP:1 mb:25 hour fcst: - 0: 3:33857:d=2021032312:RH:1 mb:25 hour fcst: - 0: 4:43627:d=2021032312:UGRD:1 mb:25 hour fcst: - 0: 5:62225:d=2021032312:VGRD:1 mb:25 hour fcst: - 0: 6:79897:d=2021032312:ABSV:1 mb:25 hour fcst: - 0: 7:96195:d=2021032312:O3MR:1 mb:25 hour fcst: - 0: 8:117327:d=2021032312:HGT:2 mb:25 hour fcst: - 0: 9:137530:d=2021032312:TMP:2 mb:25 hour fcst: - 0: 10:154182:d=2021032312:RH:2 mb:25 hour fcst: - 0: 11:162864:d=2021032312:UGRD:2 mb:25 hour fcst: - 0: 12:174427:d=2021032312:VGRD:2 mb:25 hour fcst: - 0: 13:192855:d=2021032312:ABSV:2 mb:25 hour fcst: - 0: 14:209874:d=2021032312:O3MR:2 mb:25 hour fcst: - 0: + interp_atmos_master.sh[47]export err=0 - 0: + interp_atmos_master.sh[47]err=0 - 0: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 0: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 0: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_1_0p25 - 0: + product_functions.sh[5]local filename=pgb2bfile_f025_1_0p25 - 0: + product_functions.sh[6]wgrib2 pgb2bfile_f025_1_0p25 -not_if :RH: -grib pgb2bfile_f025_1_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_1_0p25.new - 0: 1:0:d=2021032312:HGT:1 mb:25 hour fcst: - 0: 2:407129:d=2021032312:TMP:1 mb:25 hour fcst: - 0: 3:754609:d=2021032312:RH:1 mb:25 hour fcst: - 0: 4:1011544:d=2021032312:UGRD:1 mb:25 hour fcst: - 0: 5:1443549:d=2021032312:VGRD:1 mb:25 hour fcst: - 0: 6:1850190:d=2021032312:ABSV:1 mb:25 hour fcst: - 0: 7:2192850:d=2021032312:O3MR:1 mb:25 hour fcst: - 0: 8:2728058:d=2021032312:HGT:2 mb:25 hour fcst: - 0: 9:3214694:d=2021032312:TMP:2 mb:25 hour fcst: - 0: 10:3584903:d=2021032312:RH:2 mb:25 hour fcst: - 0: 11:3832850:d=2021032312:UGRD:2 mb:25 hour fcst: - 0: 12:4290183:d=2021032312:VGRD:2 mb:25 hour fcst: - 0: 13:4719847:d=2021032312:ABSV:2 mb:25 hour fcst: - 0: 14:5092110:d=2021032312:O3MR:2 mb:25 hour fcst: - 0: + product_functions.sh[10]rc=0 - 0: + product_functions.sh[11](( rc == 0 )) - 0: + product_functions.sh[11]mv pgb2bfile_f025_1_0p25.new pgb2bfile_f025_1_0p25 - 0: + product_functions.sh[12]return 0 - 0: + interp_atmos_master.sh[56]export err=0 - 0: + interp_atmos_master.sh[56]err=0 - 0: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 0: ++ interp_atmos_master.sh[62]wc -l - 0: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_1_0p25 -match 'LAND|ICEC' - 0: + interp_atmos_master.sh[62]var_count=0 - 0: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 0: + interp_atmos_master.sh[73]exit 0 - 6: + bash[8]'[' -z '' ']' - 6: + bash[9]case "$-" in - 6: + bash[12]__lmod_vx=x - 6: + bash[16]'[' -n x ']' - 6: + bash[16]set +x - 6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 6: Shell debugging restarted - 6: + bash[224]unset __lmod_vx - 6: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_7 - 6: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_7 - 6: + interp_atmos_master.sh[9]grid_string=0p25 - 6: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 6: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 6: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 6: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 6: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 6: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 6: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 6: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 6: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 6: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 6: + interp_atmos_master.sh[31]IFS=: - 6: + interp_atmos_master.sh[31]read -ra grids - 6: + interp_atmos_master.sh[33]output_grids= - 6: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 6: + interp_atmos_master.sh[35]gridopt=grid0p25 - 6: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_7_0p25' - 6: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_7 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_7_0p25 - 6: 1:0:d=2021032312:TMP:275 mb:25 hour fcst: - 6: 2:18837:d=2021032312:RH:275 mb:25 hour fcst: - 6: 3:37276:d=2021032312:TCDC:275 mb:25 hour fcst: - 6: 4:49026:d=2021032312:VVEL:275 mb:25 hour fcst: - 6: 5:72338:d=2021032312:DZDT:275 mb:25 hour fcst: - 6: 6:99971:d=2021032312:UGRD:275 mb:25 hour fcst: - 6: 7:114286:d=2021032312:VGRD:275 mb:25 hour fcst: - 6: 8:129458:d=2021032312:ABSV:275 mb:25 hour fcst: - 6: 9:151512:d=2021032312:CLMR:275 mb:25 hour fcst: - 6: 10:151691:d=2021032312:ICMR:275 mb:25 hour fcst: - 6: 11:166211:d=2021032312:RWMR:275 mb:25 hour fcst: - 6: 12:166390:d=2021032312:SNMR:275 mb:25 hour fcst: - 6: 13:176339:d=2021032312:GRLE:275 mb:25 hour fcst: - 6: 14:176649:d=2021032312:HGT:325 mb:25 hour fcst: - 6: + interp_atmos_master.sh[47]export err=0 - 6: + interp_atmos_master.sh[47]err=0 - 6: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 6: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 6: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_7_0p25 - 6: + product_functions.sh[5]local filename=pgb2bfile_f025_7_0p25 - 6: + product_functions.sh[6]wgrib2 pgb2bfile_f025_7_0p25 -not_if :RH: -grib pgb2bfile_f025_7_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_7_0p25.new - 6: 1:0:d=2021032312:TMP:275 mb:25 hour fcst: - 6: 2:416314:d=2021032312:RH:275 mb:25 hour fcst: - 6: 3:835013:d=2021032312:TCDC:275 mb:25 hour fcst: - 6: 4:1060738:d=2021032312:VVEL:275 mb:25 hour fcst: - 6: 5:1676736:d=2021032312:DZDT:275 mb:25 hour fcst: - 6: 6:2454352:d=2021032312:UGRD:275 mb:25 hour fcst: - 6: 7:2771367:d=2021032312:VGRD:275 mb:25 hour fcst: - 6: 8:3103624:d=2021032312:ABSV:275 mb:25 hour fcst: - 6: 9:3647384:d=2021032312:CLMR:275 mb:25 hour fcst: - 6: 10:3647563:d=2021032312:ICMR:275 mb:25 hour fcst: - 6: 11:3961906:d=2021032312:RWMR:275 mb:25 hour fcst: - 6: 12:3962085:d=2021032312:SNMR:275 mb:25 hour fcst: - 6: 13:4158815:d=2021032312:GRLE:275 mb:25 hour fcst: - 6: 14:4160967:d=2021032312:HGT:325 mb:25 hour fcst: - 6: + product_functions.sh[10]rc=0 - 6: + product_functions.sh[11](( rc == 0 )) - 6: + product_functions.sh[11]mv pgb2bfile_f025_7_0p25.new pgb2bfile_f025_7_0p25 - 6: + product_functions.sh[12]return 0 - 6: + interp_atmos_master.sh[56]export err=0 - 6: + interp_atmos_master.sh[56]err=0 - 6: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 6: ++ interp_atmos_master.sh[62]wc -l - 6: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_7_0p25 -match 'LAND|ICEC' - 6: + interp_atmos_master.sh[62]var_count=0 - 6: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 6: + interp_atmos_master.sh[73]exit 0 -15: + bash[8]'[' -z '' ']' -15: + bash[9]case "$-" in -15: + bash[12]__lmod_vx=x -15: + bash[16]'[' -n x ']' -15: + bash[16]set +x -15: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -15: Shell debugging restarted -15: + bash[224]unset __lmod_vx -15: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_16 -15: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_16 -15: + interp_atmos_master.sh[9]grid_string=0p25 -15: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -15: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -15: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -15: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -15: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -15: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -15: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -15: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -15: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -15: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -15: + interp_atmos_master.sh[31]IFS=: -15: + interp_atmos_master.sh[31]read -ra grids -15: + interp_atmos_master.sh[33]output_grids= -15: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -15: + interp_atmos_master.sh[35]gridopt=grid0p25 -15: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_16_0p25' -15: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_16 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_16_0p25 -15: 1:0:d=2021032312:TMP:725 mb:25 hour fcst: -15: 2:19731:d=2021032312:RH:725 mb:25 hour fcst: -15: 3:40039:d=2021032312:TCDC:725 mb:25 hour fcst: -15: 4:51915:d=2021032312:VVEL:725 mb:25 hour fcst: -15: 5:77661:d=2021032312:DZDT:725 mb:25 hour fcst: -15: 6:105000:d=2021032312:UGRD:725 mb:25 hour fcst: -15: 7:125845:d=2021032312:VGRD:725 mb:25 hour fcst: -15: 8:147499:d=2021032312:ABSV:725 mb:25 hour fcst: -15: 9:168600:d=2021032312:CLMR:725 mb:25 hour fcst: -15: 10:176450:d=2021032312:ICMR:725 mb:25 hour fcst: -15: 11:188036:d=2021032312:RWMR:725 mb:25 hour fcst: -15: 12:193736:d=2021032312:SNMR:725 mb:25 hour fcst: -15: 13:207545:d=2021032312:GRLE:725 mb:25 hour fcst: -15: 14:209465:d=2021032312:HGT:775 mb:25 hour fcst: -15: + interp_atmos_master.sh[47]export err=0 -15: + interp_atmos_master.sh[47]err=0 -15: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -15: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -15: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_16_0p25 -15: + product_functions.sh[5]local filename=pgb2bfile_f025_16_0p25 -15: + product_functions.sh[6]wgrib2 pgb2bfile_f025_16_0p25 -not_if :RH: -grib pgb2bfile_f025_16_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_16_0p25.new -15: 1:0:d=2021032312:TMP:725 mb:25 hour fcst: -15: 2:456065:d=2021032312:RH:725 mb:25 hour fcst: -15: 3:930876:d=2021032312:TCDC:725 mb:25 hour fcst: -15: 4:1165081:d=2021032312:VVEL:725 mb:25 hour fcst: -15: 5:1849732:d=2021032312:DZDT:725 mb:25 hour fcst: -15: 6:2603410:d=2021032312:UGRD:725 mb:25 hour fcst: -15: 7:3106359:d=2021032312:VGRD:725 mb:25 hour fcst: -15: 8:3639317:d=2021032312:ABSV:725 mb:25 hour fcst: -15: 9:4145307:d=2021032312:CLMR:725 mb:25 hour fcst: -15: 10:4316359:d=2021032312:ICMR:725 mb:25 hour fcst: -15: 11:4587167:d=2021032312:RWMR:725 mb:25 hour fcst: -15: 12:4697591:d=2021032312:SNMR:725 mb:25 hour fcst: -15: 13:5003081:d=2021032312:GRLE:725 mb:25 hour fcst: -15: 14:5034635:d=2021032312:HGT:775 mb:25 hour fcst: -15: + product_functions.sh[10]rc=0 -15: + product_functions.sh[11](( rc == 0 )) -15: + product_functions.sh[11]mv pgb2bfile_f025_16_0p25.new pgb2bfile_f025_16_0p25 -15: + product_functions.sh[12]return 0 -15: + interp_atmos_master.sh[56]export err=0 -15: + interp_atmos_master.sh[56]err=0 -15: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -15: ++ interp_atmos_master.sh[62]wc -l -15: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_16_0p25 -match 'LAND|ICEC' -15: + interp_atmos_master.sh[62]var_count=0 -15: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -15: + interp_atmos_master.sh[73]exit 0 - 9: + bash[8]'[' -z '' ']' - 9: + bash[9]case "$-" in - 9: + bash[12]__lmod_vx=x - 9: + bash[16]'[' -n x ']' - 9: + bash[16]set +x - 9: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 9: Shell debugging restarted - 9: + bash[224]unset __lmod_vx - 9: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_10 - 9: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_10 - 9: + interp_atmos_master.sh[9]grid_string=0p25 - 9: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 9: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 9: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 9: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 9: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 9: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 9: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 9: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 9: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 9: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 9: + interp_atmos_master.sh[31]IFS=: - 9: + interp_atmos_master.sh[31]read -ra grids - 9: + interp_atmos_master.sh[33]output_grids= - 9: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 9: + interp_atmos_master.sh[35]gridopt=grid0p25 - 9: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_10_0p25' - 9: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_10 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_10_0p25 - 9: 1:0:d=2021032312:TMP:425 mb:25 hour fcst: - 9: 2:18521:d=2021032312:RH:425 mb:25 hour fcst: - 9: 3:37823:d=2021032312:TCDC:425 mb:25 hour fcst: - 9: 4:50696:d=2021032312:VVEL:425 mb:25 hour fcst: - 9: 5:75591:d=2021032312:DZDT:425 mb:25 hour fcst: - 9: 6:103651:d=2021032312:UGRD:425 mb:25 hour fcst: - 9: 7:125488:d=2021032312:VGRD:425 mb:25 hour fcst: - 9: 8:147945:d=2021032312:ABSV:425 mb:25 hour fcst: - 9: 9:169559:d=2021032312:CLMR:425 mb:25 hour fcst: - 9: 10:170197:d=2021032312:ICMR:425 mb:25 hour fcst: - 9: 11:187112:d=2021032312:RWMR:425 mb:25 hour fcst: - 9: 12:187511:d=2021032312:SNMR:425 mb:25 hour fcst: - 9: 13:200615:d=2021032312:GRLE:425 mb:25 hour fcst: - 9: 14:201656:d=2021032312:HGT:475 mb:25 hour fcst: - 9: + interp_atmos_master.sh[47]export err=0 - 9: + interp_atmos_master.sh[47]err=0 - 9: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 9: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 9: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_10_0p25 - 9: + product_functions.sh[5]local filename=pgb2bfile_f025_10_0p25 - 9: + product_functions.sh[6]wgrib2 pgb2bfile_f025_10_0p25 -not_if :RH: -grib pgb2bfile_f025_10_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_10_0p25.new - 9: 1:0:d=2021032312:TMP:425 mb:25 hour fcst: - 9: 2:414355:d=2021032312:RH:425 mb:25 hour fcst: - 9: 3:855398:d=2021032312:TCDC:425 mb:25 hour fcst: - 9: 4:1094576:d=2021032312:VVEL:425 mb:25 hour fcst: - 9: 5:1758552:d=2021032312:DZDT:425 mb:25 hour fcst: - 9: 6:2547908:d=2021032312:UGRD:425 mb:25 hour fcst: - 9: 7:3088564:d=2021032312:VGRD:425 mb:25 hour fcst: - 9: 8:3658109:d=2021032312:ABSV:425 mb:25 hour fcst: - 9: 9:4188306:d=2021032312:CLMR:425 mb:25 hour fcst: - 9: 10:4198202:d=2021032312:ICMR:425 mb:25 hour fcst: - 9: 11:4571788:d=2021032312:RWMR:425 mb:25 hour fcst: - 9: 12:4575404:d=2021032312:SNMR:425 mb:25 hour fcst: - 9: 13:4848510:d=2021032312:GRLE:425 mb:25 hour fcst: - 9: 14:4861309:d=2021032312:HGT:475 mb:25 hour fcst: - 9: + product_functions.sh[10]rc=0 - 9: + product_functions.sh[11](( rc == 0 )) - 9: + product_functions.sh[11]mv pgb2bfile_f025_10_0p25.new pgb2bfile_f025_10_0p25 - 9: + product_functions.sh[12]return 0 - 9: + interp_atmos_master.sh[56]export err=0 - 9: + interp_atmos_master.sh[56]err=0 - 9: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 9: ++ interp_atmos_master.sh[62]wc -l - 9: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_10_0p25 -match 'LAND|ICEC' - 9: + interp_atmos_master.sh[62]var_count=0 - 9: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 9: + interp_atmos_master.sh[73]exit 0 - 2: + bash[8]'[' -z '' ']' - 2: + bash[9]case "$-" in - 2: + bash[12]__lmod_vx=x - 2: + bash[16]'[' -n x ']' - 2: + bash[16]set +x - 2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 2: Shell debugging restarted - 2: + bash[224]unset __lmod_vx - 2: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_3 - 2: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_3 - 2: + interp_atmos_master.sh[9]grid_string=0p25 - 2: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 2: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 2: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 2: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 2: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 2: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 2: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 2: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 2: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 2: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 2: + interp_atmos_master.sh[31]IFS=: - 2: + interp_atmos_master.sh[31]read -ra grids - 2: + interp_atmos_master.sh[33]output_grids= - 2: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 2: + interp_atmos_master.sh[35]gridopt=grid0p25 - 2: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_3_0p25' - 2: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_3 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_3_0p25 - 2: 1:0:d=2021032312:HGT:7 mb:25 hour fcst: - 2: 2:17665:d=2021032312:TMP:7 mb:25 hour fcst: - 2: 3:33749:d=2021032312:RH:7 mb:25 hour fcst: - 2: 4:39799:d=2021032312:UGRD:7 mb:25 hour fcst: - 2: 5:51216:d=2021032312:VGRD:7 mb:25 hour fcst: - 2: 6:69969:d=2021032312:ABSV:7 mb:25 hour fcst: - 2: 7:87704:d=2021032312:O3MR:7 mb:25 hour fcst: - 2: 8:111371:d=2021032312:TCDC:70 mb:25 hour fcst: - 2: 9:111550:d=2021032312:CLMR:70 mb:25 hour fcst: - 2: 10:111729:d=2021032312:ICMR:70 mb:25 hour fcst: - 2: 11:111908:d=2021032312:RWMR:70 mb:25 hour fcst: - 2: 12:112087:d=2021032312:SNMR:70 mb:25 hour fcst: - 2: 13:112266:d=2021032312:GRLE:70 mb:25 hour fcst: - 2: 14:112445:d=2021032312:HGT:125 mb:25 hour fcst: - 2: + interp_atmos_master.sh[47]export err=0 - 2: + interp_atmos_master.sh[47]err=0 - 2: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 2: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 2: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_3_0p25 - 2: + product_functions.sh[5]local filename=pgb2bfile_f025_3_0p25 - 2: + product_functions.sh[6]wgrib2 pgb2bfile_f025_3_0p25 -not_if :RH: -grib pgb2bfile_f025_3_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_3_0p25.new - 2: 1:0:d=2021032312:HGT:7 mb:25 hour fcst: - 2: 2:396521:d=2021032312:TMP:7 mb:25 hour fcst: - 2: 3:746511:d=2021032312:RH:7 mb:25 hour fcst: - 2: 4:921406:d=2021032312:UGRD:7 mb:25 hour fcst: - 2: 5:1366129:d=2021032312:VGRD:7 mb:25 hour fcst: - 2: 6:1798409:d=2021032312:ABSV:7 mb:25 hour fcst: - 2: 7:2192608:d=2021032312:O3MR:7 mb:25 hour fcst: - 2: 8:2796546:d=2021032312:TCDC:70 mb:25 hour fcst: - 2: 9:2796725:d=2021032312:CLMR:70 mb:25 hour fcst: - 2: 10:2796904:d=2021032312:ICMR:70 mb:25 hour fcst: - 2: 11:2797083:d=2021032312:RWMR:70 mb:25 hour fcst: - 2: 12:2797262:d=2021032312:SNMR:70 mb:25 hour fcst: - 2: 13:2797441:d=2021032312:GRLE:70 mb:25 hour fcst: - 2: 14:2797620:d=2021032312:HGT:125 mb:25 hour fcst: - 2: + product_functions.sh[10]rc=0 - 2: + product_functions.sh[11](( rc == 0 )) - 2: + product_functions.sh[11]mv pgb2bfile_f025_3_0p25.new pgb2bfile_f025_3_0p25 - 2: + product_functions.sh[12]return 0 - 2: + interp_atmos_master.sh[56]export err=0 - 2: + interp_atmos_master.sh[56]err=0 - 2: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 2: ++ interp_atmos_master.sh[62]wc -l - 2: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_3_0p25 -match 'LAND|ICEC' - 2: + interp_atmos_master.sh[62]var_count=0 - 2: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 2: + interp_atmos_master.sh[73]exit 0 - 3: + bash[8]'[' -z '' ']' - 3: + bash[9]case "$-" in - 3: + bash[12]__lmod_vx=x - 3: + bash[16]'[' -n x ']' - 3: + bash[16]set +x - 3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 3: Shell debugging restarted - 3: + bash[224]unset __lmod_vx - 3: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_4 - 3: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_4 - 3: + interp_atmos_master.sh[9]grid_string=0p25 - 3: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 3: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 3: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 3: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 3: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 3: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 3: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 3: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 3: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 3: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 3: + interp_atmos_master.sh[31]IFS=: - 3: + interp_atmos_master.sh[31]read -ra grids - 3: + interp_atmos_master.sh[33]output_grids= - 3: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 3: + interp_atmos_master.sh[35]gridopt=grid0p25 - 3: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_4_0p25' - 3: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_4 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_4_0p25 - 3: 1:0:d=2021032312:TMP:125 mb:25 hour fcst: - 3: 2:17009:d=2021032312:RH:125 mb:25 hour fcst: - 3: 3:27519:d=2021032312:TCDC:125 mb:25 hour fcst: - 3: 4:29956:d=2021032312:VVEL:125 mb:25 hour fcst: - 3: 5:56151:d=2021032312:DZDT:125 mb:25 hour fcst: - 3: 6:81243:d=2021032312:UGRD:125 mb:25 hour fcst: - 3: 7:100860:d=2021032312:VGRD:125 mb:25 hour fcst: - 3: 8:120489:d=2021032312:ABSV:125 mb:25 hour fcst: - 3: 9:138996:d=2021032312:CLMR:125 mb:25 hour fcst: - 3: 10:139175:d=2021032312:ICMR:125 mb:25 hour fcst: - 3: 11:143668:d=2021032312:RWMR:125 mb:25 hour fcst: - 3: 12:143847:d=2021032312:SNMR:125 mb:25 hour fcst: - 3: 13:147664:d=2021032312:GRLE:125 mb:25 hour fcst: - 3: 14:148075:d=2021032312:HGT:175 mb:25 hour fcst: - 3: + interp_atmos_master.sh[47]export err=0 - 3: + interp_atmos_master.sh[47]err=0 - 3: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 3: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 3: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_4_0p25 - 3: + product_functions.sh[5]local filename=pgb2bfile_f025_4_0p25 - 3: + product_functions.sh[6]wgrib2 pgb2bfile_f025_4_0p25 -not_if :RH: -grib pgb2bfile_f025_4_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_4_0p25.new - 3: 1:0:d=2021032312:TMP:125 mb:25 hour fcst: - 3: 2:377162:d=2021032312:RH:125 mb:25 hour fcst: - 3: 3:624474:d=2021032312:TCDC:125 mb:25 hour fcst: - 3: 4:675058:d=2021032312:VVEL:125 mb:25 hour fcst: - 3: 5:1392568:d=2021032312:DZDT:125 mb:25 hour fcst: - 3: 6:2075890:d=2021032312:UGRD:125 mb:25 hour fcst: - 3: 7:2559030:d=2021032312:VGRD:125 mb:25 hour fcst: - 3: 8:3038809:d=2021032312:ABSV:125 mb:25 hour fcst: - 3: 9:3458310:d=2021032312:CLMR:125 mb:25 hour fcst: - 3: 10:3458489:d=2021032312:ICMR:125 mb:25 hour fcst: - 3: 11:3561971:d=2021032312:RWMR:125 mb:25 hour fcst: - 3: 12:3562150:d=2021032312:SNMR:125 mb:25 hour fcst: - 3: 13:3630904:d=2021032312:GRLE:125 mb:25 hour fcst: - 3: 14:3635280:d=2021032312:HGT:175 mb:25 hour fcst: - 3: + product_functions.sh[10]rc=0 - 3: + product_functions.sh[11](( rc == 0 )) - 3: + product_functions.sh[11]mv pgb2bfile_f025_4_0p25.new pgb2bfile_f025_4_0p25 - 3: + product_functions.sh[12]return 0 - 3: + interp_atmos_master.sh[56]export err=0 - 3: + interp_atmos_master.sh[56]err=0 - 3: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 3: ++ interp_atmos_master.sh[62]wc -l - 3: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_4_0p25 -match 'LAND|ICEC' - 3: + interp_atmos_master.sh[62]var_count=0 - 3: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 3: + interp_atmos_master.sh[73]exit 0 -23: + bash[8]'[' -z '' ']' -23: + bash[9]case "$-" in -23: + bash[12]__lmod_vx=x -23: + bash[16]'[' -n x ']' -23: + bash[16]set +x -23: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -23: Shell debugging restarted -23: + bash[224]unset __lmod_vx -23: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_24 -23: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_24 -23: + interp_atmos_master.sh[9]grid_string=0p25 -23: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -23: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -23: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -23: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -23: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -23: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -23: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -23: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -23: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -23: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -23: + interp_atmos_master.sh[31]IFS=: -23: + interp_atmos_master.sh[31]read -ra grids -23: + interp_atmos_master.sh[33]output_grids= -23: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -23: + interp_atmos_master.sh[35]gridopt=grid0p25 -23: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_24_0p25' -23: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_24 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_24_0p25 -23: 1:0:d=2021032312:HGT:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -23: 2:17645:d=2021032312:PRES:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -23: 3:35338:d=2021032312:VWSH:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -23: 4:46750:d=2021032312:UGRD:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 5:58075:d=2021032312:VGRD:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 6:69387:d=2021032312:TMP:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 7:85324:d=2021032312:HGT:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 8:103892:d=2021032312:PRES:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 9:122570:d=2021032312:VWSH:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 10:134784:d=2021032312:UGRD:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 11:150152:d=2021032312:VGRD:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 12:161477:d=2021032312:TMP:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 13:177271:d=2021032312:HGT:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 14:195668:d=2021032312:PRES:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 15:214129:d=2021032312:VWSH:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 16:226460:d=2021032312:UGRD:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 17:237506:d=2021032312:VGRD:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 18:248492:d=2021032312:TMP:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 19:263474:d=2021032312:HGT:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 20:281403:d=2021032312:PRES:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 21:299215:d=2021032312:VWSH:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 22:311337:d=2021032312:UGRD:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 23:322384:d=2021032312:VGRD:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 24:333338:d=2021032312:TMP:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 25:348400:d=2021032312:HGT:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 26:366322:d=2021032312:PRES:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 27:384073:d=2021032312:VWSH:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: + interp_atmos_master.sh[47]export err=0 -23: + interp_atmos_master.sh[47]err=0 -23: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -23: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -23: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_24_0p25 -23: + product_functions.sh[5]local filename=pgb2bfile_f025_24_0p25 -23: + product_functions.sh[6]wgrib2 pgb2bfile_f025_24_0p25 -not_if :RH: -grib pgb2bfile_f025_24_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_24_0p25.new -23: 1:0:d=2021032312:HGT:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -23: 2:583741:d=2021032312:PRES:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -23: 3:1160560:d=2021032312:VWSH:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -23: 4:1502070:d=2021032312:UGRD:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 5:1829936:d=2021032312:VGRD:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 6:2155895:d=2021032312:TMP:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 7:2631117:d=2021032312:HGT:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 8:3226652:d=2021032312:PRES:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 9:3818277:d=2021032312:VWSH:PV=1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 10:4176602:d=2021032312:UGRD:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 11:4503976:d=2021032312:VGRD:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 12:4827109:d=2021032312:TMP:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 13:5306775:d=2021032312:HGT:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 14:5894880:d=2021032312:PRES:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 15:6483396:d=2021032312:VWSH:PV=-1e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 16:6841365:d=2021032312:UGRD:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 17:7162845:d=2021032312:VGRD:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 18:7480348:d=2021032312:TMP:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 19:7925658:d=2021032312:HGT:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 20:8492731:d=2021032312:PRES:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 21:9036530:d=2021032312:VWSH:PV=1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 22:9392321:d=2021032312:UGRD:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 23:9715895:d=2021032312:VGRD:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 24:10034374:d=2021032312:TMP:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 25:10488096:d=2021032312:HGT:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 26:11058313:d=2021032312:PRES:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: 27:11601772:d=2021032312:VWSH:PV=-1.5e-06 (Km^2/kg/s) surface:25 hour fcst: -23: + product_functions.sh[10]rc=0 -23: + product_functions.sh[11](( rc == 0 )) -23: + product_functions.sh[11]mv pgb2bfile_f025_24_0p25.new pgb2bfile_f025_24_0p25 -23: + product_functions.sh[12]return 0 -23: + interp_atmos_master.sh[56]export err=0 -23: + interp_atmos_master.sh[56]err=0 -23: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -23: ++ interp_atmos_master.sh[62]wc -l -23: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_24_0p25 -match 'LAND|ICEC' -23: + interp_atmos_master.sh[62]var_count=0 -23: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -23: + interp_atmos_master.sh[73]exit 0 - 4: + bash[8]'[' -z '' ']' - 4: + bash[9]case "$-" in - 4: + bash[12]__lmod_vx=x - 4: + bash[16]'[' -n x ']' - 4: + bash[16]set +x - 4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 4: Shell debugging restarted - 4: + bash[224]unset __lmod_vx - 4: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_5 - 4: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_5 - 4: + interp_atmos_master.sh[9]grid_string=0p25 - 4: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 4: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 4: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 4: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 4: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 4: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 4: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 4: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 4: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 4: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 4: + interp_atmos_master.sh[31]IFS=: - 4: + interp_atmos_master.sh[31]read -ra grids - 4: + interp_atmos_master.sh[33]output_grids= - 4: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 4: + interp_atmos_master.sh[35]gridopt=grid0p25 - 4: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_5_0p25' - 4: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_5 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_5_0p25 - 4: 1:0:d=2021032312:TMP:175 mb:25 hour fcst: - 4: 2:17482:d=2021032312:RH:175 mb:25 hour fcst: - 4: 3:30614:d=2021032312:TCDC:175 mb:25 hour fcst: - 4: 4:35322:d=2021032312:VVEL:175 mb:25 hour fcst: - 4: 5:63524:d=2021032312:DZDT:175 mb:25 hour fcst: - 4: 6:89584:d=2021032312:UGRD:175 mb:25 hour fcst: - 4: 7:102583:d=2021032312:VGRD:175 mb:25 hour fcst: - 4: 8:123325:d=2021032312:ABSV:175 mb:25 hour fcst: - 4: 9:142860:d=2021032312:CLMR:175 mb:25 hour fcst: - 4: 10:143039:d=2021032312:ICMR:175 mb:25 hour fcst: - 4: 11:149669:d=2021032312:RWMR:175 mb:25 hour fcst: - 4: 12:149848:d=2021032312:SNMR:175 mb:25 hour fcst: - 4: 13:153518:d=2021032312:GRLE:175 mb:25 hour fcst: - 4: 14:153997:d=2021032312:HGT:225 mb:25 hour fcst: - 4: + interp_atmos_master.sh[47]export err=0 - 4: + interp_atmos_master.sh[47]err=0 - 4: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 4: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 4: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_5_0p25 - 4: + product_functions.sh[5]local filename=pgb2bfile_f025_5_0p25 - 4: + product_functions.sh[6]wgrib2 pgb2bfile_f025_5_0p25 -not_if :RH: -grib pgb2bfile_f025_5_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_5_0p25.new - 4: 1:0:d=2021032312:TMP:175 mb:25 hour fcst: - 4: 2:392131:d=2021032312:RH:175 mb:25 hour fcst: - 4: 3:693026:d=2021032312:TCDC:175 mb:25 hour fcst: - 4: 4:788721:d=2021032312:VVEL:175 mb:25 hour fcst: - 4: 5:1568841:d=2021032312:DZDT:175 mb:25 hour fcst: - 4: 6:2283215:d=2021032312:UGRD:175 mb:25 hour fcst: - 4: 7:2799156:d=2021032312:VGRD:175 mb:25 hour fcst: - 4: 8:3317702:d=2021032312:ABSV:175 mb:25 hour fcst: - 4: 9:3765729:d=2021032312:CLMR:175 mb:25 hour fcst: - 4: 10:3765908:d=2021032312:ICMR:175 mb:25 hour fcst: - 4: 11:3913980:d=2021032312:RWMR:175 mb:25 hour fcst: - 4: 12:3914159:d=2021032312:SNMR:175 mb:25 hour fcst: - 4: 13:3979429:d=2021032312:GRLE:175 mb:25 hour fcst: - 4: 14:3985830:d=2021032312:HGT:225 mb:25 hour fcst: - 4: + product_functions.sh[10]rc=0 - 4: + product_functions.sh[11](( rc == 0 )) - 4: + product_functions.sh[11]mv pgb2bfile_f025_5_0p25.new pgb2bfile_f025_5_0p25 - 4: + product_functions.sh[12]return 0 - 4: + interp_atmos_master.sh[56]export err=0 - 4: + interp_atmos_master.sh[56]err=0 - 4: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 4: ++ interp_atmos_master.sh[62]wc -l - 4: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_5_0p25 -match 'LAND|ICEC' - 4: + interp_atmos_master.sh[62]var_count=0 - 4: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 4: + interp_atmos_master.sh[73]exit 0 -22: + bash[8]'[' -z '' ']' -22: + bash[9]case "$-" in -22: + bash[12]__lmod_vx=x -22: + bash[16]'[' -n x ']' -22: + bash[16]set +x -22: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -22: Shell debugging restarted -22: + bash[224]unset __lmod_vx -22: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_23 -22: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_23 -22: + interp_atmos_master.sh[9]grid_string=0p25 -22: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -22: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -22: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -22: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -22: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -22: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -22: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -22: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -22: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -22: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -22: + interp_atmos_master.sh[31]IFS=: -22: + interp_atmos_master.sh[31]read -ra grids -22: + interp_atmos_master.sh[33]output_grids= -22: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -22: + interp_atmos_master.sh[35]gridopt=grid0p25 -22: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_23_0p25' -22: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_23 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_23_0p25 -22: 1:0:d=2021032312:TMP:180-150 mb above ground:25 hour fcst: -22: 2:21164:d=2021032312:RH:180-150 mb above ground:25 hour fcst: -22: 3:41152:d=2021032312:SPFH:180-150 mb above ground:25 hour fcst: -22: 4:70411:d=2021032312:UGRD:180-150 mb above ground:25 hour fcst: -22: 5:91493:d=2021032312:VGRD:180-150 mb above ground:25 hour fcst: -22: 6:113138:d=2021032312:UGRD:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -22: 7:123736:d=2021032312:VGRD:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -22: 8:134301:d=2021032312:TMP:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -22: 9:148909:d=2021032312:HGT:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -22: 10:165641:d=2021032312:PRES:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -22: 11:182482:d=2021032312:VWSH:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -22: 12:193446:d=2021032312:UGRD:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -22: 13:204414:d=2021032312:VGRD:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -22: 14:218925:d=2021032312:TMP:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -22: + interp_atmos_master.sh[47]export err=0 -22: + interp_atmos_master.sh[47]err=0 -22: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -22: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -22: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_23_0p25 -22: + product_functions.sh[5]local filename=pgb2bfile_f025_23_0p25 -22: + product_functions.sh[6]wgrib2 pgb2bfile_f025_23_0p25 -not_if :RH: -grib pgb2bfile_f025_23_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_23_0p25.new -22: 1:0:d=2021032312:TMP:180-150 mb above ground:25 hour fcst: -22: 2:507278:d=2021032312:RH:180-150 mb above ground:25 hour fcst: -22: 3:962471:d=2021032312:SPFH:180-150 mb above ground:25 hour fcst: -22: 4:1756579:d=2021032312:UGRD:180-150 mb above ground:25 hour fcst: -22: 5:2259092:d=2021032312:VGRD:180-150 mb above ground:25 hour fcst: -22: 6:2789900:d=2021032312:UGRD:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -22: 7:3104173:d=2021032312:VGRD:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -22: 8:3414035:d=2021032312:TMP:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -22: 9:3879002:d=2021032312:HGT:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -22: 10:4424625:d=2021032312:PRES:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -22: 11:4978741:d=2021032312:VWSH:PV=5e-07 (Km^2/kg/s) surface:25 hour fcst: -22: 12:5313452:d=2021032312:UGRD:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -22: 13:5771649:d=2021032312:VGRD:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -22: 14:6229427:d=2021032312:TMP:PV=-5e-07 (Km^2/kg/s) surface:25 hour fcst: -22: + product_functions.sh[10]rc=0 -22: + product_functions.sh[11](( rc == 0 )) -22: + product_functions.sh[11]mv pgb2bfile_f025_23_0p25.new pgb2bfile_f025_23_0p25 -22: + product_functions.sh[12]return 0 -22: + interp_atmos_master.sh[56]export err=0 -22: + interp_atmos_master.sh[56]err=0 -22: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -22: ++ interp_atmos_master.sh[62]wc -l -22: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_23_0p25 -match 'LAND|ICEC' -22: + interp_atmos_master.sh[62]var_count=0 -22: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -22: + interp_atmos_master.sh[73]exit 0 - 1: + bash[8]'[' -z '' ']' - 1: + bash[9]case "$-" in - 1: + bash[12]__lmod_vx=x - 1: + bash[16]'[' -n x ']' - 1: + bash[16]set +x - 1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 1: Shell debugging restarted - 1: + bash[224]unset __lmod_vx - 1: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_2 - 1: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_2 - 1: + interp_atmos_master.sh[9]grid_string=0p25 - 1: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 1: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 1: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 1: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 1: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 1: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 1: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 1: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 1: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 1: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 1: + interp_atmos_master.sh[31]IFS=: - 1: + interp_atmos_master.sh[31]read -ra grids - 1: + interp_atmos_master.sh[33]output_grids= - 1: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 1: + interp_atmos_master.sh[35]gridopt=grid0p25 - 1: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_2_0p25' - 1: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_2 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_2_0p25 - 1: 1:0:d=2021032312:HGT:3 mb:25 hour fcst: - 1: 2:17800:d=2021032312:TMP:3 mb:25 hour fcst: - 1: 3:34288:d=2021032312:RH:3 mb:25 hour fcst: - 1: 4:40278:d=2021032312:UGRD:3 mb:25 hour fcst: - 1: 5:51644:d=2021032312:VGRD:3 mb:25 hour fcst: - 1: 6:69636:d=2021032312:ABSV:3 mb:25 hour fcst: - 1: 7:86554:d=2021032312:O3MR:3 mb:25 hour fcst: - 1: 8:109856:d=2021032312:HGT:5 mb:25 hour fcst: - 1: 9:127513:d=2021032312:TMP:5 mb:25 hour fcst: - 1: 10:144204:d=2021032312:RH:5 mb:25 hour fcst: - 1: 11:153394:d=2021032312:UGRD:5 mb:25 hour fcst: - 1: 12:164991:d=2021032312:VGRD:5 mb:25 hour fcst: - 1: 13:183599:d=2021032312:ABSV:5 mb:25 hour fcst: - 1: 14:201284:d=2021032312:O3MR:5 mb:25 hour fcst: - 1: + interp_atmos_master.sh[47]export err=0 - 1: + interp_atmos_master.sh[47]err=0 - 1: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 1: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 1: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_2_0p25 - 1: + product_functions.sh[5]local filename=pgb2bfile_f025_2_0p25 - 1: + product_functions.sh[6]wgrib2 pgb2bfile_f025_2_0p25 -not_if :RH: -grib pgb2bfile_f025_2_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_2_0p25.new - 1: 1:0:d=2021032312:HGT:3 mb:25 hour fcst: - 1: 2:400173:d=2021032312:TMP:3 mb:25 hour fcst: - 1: 3:763009:d=2021032312:RH:3 mb:25 hour fcst: - 1: 4:931755:d=2021032312:UGRD:3 mb:25 hour fcst: - 1: 5:1382704:d=2021032312:VGRD:3 mb:25 hour fcst: - 1: 6:1793501:d=2021032312:ABSV:3 mb:25 hour fcst: - 1: 7:2158374:d=2021032312:O3MR:3 mb:25 hour fcst: - 1: 8:2760066:d=2021032312:HGT:5 mb:25 hour fcst: - 1: 9:3156753:d=2021032312:TMP:5 mb:25 hour fcst: - 1: 10:3523553:d=2021032312:RH:5 mb:25 hour fcst: - 1: 11:3775118:d=2021032312:UGRD:5 mb:25 hour fcst: - 1: 12:4226308:d=2021032312:VGRD:5 mb:25 hour fcst: - 1: 13:4653576:d=2021032312:ABSV:5 mb:25 hour fcst: - 1: 14:5039739:d=2021032312:O3MR:5 mb:25 hour fcst: - 1: + product_functions.sh[10]rc=0 - 1: + product_functions.sh[11](( rc == 0 )) - 1: + product_functions.sh[11]mv pgb2bfile_f025_2_0p25.new pgb2bfile_f025_2_0p25 - 1: + product_functions.sh[12]return 0 - 1: + interp_atmos_master.sh[56]export err=0 - 1: + interp_atmos_master.sh[56]err=0 - 1: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 1: ++ interp_atmos_master.sh[62]wc -l - 1: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_2_0p25 -match 'LAND|ICEC' - 1: + interp_atmos_master.sh[62]var_count=0 - 1: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 1: + interp_atmos_master.sh[73]exit 0 - 7: + bash[8]'[' -z '' ']' - 7: + bash[9]case "$-" in - 7: + bash[12]__lmod_vx=x - 7: + bash[16]'[' -n x ']' - 7: + bash[16]set +x - 7: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 7: Shell debugging restarted - 7: + bash[224]unset __lmod_vx - 7: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_8 - 7: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_8 - 7: + interp_atmos_master.sh[9]grid_string=0p25 - 7: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 7: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 7: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 7: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 7: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 7: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 7: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 7: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 7: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 7: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 7: + interp_atmos_master.sh[31]IFS=: - 7: + interp_atmos_master.sh[31]read -ra grids - 7: + interp_atmos_master.sh[33]output_grids= - 7: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 7: + interp_atmos_master.sh[35]gridopt=grid0p25 - 7: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_8_0p25' - 7: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_8 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_8_0p25 - 7: 1:0:d=2021032312:TMP:325 mb:25 hour fcst: - 7: 2:18354:d=2021032312:RH:325 mb:25 hour fcst: - 7: 3:37315:d=2021032312:TCDC:325 mb:25 hour fcst: - 7: 4:50472:d=2021032312:VVEL:325 mb:25 hour fcst: - 7: 5:74756:d=2021032312:DZDT:325 mb:25 hour fcst: - 7: 6:103099:d=2021032312:UGRD:325 mb:25 hour fcst: - 7: 7:117776:d=2021032312:VGRD:325 mb:25 hour fcst: - 7: 8:133107:d=2021032312:ABSV:325 mb:25 hour fcst: - 7: 9:155421:d=2021032312:CLMR:325 mb:25 hour fcst: - 7: 10:155600:d=2021032312:ICMR:325 mb:25 hour fcst: - 7: 11:172364:d=2021032312:RWMR:325 mb:25 hour fcst: - 7: 12:172543:d=2021032312:SNMR:325 mb:25 hour fcst: - 7: 13:184513:d=2021032312:GRLE:325 mb:25 hour fcst: - 7: 14:184947:d=2021032312:HGT:375 mb:25 hour fcst: - 7: + interp_atmos_master.sh[47]export err=0 - 7: + interp_atmos_master.sh[47]err=0 - 7: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 7: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 7: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_8_0p25 - 7: + product_functions.sh[5]local filename=pgb2bfile_f025_8_0p25 - 7: + product_functions.sh[6]wgrib2 pgb2bfile_f025_8_0p25 -not_if :RH: -grib pgb2bfile_f025_8_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_8_0p25.new - 7: 1:0:d=2021032312:TMP:325 mb:25 hour fcst: - 7: 2:407633:d=2021032312:RH:325 mb:25 hour fcst: - 7: 3:834165:d=2021032312:TCDC:325 mb:25 hour fcst: - 7: 4:1077497:d=2021032312:VVEL:325 mb:25 hour fcst: - 7: 5:1722983:d=2021032312:DZDT:325 mb:25 hour fcst: - 7: 6:2516173:d=2021032312:UGRD:325 mb:25 hour fcst: - 7: 7:2831152:d=2021032312:VGRD:325 mb:25 hour fcst: - 7: 8:3164070:d=2021032312:ABSV:325 mb:25 hour fcst: - 7: 9:3715487:d=2021032312:CLMR:325 mb:25 hour fcst: - 7: 10:3715666:d=2021032312:ICMR:325 mb:25 hour fcst: - 7: 11:4082060:d=2021032312:RWMR:325 mb:25 hour fcst: - 7: 12:4082239:d=2021032312:SNMR:325 mb:25 hour fcst: - 7: 13:4327008:d=2021032312:GRLE:325 mb:25 hour fcst: - 7: 14:4332787:d=2021032312:HGT:375 mb:25 hour fcst: - 7: + product_functions.sh[10]rc=0 - 7: + product_functions.sh[11](( rc == 0 )) - 7: + product_functions.sh[11]mv pgb2bfile_f025_8_0p25.new pgb2bfile_f025_8_0p25 - 7: + product_functions.sh[12]return 0 - 7: + interp_atmos_master.sh[56]export err=0 - 7: + interp_atmos_master.sh[56]err=0 - 7: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 7: ++ interp_atmos_master.sh[62]wc -l - 7: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_8_0p25 -match 'LAND|ICEC' - 7: + interp_atmos_master.sh[62]var_count=0 - 7: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 7: + interp_atmos_master.sh[73]exit 0 -21: + bash[8]'[' -z '' ']' -21: + bash[9]case "$-" in -21: + bash[12]__lmod_vx=x -21: + bash[16]'[' -n x ']' -21: + bash[16]set +x -21: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -21: Shell debugging restarted -21: + bash[224]unset __lmod_vx -21: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_22 -21: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_22 -21: + interp_atmos_master.sh[9]grid_string=0p25 -21: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -21: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -21: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -21: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -21: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -21: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -21: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -21: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -21: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -21: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -21: + interp_atmos_master.sh[31]IFS=: -21: + interp_atmos_master.sh[31]read -ra grids -21: + interp_atmos_master.sh[33]output_grids= -21: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -21: + interp_atmos_master.sh[35]gridopt=grid0p25 -21: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_22_0p25' -21: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_22 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_22_0p25 -21: 1:0:d=2021032312:RH:90-60 mb above ground:25 hour fcst: -21: 2:19079:d=2021032312:SPFH:90-60 mb above ground:25 hour fcst: -21: 3:48035:d=2021032312:UGRD:90-60 mb above ground:25 hour fcst: -21: 4:69438:d=2021032312:VGRD:90-60 mb above ground:25 hour fcst: -21: 5:91399:d=2021032312:TMP:120-90 mb above ground:25 hour fcst: -21: 6:112341:d=2021032312:RH:120-90 mb above ground:25 hour fcst: -21: 7:131689:d=2021032312:SPFH:120-90 mb above ground:25 hour fcst: -21: 8:160623:d=2021032312:UGRD:120-90 mb above ground:25 hour fcst: -21: 9:181801:d=2021032312:VGRD:120-90 mb above ground:25 hour fcst: -21: 10:203591:d=2021032312:TMP:150-120 mb above ground:25 hour fcst: -21: 11:224497:d=2021032312:RH:150-120 mb above ground:25 hour fcst: -21: 12:244201:d=2021032312:SPFH:150-120 mb above ground:25 hour fcst: -21: 13:273317:d=2021032312:UGRD:150-120 mb above ground:25 hour fcst: -21: 14:294339:d=2021032312:VGRD:150-120 mb above ground:25 hour fcst: -21: + interp_atmos_master.sh[47]export err=0 -21: + interp_atmos_master.sh[47]err=0 -21: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -21: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -21: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_22_0p25 -21: + product_functions.sh[5]local filename=pgb2bfile_f025_22_0p25 -21: + product_functions.sh[6]wgrib2 pgb2bfile_f025_22_0p25 -not_if :RH: -grib pgb2bfile_f025_22_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_22_0p25.new -21: 1:0:d=2021032312:RH:90-60 mb above ground:25 hour fcst: -21: 2:426163:d=2021032312:SPFH:90-60 mb above ground:25 hour fcst: -21: 3:1215486:d=2021032312:UGRD:90-60 mb above ground:25 hour fcst: -21: 4:1730107:d=2021032312:VGRD:90-60 mb above ground:25 hour fcst: -21: 5:2277245:d=2021032312:TMP:120-90 mb above ground:25 hour fcst: -21: 6:2776299:d=2021032312:RH:120-90 mb above ground:25 hour fcst: -21: 7:3210658:d=2021032312:SPFH:120-90 mb above ground:25 hour fcst: -21: 8:3999500:d=2021032312:UGRD:120-90 mb above ground:25 hour fcst: -21: 9:4507762:d=2021032312:VGRD:120-90 mb above ground:25 hour fcst: -21: 10:5047279:d=2021032312:TMP:150-120 mb above ground:25 hour fcst: -21: 11:5548766:d=2021032312:RH:150-120 mb above ground:25 hour fcst: -21: 12:5995903:d=2021032312:SPFH:150-120 mb above ground:25 hour fcst: -21: 13:6788551:d=2021032312:UGRD:150-120 mb above ground:25 hour fcst: -21: 14:7292418:d=2021032312:VGRD:150-120 mb above ground:25 hour fcst: -21: + product_functions.sh[10]rc=0 -21: + product_functions.sh[11](( rc == 0 )) -21: + product_functions.sh[11]mv pgb2bfile_f025_22_0p25.new pgb2bfile_f025_22_0p25 -21: + product_functions.sh[12]return 0 -21: + interp_atmos_master.sh[56]export err=0 -21: + interp_atmos_master.sh[56]err=0 -21: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -21: ++ interp_atmos_master.sh[62]wc -l -21: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_22_0p25 -match 'LAND|ICEC' -21: + interp_atmos_master.sh[62]var_count=0 -21: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -21: + interp_atmos_master.sh[73]exit 0 -12: + bash[8]'[' -z '' ']' -12: + bash[9]case "$-" in -12: + bash[12]__lmod_vx=x -12: + bash[16]'[' -n x ']' -12: + bash[16]set +x -12: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -12: Shell debugging restarted -12: + bash[224]unset __lmod_vx -12: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_13 -12: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_13 -12: + interp_atmos_master.sh[9]grid_string=0p25 -12: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -12: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -12: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -12: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -12: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -12: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -12: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -12: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -12: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -12: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -12: + interp_atmos_master.sh[31]IFS=: -12: + interp_atmos_master.sh[31]read -ra grids -12: + interp_atmos_master.sh[33]output_grids= -12: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -12: + interp_atmos_master.sh[35]gridopt=grid0p25 -12: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_13_0p25' -12: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_13 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_13_0p25 -12: 1:0:d=2021032312:TMP:575 mb:25 hour fcst: -12: 2:18878:d=2021032312:RH:575 mb:25 hour fcst: -12: 3:38796:d=2021032312:TCDC:575 mb:25 hour fcst: -12: 4:51030:d=2021032312:VVEL:575 mb:25 hour fcst: -12: 5:76254:d=2021032312:DZDT:575 mb:25 hour fcst: -12: 6:103673:d=2021032312:UGRD:575 mb:25 hour fcst: -12: 7:124573:d=2021032312:VGRD:575 mb:25 hour fcst: -12: 8:146250:d=2021032312:ABSV:575 mb:25 hour fcst: -12: 9:167106:d=2021032312:CLMR:575 mb:25 hour fcst: -12: 10:171287:d=2021032312:ICMR:575 mb:25 hour fcst: -12: 11:187340:d=2021032312:RWMR:575 mb:25 hour fcst: -12: 12:189722:d=2021032312:SNMR:575 mb:25 hour fcst: -12: 13:204307:d=2021032312:GRLE:575 mb:25 hour fcst: -12: 14:206752:d=2021032312:HGT:625 mb:25 hour fcst: -12: + interp_atmos_master.sh[47]export err=0 -12: + interp_atmos_master.sh[47]err=0 -12: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -12: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -12: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_13_0p25 -12: + product_functions.sh[5]local filename=pgb2bfile_f025_13_0p25 -12: + product_functions.sh[6]wgrib2 pgb2bfile_f025_13_0p25 -not_if :RH: -grib pgb2bfile_f025_13_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_13_0p25.new -12: 1:0:d=2021032312:TMP:575 mb:25 hour fcst: -12: 2:424136:d=2021032312:RH:575 mb:25 hour fcst: -12: 3:878740:d=2021032312:TCDC:575 mb:25 hour fcst: -12: 4:1107074:d=2021032312:VVEL:575 mb:25 hour fcst: -12: 5:1778354:d=2021032312:DZDT:575 mb:25 hour fcst: -12: 6:2546967:d=2021032312:UGRD:575 mb:25 hour fcst: -12: 7:3056574:d=2021032312:VGRD:575 mb:25 hour fcst: -12: 8:3594347:d=2021032312:ABSV:575 mb:25 hour fcst: -12: 9:4080562:d=2021032312:CLMR:575 mb:25 hour fcst: -12: 10:4167602:d=2021032312:ICMR:575 mb:25 hour fcst: -12: 11:4522287:d=2021032312:RWMR:575 mb:25 hour fcst: -12: 12:4566860:d=2021032312:SNMR:575 mb:25 hour fcst: -12: 13:4880586:d=2021032312:GRLE:575 mb:25 hour fcst: -12: 14:4926176:d=2021032312:HGT:625 mb:25 hour fcst: -12: + product_functions.sh[10]rc=0 -12: + product_functions.sh[11](( rc == 0 )) -12: + product_functions.sh[11]mv pgb2bfile_f025_13_0p25.new pgb2bfile_f025_13_0p25 -12: + product_functions.sh[12]return 0 -12: + interp_atmos_master.sh[56]export err=0 -12: + interp_atmos_master.sh[56]err=0 -12: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -12: ++ interp_atmos_master.sh[62]wc -l -12: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_13_0p25 -match 'LAND|ICEC' -12: + interp_atmos_master.sh[62]var_count=0 -12: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -12: + interp_atmos_master.sh[73]exit 0 -19: + bash[8]'[' -z '' ']' -19: + bash[9]case "$-" in -19: + bash[12]__lmod_vx=x -19: + bash[16]'[' -n x ']' -19: + bash[16]set +x -19: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -19: Shell debugging restarted -19: + bash[224]unset __lmod_vx -19: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_20 -19: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_20 -19: + interp_atmos_master.sh[9]grid_string=0p25 -19: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -19: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -19: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -19: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -19: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -19: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -19: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -19: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -19: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -19: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -19: + interp_atmos_master.sh[31]IFS=: -19: + interp_atmos_master.sh[31]read -ra grids -19: + interp_atmos_master.sh[33]output_grids= -19: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -19: + interp_atmos_master.sh[35]gridopt=grid0p25 -19: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_20_0p25' -19: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_20 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_20_0p25 -19: 1:0:d=2021032312:SOILL:0.1-0.4 m below ground:25 hour fcst: -19: 2:8894:d=2021032312:SOILL:0.4-1 m below ground:25 hour fcst: -19: 3:17813:d=2021032312:SOILL:1-2 m below ground:25 hour fcst: -19: 4:26891:d=2021032312:CNWAT:surface:25 hour fcst: -19: 5:32435:d=2021032312:ICETK:surface:25 hour fcst: -19: 6:35852:d=2021032312:DUVB:surface:24-25 hour ave fcst: -19: 7:52568:d=2021032312:CDUVB:surface:24-25 hour ave fcst: -19: 8:65734:d=2021032312:TMP:305 m above mean sea level:25 hour fcst: -19: 9:83935:d=2021032312:UGRD:305 m above mean sea level:25 hour fcst: -19: 10:103090:d=2021032312:VGRD:305 m above mean sea level:25 hour fcst: -19: 11:122727:d=2021032312:TMP:457 m above mean sea level:25 hour fcst: -19: 12:142252:d=2021032312:UGRD:457 m above mean sea level:25 hour fcst: -19: 13:162459:d=2021032312:VGRD:457 m above mean sea level:25 hour fcst: -19: 14:183010:d=2021032312:TMP:610 m above mean sea level:25 hour fcst: -19: + interp_atmos_master.sh[47]export err=0 -19: + interp_atmos_master.sh[47]err=0 -19: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -19: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -19: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_20_0p25 -19: + product_functions.sh[5]local filename=pgb2bfile_f025_20_0p25 -19: + product_functions.sh[6]wgrib2 pgb2bfile_f025_20_0p25 -not_if :RH: -grib pgb2bfile_f025_20_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_20_0p25.new -19: 1:0:d=2021032312:SOILL:0.1-0.4 m below ground:25 hour fcst: -19: 2:294071:d=2021032312:SOILL:0.4-1 m below ground:25 hour fcst: -19: 3:589665:d=2021032312:SOILL:1-2 m below ground:25 hour fcst: -19: 4:885684:d=2021032312:CNWAT:surface:25 hour fcst: -19: 5:1071820:d=2021032312:ICETK:surface:25 hour fcst: -19: 6:1125258:d=2021032312:DUVB:surface:24-25 hour ave fcst: -19: 7:1553634:d=2021032312:CDUVB:surface:24-25 hour ave fcst: -19: 8:1875236:d=2021032312:TMP:305 m above mean sea level:25 hour fcst: -19: 9:2387507:d=2021032312:UGRD:305 m above mean sea level:25 hour fcst: -19: 10:2947410:d=2021032312:VGRD:305 m above mean sea level:25 hour fcst: -19: 11:3523878:d=2021032312:TMP:457 m above mean sea level:25 hour fcst: -19: 12:4067395:d=2021032312:UGRD:457 m above mean sea level:25 hour fcst: -19: 13:4653061:d=2021032312:VGRD:457 m above mean sea level:25 hour fcst: -19: 14:5257788:d=2021032312:TMP:610 m above mean sea level:25 hour fcst: -19: + product_functions.sh[10]rc=0 -19: + product_functions.sh[11](( rc == 0 )) -19: + product_functions.sh[11]mv pgb2bfile_f025_20_0p25.new pgb2bfile_f025_20_0p25 -19: + product_functions.sh[12]return 0 -19: + interp_atmos_master.sh[56]export err=0 -19: + interp_atmos_master.sh[56]err=0 -19: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -19: ++ interp_atmos_master.sh[62]wc -l -19: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_20_0p25 -match 'LAND|ICEC' -19: + interp_atmos_master.sh[62]var_count=0 -19: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -19: + interp_atmos_master.sh[73]exit 0 - 8: + bash[8]'[' -z '' ']' - 8: + bash[9]case "$-" in - 8: + bash[12]__lmod_vx=x - 8: + bash[16]'[' -n x ']' - 8: + bash[16]set +x - 8: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 8: Shell debugging restarted - 8: + bash[224]unset __lmod_vx - 8: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_9 - 8: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_9 - 8: + interp_atmos_master.sh[9]grid_string=0p25 - 8: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 8: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 8: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 8: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 8: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 8: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 8: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 8: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 8: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 8: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 8: + interp_atmos_master.sh[31]IFS=: - 8: + interp_atmos_master.sh[31]read -ra grids - 8: + interp_atmos_master.sh[33]output_grids= - 8: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 8: + interp_atmos_master.sh[35]gridopt=grid0p25 - 8: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_9_0p25' - 8: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_9 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_9_0p25 - 8: 1:0:d=2021032312:TMP:375 mb:25 hour fcst: - 8: 2:18332:d=2021032312:RH:375 mb:25 hour fcst: - 8: 3:37451:d=2021032312:TCDC:375 mb:25 hour fcst: - 8: 4:50614:d=2021032312:VVEL:375 mb:25 hour fcst: - 8: 5:75445:d=2021032312:DZDT:375 mb:25 hour fcst: - 8: 6:103516:d=2021032312:UGRD:375 mb:25 hour fcst: - 8: 7:117963:d=2021032312:VGRD:375 mb:25 hour fcst: - 8: 8:133071:d=2021032312:ABSV:375 mb:25 hour fcst: - 8: 9:155237:d=2021032312:CLMR:375 mb:25 hour fcst: - 8: 10:155416:d=2021032312:ICMR:375 mb:25 hour fcst: - 8: 11:172516:d=2021032312:RWMR:375 mb:25 hour fcst: - 8: 12:172695:d=2021032312:SNMR:375 mb:25 hour fcst: - 8: 13:185863:d=2021032312:GRLE:375 mb:25 hour fcst: - 8: 14:186496:d=2021032312:HGT:425 mb:25 hour fcst: - 8: + interp_atmos_master.sh[47]export err=0 - 8: + interp_atmos_master.sh[47]err=0 - 8: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 8: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 8: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_9_0p25 - 8: + product_functions.sh[5]local filename=pgb2bfile_f025_9_0p25 - 8: + product_functions.sh[6]wgrib2 pgb2bfile_f025_9_0p25 -not_if :RH: -grib pgb2bfile_f025_9_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_9_0p25.new - 8: 1:0:d=2021032312:TMP:375 mb:25 hour fcst: - 8: 2:410019:d=2021032312:RH:375 mb:25 hour fcst: - 8: 3:844085:d=2021032312:TCDC:375 mb:25 hour fcst: - 8: 4:1083686:d=2021032312:VVEL:375 mb:25 hour fcst: - 8: 5:1742288:d=2021032312:DZDT:375 mb:25 hour fcst: - 8: 6:2536406:d=2021032312:UGRD:375 mb:25 hour fcst: - 8: 7:2845923:d=2021032312:VGRD:375 mb:25 hour fcst: - 8: 8:3174138:d=2021032312:ABSV:375 mb:25 hour fcst: - 8: 9:3717684:d=2021032312:CLMR:375 mb:25 hour fcst: - 8: 10:3717863:d=2021032312:ICMR:375 mb:25 hour fcst: - 8: 11:4093454:d=2021032312:RWMR:375 mb:25 hour fcst: - 8: 12:4093633:d=2021032312:SNMR:375 mb:25 hour fcst: - 8: 13:4367355:d=2021032312:GRLE:375 mb:25 hour fcst: - 8: 14:4375566:d=2021032312:HGT:425 mb:25 hour fcst: - 8: + product_functions.sh[10]rc=0 - 8: + product_functions.sh[11](( rc == 0 )) - 8: + product_functions.sh[11]mv pgb2bfile_f025_9_0p25.new pgb2bfile_f025_9_0p25 - 8: + product_functions.sh[12]return 0 - 8: + interp_atmos_master.sh[56]export err=0 - 8: + interp_atmos_master.sh[56]err=0 - 8: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 8: ++ interp_atmos_master.sh[62]wc -l - 8: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_9_0p25 -match 'LAND|ICEC' - 8: + interp_atmos_master.sh[62]var_count=0 - 8: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 8: + interp_atmos_master.sh[73]exit 0 -16: + bash[8]'[' -z '' ']' -16: + bash[9]case "$-" in -16: + bash[12]__lmod_vx=x -16: + bash[16]'[' -n x ']' -16: + bash[16]set +x -16: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -16: Shell debugging restarted -16: + bash[224]unset __lmod_vx -16: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_17 -16: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_17 -16: + interp_atmos_master.sh[9]grid_string=0p25 -16: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -16: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -16: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -16: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -16: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -16: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -16: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -16: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -16: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -16: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -16: + interp_atmos_master.sh[31]IFS=: -16: + interp_atmos_master.sh[31]read -ra grids -16: + interp_atmos_master.sh[33]output_grids= -16: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -16: + interp_atmos_master.sh[35]gridopt=grid0p25 -16: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_17_0p25' -16: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_17 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_17_0p25 -16: 1:0:d=2021032312:TMP:775 mb:25 hour fcst: -16: 2:20134:d=2021032312:RH:775 mb:25 hour fcst: -16: 3:40603:d=2021032312:TCDC:775 mb:25 hour fcst: -16: 4:53362:d=2021032312:VVEL:775 mb:25 hour fcst: -16: 5:79105:d=2021032312:DZDT:775 mb:25 hour fcst: -16: 6:106096:d=2021032312:UGRD:775 mb:25 hour fcst: -16: 7:126951:d=2021032312:VGRD:775 mb:25 hour fcst: -16: 8:148675:d=2021032312:ABSV:775 mb:25 hour fcst: -16: 9:169864:d=2021032312:CLMR:775 mb:25 hour fcst: -16: 10:180226:d=2021032312:ICMR:775 mb:25 hour fcst: -16: 11:191286:d=2021032312:RWMR:775 mb:25 hour fcst: -16: 12:198001:d=2021032312:SNMR:775 mb:25 hour fcst: -16: 13:211133:d=2021032312:GRLE:775 mb:25 hour fcst: -16: 14:213412:d=2021032312:HGT:825 mb:25 hour fcst: -16: + interp_atmos_master.sh[47]export err=0 -16: + interp_atmos_master.sh[47]err=0 -16: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -16: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -16: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_17_0p25 -16: + product_functions.sh[5]local filename=pgb2bfile_f025_17_0p25 -16: + product_functions.sh[6]wgrib2 pgb2bfile_f025_17_0p25 -not_if :RH: -grib pgb2bfile_f025_17_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_17_0p25.new -16: 1:0:d=2021032312:TMP:775 mb:25 hour fcst: -16: 2:469743:d=2021032312:RH:775 mb:25 hour fcst: -16: 3:951284:d=2021032312:TCDC:775 mb:25 hour fcst: -16: 4:1206153:d=2021032312:VVEL:775 mb:25 hour fcst: -16: 5:1891030:d=2021032312:DZDT:775 mb:25 hour fcst: -16: 6:2638029:d=2021032312:UGRD:775 mb:25 hour fcst: -16: 7:3143312:d=2021032312:VGRD:775 mb:25 hour fcst: -16: 8:3679171:d=2021032312:ABSV:775 mb:25 hour fcst: -16: 9:4183322:d=2021032312:CLMR:775 mb:25 hour fcst: -16: 10:4409335:d=2021032312:ICMR:775 mb:25 hour fcst: -16: 11:4671107:d=2021032312:RWMR:775 mb:25 hour fcst: -16: 12:4796437:d=2021032312:SNMR:775 mb:25 hour fcst: -16: 13:5091626:d=2021032312:GRLE:775 mb:25 hour fcst: -16: 14:5132223:d=2021032312:HGT:825 mb:25 hour fcst: -16: + product_functions.sh[10]rc=0 -16: + product_functions.sh[11](( rc == 0 )) -16: + product_functions.sh[11]mv pgb2bfile_f025_17_0p25.new pgb2bfile_f025_17_0p25 -16: + product_functions.sh[12]return 0 -16: + interp_atmos_master.sh[56]export err=0 -16: + interp_atmos_master.sh[56]err=0 -16: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -16: ++ interp_atmos_master.sh[62]wc -l -16: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_17_0p25 -match 'LAND|ICEC' -16: + interp_atmos_master.sh[62]var_count=0 -16: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -16: + interp_atmos_master.sh[73]exit 0 -18: + bash[8]'[' -z '' ']' -18: + bash[9]case "$-" in -18: + bash[12]__lmod_vx=x -18: + bash[16]'[' -n x ']' -18: + bash[16]set +x -18: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -18: Shell debugging restarted -18: + bash[224]unset __lmod_vx -18: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_19 -18: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_19 -18: + interp_atmos_master.sh[9]grid_string=0p25 -18: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -18: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -18: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -18: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -18: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -18: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -18: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -18: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -18: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -18: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -18: + interp_atmos_master.sh[31]IFS=: -18: + interp_atmos_master.sh[31]read -ra grids -18: + interp_atmos_master.sh[33]output_grids= -18: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -18: + interp_atmos_master.sh[35]gridopt=grid0p25 -18: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_19_0p25' -18: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_19 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_19_0p25 -18: 1:0:d=2021032312:TMP:875 mb:25 hour fcst: -18: 2:20821:d=2021032312:RH:875 mb:25 hour fcst: -18: 3:40801:d=2021032312:TCDC:875 mb:25 hour fcst: -18: 4:55138:d=2021032312:VVEL:875 mb:25 hour fcst: -18: 5:80473:d=2021032312:DZDT:875 mb:25 hour fcst: -18: 6:106883:d=2021032312:UGRD:875 mb:25 hour fcst: -18: 7:128093:d=2021032312:VGRD:875 mb:25 hour fcst: -18: 8:149944:d=2021032312:ABSV:875 mb:25 hour fcst: -18: 9:171181:d=2021032312:CLMR:875 mb:25 hour fcst: -18: 10:183484:d=2021032312:ICMR:875 mb:25 hour fcst: -18: 11:193394:d=2021032312:RWMR:875 mb:25 hour fcst: -18: 12:204917:d=2021032312:SNMR:875 mb:25 hour fcst: -18: 13:215884:d=2021032312:GRLE:875 mb:25 hour fcst: -18: 14:218947:d=2021032312:SOILL:0-0.1 m below ground:25 hour fcst: -18: + interp_atmos_master.sh[47]export err=0 -18: + interp_atmos_master.sh[47]err=0 -18: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -18: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -18: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_19_0p25 -18: + product_functions.sh[5]local filename=pgb2bfile_f025_19_0p25 -18: + product_functions.sh[6]wgrib2 pgb2bfile_f025_19_0p25 -not_if :RH: -grib pgb2bfile_f025_19_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_19_0p25.new -18: 1:0:d=2021032312:TMP:875 mb:25 hour fcst: -18: 2:493145:d=2021032312:RH:875 mb:25 hour fcst: -18: 3:960087:d=2021032312:TCDC:875 mb:25 hour fcst: -18: 4:1251812:d=2021032312:VVEL:875 mb:25 hour fcst: -18: 5:1925258:d=2021032312:DZDT:875 mb:25 hour fcst: -18: 6:2649583:d=2021032312:UGRD:875 mb:25 hour fcst: -18: 7:3162706:d=2021032312:VGRD:875 mb:25 hour fcst: -18: 8:3707042:d=2021032312:ABSV:875 mb:25 hour fcst: -18: 9:4218599:d=2021032312:CLMR:875 mb:25 hour fcst: -18: 10:4490102:d=2021032312:ICMR:875 mb:25 hour fcst: -18: 11:4731195:d=2021032312:RWMR:875 mb:25 hour fcst: -18: 12:4958349:d=2021032312:SNMR:875 mb:25 hour fcst: -18: 13:5208401:d=2021032312:GRLE:875 mb:25 hour fcst: -18: 14:5267248:d=2021032312:SOILL:0-0.1 m below ground:25 hour fcst: -18: + product_functions.sh[10]rc=0 -18: + product_functions.sh[11](( rc == 0 )) -18: + product_functions.sh[11]mv pgb2bfile_f025_19_0p25.new pgb2bfile_f025_19_0p25 -18: + product_functions.sh[12]return 0 -18: + interp_atmos_master.sh[56]export err=0 -18: + interp_atmos_master.sh[56]err=0 -18: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -18: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_19_0p25 -match 'LAND|ICEC' -18: ++ interp_atmos_master.sh[62]wc -l -18: + interp_atmos_master.sh[62]var_count=0 -18: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -18: + interp_atmos_master.sh[73]exit 0 -13: + bash[8]'[' -z '' ']' -13: + bash[9]case "$-" in -13: + bash[12]__lmod_vx=x -13: + bash[16]'[' -n x ']' -13: + bash[16]set +x -13: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -13: Shell debugging restarted -13: + bash[224]unset __lmod_vx -13: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_14 -13: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_14 -13: + interp_atmos_master.sh[9]grid_string=0p25 -13: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -13: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -13: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -13: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -13: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -13: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -13: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -13: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -13: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -13: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -13: + interp_atmos_master.sh[31]IFS=: -13: + interp_atmos_master.sh[31]read -ra grids -13: + interp_atmos_master.sh[33]output_grids= -13: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -13: + interp_atmos_master.sh[35]gridopt=grid0p25 -13: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_14_0p25' -13: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_14 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_14_0p25 -13: 1:0:d=2021032312:TMP:625 mb:25 hour fcst: -13: 2:19134:d=2021032312:RH:625 mb:25 hour fcst: -13: 3:39257:d=2021032312:TCDC:625 mb:25 hour fcst: -13: 4:51076:d=2021032312:VVEL:625 mb:25 hour fcst: -13: 5:76564:d=2021032312:DZDT:625 mb:25 hour fcst: -13: 6:103985:d=2021032312:UGRD:625 mb:25 hour fcst: -13: 7:124855:d=2021032312:VGRD:625 mb:25 hour fcst: -13: 8:146489:d=2021032312:ABSV:625 mb:25 hour fcst: -13: 9:167280:d=2021032312:CLMR:625 mb:25 hour fcst: -13: 10:172212:d=2021032312:ICMR:625 mb:25 hour fcst: -13: 11:185304:d=2021032312:RWMR:625 mb:25 hour fcst: -13: 12:189535:d=2021032312:SNMR:625 mb:25 hour fcst: -13: 13:204868:d=2021032312:GRLE:625 mb:25 hour fcst: -13: 14:207329:d=2021032312:HGT:675 mb:25 hour fcst: -13: + interp_atmos_master.sh[47]export err=0 -13: + interp_atmos_master.sh[47]err=0 -13: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -13: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -13: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_14_0p25 -13: + product_functions.sh[5]local filename=pgb2bfile_f025_14_0p25 -13: + product_functions.sh[6]wgrib2 pgb2bfile_f025_14_0p25 -not_if :RH: -grib pgb2bfile_f025_14_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_14_0p25.new -13: 1:0:d=2021032312:TMP:625 mb:25 hour fcst: -13: 2:427349:d=2021032312:RH:625 mb:25 hour fcst: -13: 3:896627:d=2021032312:TCDC:625 mb:25 hour fcst: -13: 4:1118848:d=2021032312:VVEL:625 mb:25 hour fcst: -13: 5:1793376:d=2021032312:DZDT:625 mb:25 hour fcst: -13: 6:2558506:d=2021032312:UGRD:625 mb:25 hour fcst: -13: 7:3064340:d=2021032312:VGRD:625 mb:25 hour fcst: -13: 8:3599629:d=2021032312:ABSV:625 mb:25 hour fcst: -13: 9:4083367:d=2021032312:CLMR:625 mb:25 hour fcst: -13: 10:4182359:d=2021032312:ICMR:625 mb:25 hour fcst: -13: 11:4470892:d=2021032312:RWMR:625 mb:25 hour fcst: -13: 12:4557461:d=2021032312:SNMR:625 mb:25 hour fcst: -13: 13:4893842:d=2021032312:GRLE:625 mb:25 hour fcst: -13: 14:4936835:d=2021032312:HGT:675 mb:25 hour fcst: -13: + product_functions.sh[10]rc=0 -13: + product_functions.sh[11](( rc == 0 )) -13: + product_functions.sh[11]mv pgb2bfile_f025_14_0p25.new pgb2bfile_f025_14_0p25 -13: + product_functions.sh[12]return 0 -13: + interp_atmos_master.sh[56]export err=0 -13: + interp_atmos_master.sh[56]err=0 -13: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -13: ++ interp_atmos_master.sh[62]wc -l -13: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_14_0p25 -match 'LAND|ICEC' -13: + interp_atmos_master.sh[62]var_count=0 -13: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -13: + interp_atmos_master.sh[73]exit 0 -10: + bash[8]'[' -z '' ']' -10: + bash[9]case "$-" in -10: + bash[12]__lmod_vx=x -10: + bash[16]'[' -n x ']' -10: + bash[16]set +x -10: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -10: Shell debugging restarted -10: + bash[224]unset __lmod_vx -10: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_11 -10: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_11 -10: + interp_atmos_master.sh[9]grid_string=0p25 -10: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -10: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -10: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -10: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -10: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -10: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -10: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -10: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -10: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -10: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -10: + interp_atmos_master.sh[31]IFS=: -10: + interp_atmos_master.sh[31]read -ra grids -10: + interp_atmos_master.sh[33]output_grids= -10: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -10: + interp_atmos_master.sh[35]gridopt=grid0p25 -10: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_11_0p25' -10: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_11 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_11_0p25 -10: 1:0:d=2021032312:TMP:475 mb:25 hour fcst: -10: 2:18645:d=2021032312:RH:475 mb:25 hour fcst: -10: 3:38151:d=2021032312:TCDC:475 mb:25 hour fcst: -10: 4:50563:d=2021032312:VVEL:475 mb:25 hour fcst: -10: 5:75580:d=2021032312:DZDT:475 mb:25 hour fcst: -10: 6:103407:d=2021032312:UGRD:475 mb:25 hour fcst: -10: 7:124973:d=2021032312:VGRD:475 mb:25 hour fcst: -10: 8:147172:d=2021032312:ABSV:475 mb:25 hour fcst: -10: 9:168437:d=2021032312:CLMR:475 mb:25 hour fcst: -10: 10:169696:d=2021032312:ICMR:475 mb:25 hour fcst: -10: 11:185982:d=2021032312:RWMR:475 mb:25 hour fcst: -10: 12:186696:d=2021032312:SNMR:475 mb:25 hour fcst: -10: 13:200721:d=2021032312:GRLE:475 mb:25 hour fcst: -10: 14:202234:d=2021032312:HGT:525 mb:25 hour fcst: -10: + interp_atmos_master.sh[47]export err=0 -10: + interp_atmos_master.sh[47]err=0 -10: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -10: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -10: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_11_0p25 -10: + product_functions.sh[5]local filename=pgb2bfile_f025_11_0p25 -10: + product_functions.sh[6]wgrib2 pgb2bfile_f025_11_0p25 -not_if :RH: -grib pgb2bfile_f025_11_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_11_0p25.new -10: 1:0:d=2021032312:TMP:475 mb:25 hour fcst: -10: 2:422205:d=2021032312:RH:475 mb:25 hour fcst: -10: 3:872422:d=2021032312:TCDC:475 mb:25 hour fcst: -10: 4:1106731:d=2021032312:VVEL:475 mb:25 hour fcst: -10: 5:1774800:d=2021032312:DZDT:475 mb:25 hour fcst: -10: 6:2557577:d=2021032312:UGRD:475 mb:25 hour fcst: -10: 7:3086684:d=2021032312:VGRD:475 mb:25 hour fcst: -10: 8:3644135:d=2021032312:ABSV:475 mb:25 hour fcst: -10: 9:4162246:d=2021032312:CLMR:475 mb:25 hour fcst: -10: 10:4186725:d=2021032312:ICMR:475 mb:25 hour fcst: -10: 11:4548518:d=2021032312:RWMR:475 mb:25 hour fcst: -10: 12:4558822:d=2021032312:SNMR:475 mb:25 hour fcst: -10: 13:4859036:d=2021032312:GRLE:475 mb:25 hour fcst: -10: 14:4880212:d=2021032312:HGT:525 mb:25 hour fcst: -10: + product_functions.sh[10]rc=0 -10: + product_functions.sh[11](( rc == 0 )) -10: + product_functions.sh[11]mv pgb2bfile_f025_11_0p25.new pgb2bfile_f025_11_0p25 -10: + product_functions.sh[12]return 0 -10: + interp_atmos_master.sh[56]export err=0 -10: + interp_atmos_master.sh[56]err=0 -10: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -10: ++ interp_atmos_master.sh[62]wc -l -10: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_11_0p25 -match 'LAND|ICEC' -10: + interp_atmos_master.sh[62]var_count=0 -10: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -10: + interp_atmos_master.sh[73]exit 0 -20: + bash[8]'[' -z '' ']' -20: + bash[9]case "$-" in -20: + bash[12]__lmod_vx=x -20: + bash[16]'[' -n x ']' -20: + bash[16]set +x -20: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -20: Shell debugging restarted -20: + bash[224]unset __lmod_vx -20: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_21 -20: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_21 -20: + interp_atmos_master.sh[9]grid_string=0p25 -20: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -20: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -20: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -20: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -20: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -20: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -20: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -20: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -20: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -20: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -20: + interp_atmos_master.sh[31]IFS=: -20: + interp_atmos_master.sh[31]read -ra grids -20: + interp_atmos_master.sh[33]output_grids= -20: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -20: + interp_atmos_master.sh[35]gridopt=grid0p25 -20: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_21_0p25' -20: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_21 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_21_0p25 -20: 1:0:d=2021032312:UGRD:610 m above mean sea level:25 hour fcst: -20: 2:20836:d=2021032312:VGRD:610 m above mean sea level:25 hour fcst: -20: 3:41986:d=2021032312:TMP:914 m above mean sea level:25 hour fcst: -20: 4:62836:d=2021032312:UGRD:914 m above mean sea level:25 hour fcst: -20: 5:84229:d=2021032312:VGRD:914 m above mean sea level:25 hour fcst: -20: 6:106064:d=2021032312:TMP:4572 m above mean sea level:25 hour fcst: -20: 7:127214:d=2021032312:UGRD:4572 m above mean sea level:25 hour fcst: -20: 8:150435:d=2021032312:VGRD:4572 m above mean sea level:25 hour fcst: -20: 9:174392:d=2021032312:TMP:60-30 mb above ground:25 hour fcst: -20: 10:195198:d=2021032312:RH:60-30 mb above ground:25 hour fcst: -20: 11:213870:d=2021032312:SPFH:60-30 mb above ground:25 hour fcst: -20: 12:242611:d=2021032312:UGRD:60-30 mb above ground:25 hour fcst: -20: 13:264194:d=2021032312:VGRD:60-30 mb above ground:25 hour fcst: -20: 14:286328:d=2021032312:TMP:90-60 mb above ground:25 hour fcst: -20: + interp_atmos_master.sh[47]export err=0 -20: + interp_atmos_master.sh[47]err=0 -20: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -20: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -20: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_21_0p25 -20: + product_functions.sh[5]local filename=pgb2bfile_f025_21_0p25 -20: + product_functions.sh[6]wgrib2 pgb2bfile_f025_21_0p25 -not_if :RH: -grib pgb2bfile_f025_21_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_21_0p25.new -20: 1:0:d=2021032312:UGRD:610 m above mean sea level:25 hour fcst: -20: 2:591884:d=2021032312:VGRD:610 m above mean sea level:25 hour fcst: -20: 3:1209768:d=2021032312:TMP:914 m above mean sea level:25 hour fcst: -20: 4:1795113:d=2021032312:UGRD:914 m above mean sea level:25 hour fcst: -20: 5:2397736:d=2021032312:VGRD:914 m above mean sea level:25 hour fcst: -20: 6:3028747:d=2021032312:TMP:4572 m above mean sea level:25 hour fcst: -20: 7:3602186:d=2021032312:UGRD:4572 m above mean sea level:25 hour fcst: -20: 8:4248501:d=2021032312:VGRD:4572 m above mean sea level:25 hour fcst: -20: 9:4922641:d=2021032312:TMP:60-30 mb above ground:25 hour fcst: -20: 10:5408733:d=2021032312:RH:60-30 mb above ground:25 hour fcst: -20: 11:5819612:d=2021032312:SPFH:60-30 mb above ground:25 hour fcst: -20: 12:6604522:d=2021032312:UGRD:60-30 mb above ground:25 hour fcst: -20: 13:7126750:d=2021032312:VGRD:60-30 mb above ground:25 hour fcst: -20: 14:7678430:d=2021032312:TMP:90-60 mb above ground:25 hour fcst: -20: + product_functions.sh[10]rc=0 -20: + product_functions.sh[11](( rc == 0 )) -20: + product_functions.sh[11]mv pgb2bfile_f025_21_0p25.new pgb2bfile_f025_21_0p25 -20: + product_functions.sh[12]return 0 -20: + interp_atmos_master.sh[56]export err=0 -20: + interp_atmos_master.sh[56]err=0 -20: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -20: ++ interp_atmos_master.sh[62]wc -l -20: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_21_0p25 -match 'LAND|ICEC' -20: + interp_atmos_master.sh[62]var_count=0 -20: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -20: + interp_atmos_master.sh[73]exit 0 -14: + bash[8]'[' -z '' ']' -14: + bash[9]case "$-" in -14: + bash[12]__lmod_vx=x -14: + bash[16]'[' -n x ']' -14: + bash[16]set +x -14: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -14: Shell debugging restarted -14: + bash[224]unset __lmod_vx -14: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_15 -14: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_15 -14: + interp_atmos_master.sh[9]grid_string=0p25 -14: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -14: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -14: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -14: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -14: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -14: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -14: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -14: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -14: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -14: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -14: + interp_atmos_master.sh[31]IFS=: -14: + interp_atmos_master.sh[31]read -ra grids -14: + interp_atmos_master.sh[33]output_grids= -14: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -14: + interp_atmos_master.sh[35]gridopt=grid0p25 -14: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_15_0p25' -14: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_15 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_15_0p25 -14: 1:0:d=2021032312:TMP:675 mb:25 hour fcst: -14: 2:19445:d=2021032312:RH:675 mb:25 hour fcst: -14: 3:39688:d=2021032312:TCDC:675 mb:25 hour fcst: -14: 4:51351:d=2021032312:VVEL:675 mb:25 hour fcst: -14: 5:77043:d=2021032312:DZDT:675 mb:25 hour fcst: -14: 6:104434:d=2021032312:UGRD:675 mb:25 hour fcst: -14: 7:125486:d=2021032312:VGRD:675 mb:25 hour fcst: -14: 8:146993:d=2021032312:ABSV:675 mb:25 hour fcst: -14: 9:168037:d=2021032312:CLMR:675 mb:25 hour fcst: -14: 10:173575:d=2021032312:ICMR:675 mb:25 hour fcst: -14: 11:185437:d=2021032312:RWMR:675 mb:25 hour fcst: -14: 12:190485:d=2021032312:SNMR:675 mb:25 hour fcst: -14: 13:205062:d=2021032312:GRLE:675 mb:25 hour fcst: -14: 14:207207:d=2021032312:HGT:725 mb:25 hour fcst: -14: + interp_atmos_master.sh[47]export err=0 -14: + interp_atmos_master.sh[47]err=0 -14: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -14: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -14: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_15_0p25 -14: + product_functions.sh[5]local filename=pgb2bfile_f025_15_0p25 -14: + product_functions.sh[6]wgrib2 pgb2bfile_f025_15_0p25 -not_if :RH: -grib pgb2bfile_f025_15_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_15_0p25.new -14: 1:0:d=2021032312:TMP:675 mb:25 hour fcst: -14: 2:439958:d=2021032312:RH:675 mb:25 hour fcst: -14: 3:912460:d=2021032312:TCDC:675 mb:25 hour fcst: -14: 4:1144319:d=2021032312:VVEL:675 mb:25 hour fcst: -14: 5:1824366:d=2021032312:DZDT:675 mb:25 hour fcst: -14: 6:2579024:d=2021032312:UGRD:675 mb:25 hour fcst: -14: 7:3082863:d=2021032312:VGRD:675 mb:25 hour fcst: -14: 8:3616808:d=2021032312:ABSV:675 mb:25 hour fcst: -14: 9:4108948:d=2021032312:CLMR:675 mb:25 hour fcst: -14: 10:4225919:d=2021032312:ICMR:675 mb:25 hour fcst: -14: 11:4496188:d=2021032312:RWMR:675 mb:25 hour fcst: -14: 12:4596207:d=2021032312:SNMR:675 mb:25 hour fcst: -14: 13:4915729:d=2021032312:GRLE:675 mb:25 hour fcst: -14: 14:4950681:d=2021032312:HGT:725 mb:25 hour fcst: -14: + product_functions.sh[10]rc=0 -14: + product_functions.sh[11](( rc == 0 )) -14: + product_functions.sh[11]mv pgb2bfile_f025_15_0p25.new pgb2bfile_f025_15_0p25 -14: + product_functions.sh[12]return 0 -14: + interp_atmos_master.sh[56]export err=0 -14: + interp_atmos_master.sh[56]err=0 -14: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -14: ++ interp_atmos_master.sh[62]wc -l -14: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_15_0p25 -match 'LAND|ICEC' -14: + interp_atmos_master.sh[62]var_count=0 -14: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -14: + interp_atmos_master.sh[73]exit 0 - 5: + bash[8]'[' -z '' ']' - 5: + bash[9]case "$-" in - 5: + bash[12]__lmod_vx=x - 5: + bash[16]'[' -n x ']' - 5: + bash[16]set +x - 5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 5: Shell debugging restarted - 5: + bash[224]unset __lmod_vx - 5: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_6 - 5: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_6 - 5: + interp_atmos_master.sh[9]grid_string=0p25 - 5: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 5: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 5: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 5: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 5: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 5: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 5: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 5: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 5: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 5: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 5: + interp_atmos_master.sh[31]IFS=: - 5: + interp_atmos_master.sh[31]read -ra grids - 5: + interp_atmos_master.sh[33]output_grids= - 5: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 5: + interp_atmos_master.sh[35]gridopt=grid0p25 - 5: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_6_0p25' - 5: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_6 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_6_0p25 - 5: 1:0:d=2021032312:TMP:225 mb:25 hour fcst: - 5: 2:18599:d=2021032312:RH:225 mb:25 hour fcst: - 5: 3:35246:d=2021032312:TCDC:225 mb:25 hour fcst: - 5: 4:44059:d=2021032312:VVEL:225 mb:25 hour fcst: - 5: 5:71670:d=2021032312:DZDT:225 mb:25 hour fcst: - 5: 6:98420:d=2021032312:UGRD:225 mb:25 hour fcst: - 5: 7:112382:d=2021032312:VGRD:225 mb:25 hour fcst: - 5: 8:126755:d=2021032312:ABSV:225 mb:25 hour fcst: - 5: 9:147685:d=2021032312:CLMR:225 mb:25 hour fcst: - 5: 10:147864:d=2021032312:ICMR:225 mb:25 hour fcst: - 5: 11:159372:d=2021032312:RWMR:225 mb:25 hour fcst: - 5: 12:159551:d=2021032312:SNMR:225 mb:25 hour fcst: - 5: 13:166365:d=2021032312:GRLE:225 mb:25 hour fcst: - 5: 14:166638:d=2021032312:HGT:275 mb:25 hour fcst: - 5: + interp_atmos_master.sh[47]export err=0 - 5: + interp_atmos_master.sh[47]err=0 - 5: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 5: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 5: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_6_0p25 - 5: + product_functions.sh[5]local filename=pgb2bfile_f025_6_0p25 - 5: + product_functions.sh[6]wgrib2 pgb2bfile_f025_6_0p25 -not_if :RH: -grib pgb2bfile_f025_6_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_6_0p25.new - 5: 1:0:d=2021032312:TMP:225 mb:25 hour fcst: - 5: 2:419948:d=2021032312:RH:225 mb:25 hour fcst: - 5: 3:805169:d=2021032312:TCDC:225 mb:25 hour fcst: - 5: 4:982010:d=2021032312:VVEL:225 mb:25 hour fcst: - 5: 5:1743080:d=2021032312:DZDT:225 mb:25 hour fcst: - 5: 6:2493686:d=2021032312:UGRD:225 mb:25 hour fcst: - 5: 7:2804221:d=2021032312:VGRD:225 mb:25 hour fcst: - 5: 8:3124704:d=2021032312:ABSV:225 mb:25 hour fcst: - 5: 9:3628225:d=2021032312:CLMR:225 mb:25 hour fcst: - 5: 10:3628404:d=2021032312:ICMR:225 mb:25 hour fcst: - 5: 11:3876351:d=2021032312:RWMR:225 mb:25 hour fcst: - 5: 12:3876530:d=2021032312:SNMR:225 mb:25 hour fcst: - 5: 13:4006695:d=2021032312:GRLE:225 mb:25 hour fcst: - 5: 14:4008027:d=2021032312:HGT:275 mb:25 hour fcst: - 5: + product_functions.sh[10]rc=0 - 5: + product_functions.sh[11](( rc == 0 )) - 5: + product_functions.sh[11]mv pgb2bfile_f025_6_0p25.new pgb2bfile_f025_6_0p25 - 5: + product_functions.sh[12]return 0 - 5: + interp_atmos_master.sh[56]export err=0 - 5: + interp_atmos_master.sh[56]err=0 - 5: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 5: ++ interp_atmos_master.sh[62]wc -l - 5: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_6_0p25 -match 'LAND|ICEC' - 5: + interp_atmos_master.sh[62]var_count=0 - 5: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 5: + interp_atmos_master.sh[73]exit 0 -11: + bash[8]'[' -z '' ']' -11: + bash[9]case "$-" in -11: + bash[12]__lmod_vx=x -11: + bash[16]'[' -n x ']' -11: + bash[16]set +x -11: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -11: Shell debugging restarted -11: + bash[224]unset __lmod_vx -11: + interp_atmos_master.sh[7]input_file=tmpfileb_f025_12 -11: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f025_12 -11: + interp_atmos_master.sh[9]grid_string=0p25 -11: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -11: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -11: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -11: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -11: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -11: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -11: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -11: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -11: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -11: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -11: + interp_atmos_master.sh[31]IFS=: -11: + interp_atmos_master.sh[31]read -ra grids -11: + interp_atmos_master.sh[33]output_grids= -11: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -11: + interp_atmos_master.sh[35]gridopt=grid0p25 -11: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_12_0p25' -11: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f025_12 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f025_12_0p25 -11: 1:0:d=2021032312:TMP:525 mb:25 hour fcst: -11: 2:18821:d=2021032312:RH:525 mb:25 hour fcst: -11: 3:38498:d=2021032312:TCDC:525 mb:25 hour fcst: -11: 4:50716:d=2021032312:VVEL:525 mb:25 hour fcst: -11: 5:75807:d=2021032312:DZDT:525 mb:25 hour fcst: -11: 6:103404:d=2021032312:UGRD:525 mb:25 hour fcst: -11: 7:124545:d=2021032312:VGRD:525 mb:25 hour fcst: -11: 8:146441:d=2021032312:ABSV:525 mb:25 hour fcst: -11: 9:167453:d=2021032312:CLMR:525 mb:25 hour fcst: -11: 10:169539:d=2021032312:ICMR:525 mb:25 hour fcst: -11: 11:186566:d=2021032312:RWMR:525 mb:25 hour fcst: -11: 12:187668:d=2021032312:SNMR:525 mb:25 hour fcst: -11: 13:202107:d=2021032312:GRLE:525 mb:25 hour fcst: -11: 14:204143:d=2021032312:HGT:575 mb:25 hour fcst: -11: + interp_atmos_master.sh[47]export err=0 -11: + interp_atmos_master.sh[47]err=0 -11: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -11: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -11: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f025_12_0p25 -11: + product_functions.sh[5]local filename=pgb2bfile_f025_12_0p25 -11: + product_functions.sh[6]wgrib2 pgb2bfile_f025_12_0p25 -not_if :RH: -grib pgb2bfile_f025_12_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f025_12_0p25.new -11: 1:0:d=2021032312:TMP:525 mb:25 hour fcst: -11: 2:430407:d=2021032312:RH:525 mb:25 hour fcst: -11: 3:879816:d=2021032312:TCDC:525 mb:25 hour fcst: -11: 4:1113140:d=2021032312:VVEL:525 mb:25 hour fcst: -11: 5:1782620:d=2021032312:DZDT:525 mb:25 hour fcst: -11: 6:2557970:d=2021032312:UGRD:525 mb:25 hour fcst: -11: 7:3075827:d=2021032312:VGRD:525 mb:25 hour fcst: -11: 8:3620977:d=2021032312:ABSV:525 mb:25 hour fcst: -11: 9:4120508:d=2021032312:CLMR:525 mb:25 hour fcst: -11: 10:4163145:d=2021032312:ICMR:525 mb:25 hour fcst: -11: 11:4546886:d=2021032312:RWMR:525 mb:25 hour fcst: -11: 12:4564296:d=2021032312:SNMR:525 mb:25 hour fcst: -11: 13:4875186:d=2021032312:GRLE:525 mb:25 hour fcst: -11: 14:4906861:d=2021032312:HGT:575 mb:25 hour fcst: -11: + product_functions.sh[10]rc=0 -11: + product_functions.sh[11](( rc == 0 )) -11: + product_functions.sh[11]mv pgb2bfile_f025_12_0p25.new pgb2bfile_f025_12_0p25 -11: + product_functions.sh[12]return 0 -11: + interp_atmos_master.sh[56]export err=0 -11: + interp_atmos_master.sh[56]err=0 -11: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -11: ++ interp_atmos_master.sh[62]wc -l -11: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f025_12_0p25 -match 'LAND|ICEC' -11: + interp_atmos_master.sh[62]var_count=0 -11: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -11: + interp_atmos_master.sh[73]exit 0 -+ run_mpmd.sh[113]exit 0 -+ run_mpmd.sh[1]postamble run_mpmd.sh 1753756030 0 -+ preamble.sh[62]set +x -End run_mpmd.sh at 02:27:24 with error code 0 (time elapsed: 00:00:14) -+ exglobal_atmos_products.sh[142]true -+ exglobal_atmos_products.sh[143]export err=0 -+ exglobal_atmos_products.sh[143]err=0 -+ exglobal_atmos_products.sh[144][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[149]mv mpmd.out mpmd_2.out -+ exglobal_atmos_products.sh[153]echo 'Concatenating processor-specific grib2 files into a single product file' -Concatenating processor-specific grib2 files into a single product file -+ exglobal_atmos_products.sh[154](( iproc = 1 )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_1_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_1_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_1 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_2_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_2_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_2 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_3_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_3_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_3 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_4_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_4_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_4 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_5_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_5_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_5 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_6_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_6_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_6 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_7_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_7_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_7 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_8_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_8_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_8 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_9_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_9_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_9 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_10_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_10_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_10 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_11_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_11_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_11 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_12_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_12_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_12 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_13_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_13_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_13 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_14_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_14_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_14 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_15_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_15_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_15 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_16_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_16_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_16 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_17_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_17_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_17 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_18_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_18_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_18 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_19_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_19_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_19 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_20_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_20_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_20 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_21_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_21_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_21 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_22_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_22_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_22 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_23_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_23_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_23 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f025_24_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f025_24_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f025_24 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[164]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[165]prod_dir=COMOUT_ATMOS_GRIB_0p25 -+ exglobal_atmos_products.sh[166]cpfs pgb2bfile_f025_0p25 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f025 -+ cpfs[3]'[' 2 -ne 2 ']' -+ cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f025 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f025 = ./ ']' -+ cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f025 ']' -+ cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f025 -+ cpfs[16]cp pgb2bfile_f025_0p25 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f025.cptmp -+ cpfs[18]'[' 0 -ne 0 ']' -+ cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f025.cptmp -+ cpfs[23]'[' 0 -ne 0 ']' -+ cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f025.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f025 -+ cpfs[28]'[' 0 -ne 0 ']' -+ exglobal_atmos_products.sh[167]wgrib2 -s pgb2bfile_f025_0p25 -+ exglobal_atmos_products.sh[170]echo 'Finished processing nset = 2' -Finished processing nset = 2 -+ exglobal_atmos_products.sh[154](( nset++ )) -+ exglobal_atmos_products.sh[154](( nset <= downset )) -+ exglobal_atmos_products.sh[177]FLUX_FILE=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.sfluxgrbf025.grib2 -+ exglobal_atmos_products.sh[178][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.sfluxgrbf025.grib2 ]] -+ exglobal_atmos_products.sh[179]wgrib2 -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.sfluxgrbf025.grib2 -+ exglobal_atmos_products.sh[185][[ YES == \Y\E\S ]] -+ exglobal_atmos_products.sh[188]input_file=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.sfluxgrbf025.grib2 -+ exglobal_atmos_products.sh[189]output_file_prefix=sflux_f025 -+ exglobal_atmos_products.sh[190]grid_string=1p00 -+ exglobal_atmos_products.sh[191]/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_sflux.sh /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.sfluxgrbf025.grib2 sflux_f025 1p00 -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ interp_atmos_sflux.sh[6]input_file=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.sfluxgrbf025.grib2 -+ interp_atmos_sflux.sh[7]output_file_prefix=sflux_f025 -+ interp_atmos_sflux.sh[8]grid_string=1p00 -+ interp_atmos_sflux.sh[11]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -+ interp_atmos_sflux.sh[12]interp_winds='-new_grid_winds earth' -+ interp_atmos_sflux.sh[13]interp_bilinear='-new_grid_interpolation bilinear' -+ interp_atmos_sflux.sh[14]interp_neighbor='-if :(LAND|CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -+ interp_atmos_sflux.sh[15]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -+ interp_atmos_sflux.sh[16]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -+ interp_atmos_sflux.sh[20]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -+ interp_atmos_sflux.sh[22]grid0p50='latlon 0:720:0.5 90:361:-0.5' -+ interp_atmos_sflux.sh[24]grid1p00='latlon 0:360:1.0 90:181:-1.0' -+ interp_atmos_sflux.sh[27]IFS=: -+ interp_atmos_sflux.sh[27]read -ra grids -+ interp_atmos_sflux.sh[29]output_grids= -+ interp_atmos_sflux.sh[30]for grid in "${grids[@]}" -+ interp_atmos_sflux.sh[31]gridopt=grid1p00 -+ interp_atmos_sflux.sh[32]output_grids=' -new_grid latlon 0:360:1.0 90:181:-1.0 sflux_f025_1p00' -+ interp_atmos_sflux.sh[36]wgrib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.sfluxgrbf025.grib2 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(LAND|CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:360:1.0 90:181:-1.0 sflux_f025_1p00 -1:0:d=2021032312:HGT:1 hybrid level:25 hour fcst: -2:11772:d=2021032312:TMP:1 hybrid level:25 hour fcst: -3:32639:d=2021032312:SPFH:1 hybrid level:25 hour fcst: -4:61495:d=2021032312:UGRD:1 hybrid level:25 hour fcst: -5:81796:d=2021032312:VGRD:1 hybrid level:25 hour fcst: -6:102529:d=2021032312:PRES:surface:25 hour fcst: -7:128137:d=2021032312:HGT:surface:25 hour fcst: -8:144207:d=2021032312:TMP:surface:25 hour fcst: -9:160850:d=2021032312:TSOIL:0-0.1 m below ground:25 hour fcst: -10:172774:d=2021032312:SOILW:0-0.1 m below ground:25 hour fcst: -11:181688:d=2021032312:SOILL:0-0.1 m below ground:25 hour fcst: -12:190494:d=2021032312:TSOIL:0.1-0.4 m below ground:25 hour fcst: -13:205970:d=2021032312:SOILW:0.1-0.4 m below ground:25 hour fcst: -14:214986:d=2021032312:SOILL:0.1-0.4 m below ground:25 hour fcst: -15:223880:d=2021032312:TSOIL:0.4-1 m below ground:25 hour fcst: -16:238326:d=2021032312:SOILW:0.4-1 m below ground:25 hour fcst: -17:247463:d=2021032312:SOILL:0.4-1 m below ground:25 hour fcst: -18:256382:d=2021032312:TSOIL:1-2 m below ground:25 hour fcst: -19:270668:d=2021032312:SOILW:1-2 m below ground:25 hour fcst: -20:279731:d=2021032312:SOILL:1-2 m below ground:25 hour fcst: -21:288809:d=2021032312:CISOILM:0-2 m below ground:25 hour fcst: -22:299319:d=2021032312:CNWAT:surface:25 hour fcst: -23:304863:d=2021032312:WEASD:surface:25 hour fcst: -24:317676:d=2021032312:SNOWC:surface:24-25 hour ave fcst: -25:322604:d=2021032312:SNOD:surface:25 hour fcst: -26:336966:d=2021032312:PEVPR:surface:25 hour fcst: -27:352734:d=2021032312:ICETK:surface:25 hour fcst: -28:356151:d=2021032312:ACOND:surface:25 hour fcst: -29:373694:d=2021032312:EVCW:surface:24-25 hour ave fcst: -30:383353:d=2021032312:EVBS:surface:24-25 hour ave fcst: -31:397822:d=2021032312:TRANS:surface:24-25 hour ave fcst: -32:406499:d=2021032312:SBSNO:surface:24-25 hour ave fcst: -33:413937:d=2021032312:var discipline=2 center=7 local_table=1 parmcat=0 parm=233:surface:25 hour fcst: -34:421958:d=2021032312:var discipline=2 center=7 local_table=1 parmcat=0 parm=235:surface:25 hour fcst: -35:426059:d=2021032312:var discipline=2 center=7 local_table=1 parmcat=0 parm=236:surface:25 hour fcst: -36:432747:d=2021032312:var discipline=2 center=7 local_table=1 parmcat=0 parm=237:surface:24-25 hour acc fcst: -37:438078:d=2021032312:var discipline=2 center=7 local_table=1 parmcat=0 parm=238:surface:24-25 hour acc fcst: -38:443645:d=2021032312:var discipline=2 center=7 local_table=1 parmcat=0 parm=239:surface:24-25 hour acc fcst: -39:451924:d=2021032312:var discipline=2 center=7 local_table=1 parmcat=0 parm=235:surface:24-25 hour ave fcst: -40:456152:d=2021032312:TMP:2 m above ground:25 hour fcst: -41:476953:d=2021032312:SPFH:2 m above ground:25 hour fcst: -42:505871:d=2021032312:TMAX:2 m above ground:24-25 hour max fcst: -43:526630:d=2021032312:TMIN:2 m above ground:24-25 hour min fcst: -44:547450:d=2021032312:QMAX:2 m above ground:24-25 hour max fcst: -45:564591:d=2021032312:QMIN:2 m above ground:24-25 hour min fcst: -46:581725:d=2021032312:UGRD:10 m above ground:25 hour fcst: -47:602236:d=2021032312:VGRD:10 m above ground:25 hour fcst: -48:623134:d=2021032312:CPOFP:surface:25 hour fcst: -49:643507:d=2021032312:CPRAT:surface:24-25 hour ave fcst: -50:667311:d=2021032312:PRATE:surface:24-25 hour ave fcst: -51:692626:d=2021032312:SSRUN:surface:24-25 hour acc fcst: -52:699642:d=2021032312:WATR:surface:24-25 hour acc fcst: -53:712385:d=2021032312:LHTFL:surface:24-25 hour ave fcst: -54:734191:d=2021032312:SHTFL:surface:24-25 hour ave fcst: -55:755187:d=2021032312:GFLUX:surface:24-25 hour ave fcst: -56:767259:d=2021032312:SNOHF:surface:24-25 hour ave fcst: -57:770745:d=2021032312:UFLX:surface:24-25 hour ave fcst: -58:787146:d=2021032312:VFLX:surface:24-25 hour ave fcst: -59:803662:d=2021032312:SFCR:surface:25 hour fcst: -60:822085:d=2021032312:FRICV:surface:25 hour fcst: -61:843822:d=2021032312:U-GWD:surface:24-25 hour ave fcst: -62:858694:d=2021032312:V-GWD:surface:24-25 hour ave fcst: -63:873634:d=2021032312:SHTFL:surface:25 hour fcst: -64:894627:d=2021032312:LHTFL:surface:25 hour fcst: -65:916437:d=2021032312:SFEXC:surface:25 hour fcst: -66:932964:d=2021032312:VEG:surface:25 hour fcst: -67:940998:d=2021032312:GFLUX:surface:25 hour fcst: -68:953020:d=2021032312:VGTYP:surface:25 hour fcst: -69:961553:d=2021032312:SOTYP:surface:25 hour fcst: -70:969512:d=2021032312:SLTYP:surface:25 hour fcst: -71:979724:d=2021032312:WILT:surface:25 hour fcst: -72:989634:d=2021032312:FLDCP:surface:25 hour fcst: -73:999840:d=2021032312:SUNSD:surface:25 hour fcst: -74:1007003:d=2021032312:PEVPR:surface:24-25 hour ave fcst: -75:1022739:d=2021032312:PWAT:entire atmosphere (considered as a single layer):25 hour fcst: -76:1049680:d=2021032312:LCDC:low cloud layer:24-25 hour ave fcst: -77:1067497:d=2021032312:MCDC:middle cloud layer:24-25 hour ave fcst: -78:1082962:d=2021032312:HCDC:high cloud layer:24-25 hour ave fcst: -79:1098805:d=2021032312:TCDC:entire atmosphere (considered as a single layer):24-25 hour ave fcst: -80:1116035:d=2021032312:PRES:convective cloud bottom level:25 hour fcst: -81:1129693:d=2021032312:PRES:low cloud bottom level:24-25 hour ave fcst: -82:1150781:d=2021032312:PRES:middle cloud bottom level:24-25 hour ave fcst: -83:1167071:d=2021032312:PRES:high cloud bottom level:24-25 hour ave fcst: -84:1183841:d=2021032312:PRES:convective cloud top level:25 hour fcst: -85:1199204:d=2021032312:PRES:low cloud top level:24-25 hour ave fcst: -86:1220292:d=2021032312:PRES:middle cloud top level:24-25 hour ave fcst: -87:1236403:d=2021032312:PRES:high cloud top level:24-25 hour ave fcst: -88:1253698:d=2021032312:TMP:low cloud top level:24-25 hour ave fcst: -89:1269581:d=2021032312:TMP:middle cloud top level:24-25 hour ave fcst: -90:1282626:d=2021032312:TMP:high cloud top level:24-25 hour ave fcst: -91:1296994:d=2021032312:TCDC:convective cloud layer:25 hour fcst: -92:1317629:d=2021032312:TCDC:boundary layer cloud layer:24-25 hour ave fcst: -93:1332985:d=2021032312:CWORK:entire atmosphere (considered as a single layer):24-25 hour ave fcst: -94:1341350:d=2021032312:DSWRF:surface:24-25 hour ave fcst: -95:1359207:d=2021032312:DUVB:surface:24-25 hour ave fcst: -96:1375923:d=2021032312:CDUVB:surface:24-25 hour ave fcst: -97:1389089:d=2021032312:DLWRF:surface:24-25 hour ave fcst: -98:1410800:d=2021032312:USWRF:surface:24-25 hour ave fcst: -99:1428092:d=2021032312:ULWRF:surface:24-25 hour ave fcst: -100:1447046:d=2021032312:USWRF:top of atmosphere:24-25 hour ave fcst: -101:1465130:d=2021032312:ULWRF:top of atmosphere:24-25 hour ave fcst: -102:1486276:d=2021032312:DSWRF:surface:25 hour fcst: -103:1503464:d=2021032312:DLWRF:surface:25 hour fcst: -104:1525177:d=2021032312:USWRF:surface:25 hour fcst: -105:1541866:d=2021032312:ULWRF:surface:25 hour fcst: -106:1560791:d=2021032312:CSDSF:surface:24-25 hour ave fcst: -107:1574461:d=2021032312:CSUSF:surface:24-25 hour ave fcst: -108:1590088:d=2021032312:CSUSF:top of atmosphere:24-25 hour ave fcst: -109:1606045:d=2021032312:DSWRF:top of atmosphere:24-25 hour ave fcst: -110:1614540:d=2021032312:CSDLF:surface:24-25 hour ave fcst: -111:1632779:d=2021032312:CSULF:surface:24-25 hour ave fcst: -112:1651510:d=2021032312:CSULF:top of atmosphere:24-25 hour ave fcst: -113:1668286:d=2021032312:VBDSF:surface:24-25 hour ave fcst: -114:1684838:d=2021032312:VDDSF:surface:24-25 hour ave fcst: -115:1702708:d=2021032312:NBDSF:surface:24-25 hour ave fcst: -116:1720272:d=2021032312:NDDSF:surface:24-25 hour ave fcst: -117:1738444:d=2021032312:HPBL:surface:25 hour fcst: -118:1769475:d=2021032312:LAND:surface:25 hour fcst: -119:1771416:d=2021032312:ICEC:surface:25 hour fcst: -120:1774585:d=2021032312:ALBDO:surface:24-25 hour ave fcst: -121:1789069:d=2021032312:FDNSSTMP:surface:25 hour fcst: -+ interp_atmos_sflux.sh[43]export err=0 -+ interp_atmos_sflux.sh[43]err=0 -+ interp_atmos_sflux.sh[44][[ 0 -ne 0 ]] -+ interp_atmos_sflux.sh[49]exit 0 -+ exglobal_atmos_products.sh[191]true -+ exglobal_atmos_products.sh[192]export err=0 -+ exglobal_atmos_products.sh[192]err=0 -+ exglobal_atmos_products.sh[193][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[198]IFS=: -+ exglobal_atmos_products.sh[198]read -ra grids -+ exglobal_atmos_products.sh[199]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[200]prod_dir=COMOUT_ATMOS_GRIB_1p00 -+ exglobal_atmos_products.sh[201]cpfs sflux_f025_1p00 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f025 -+ cpfs[3]'[' 2 -ne 2 ']' -+ cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f025 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f025 = ./ ']' -+ cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f025 ']' -+ cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f025 -+ cpfs[16]cp sflux_f025_1p00 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f025.cptmp -+ cpfs[18]'[' 0 -ne 0 ']' -+ cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f025.cptmp -+ cpfs[23]'[' 0 -ne 0 ']' -+ cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f025.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f025 -+ cpfs[28]'[' 0 -ne 0 ']' -+ exglobal_atmos_products.sh[202]wgrib2 -s sflux_f025_1p00 -+ exglobal_atmos_products.sh[207][[ YES == \Y\E\S ]] -+ exglobal_atmos_products.sh[208]grp= -+ exglobal_atmos_products.sh[209][[ 25 -gt 0 ]] -+ exglobal_atmos_products.sh[209][[ 25 -le 180 ]] -+ exglobal_atmos_products.sh[211]wgrib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f025 -d 597 -grib /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.wgne.f025 -597:256718351:d=2021032312:APCP:surface:24-25 hour acc fcst: -+ exglobal_atmos_products.sh[219][[ NO == \Y\E\S ]] -+ exglobal_atmos_products.sh[290]exit 0 -+ JGLOBAL_ATMOS_PRODUCTS[27]true -+ JGLOBAL_ATMOS_PRODUCTS[28]export err=0 -+ JGLOBAL_ATMOS_PRODUCTS[28]err=0 -+ JGLOBAL_ATMOS_PRODUCTS[29][[ 0 -ne 0 ]] -+ JGLOBAL_ATMOS_PRODUCTS[40][[ -e OUTPUT.548446 ]] -+ JGLOBAL_ATMOS_PRODUCTS[47]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312 -+ JGLOBAL_ATMOS_PRODUCTS[48][[ NO == \N\O ]] -+ JGLOBAL_ATMOS_PRODUCTS[49]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f025.533601 -+ JGLOBAL_ATMOS_PRODUCTS[53]exit 0 -+ JGLOBAL_ATMOS_PRODUCTS[1]postamble /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_ATMOS_PRODUCTS 1753755969 0 -+ preamble.sh[62]set +x -End /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_ATMOS_PRODUCTS at 02:27:31 with error code 0 (time elapsed: 00:01:22) -+ status=0 -+ [[ 0 -ne 0 ]] -+ for FORECAST_HOUR in "${fhr_list[@]}" -++ printf %03d 26 -+ fhr3=026 -+ jobid=atmos_products_f026.533601 -+ /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_ATMOS_PRODUCTS -Begin /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_ATMOS_PRODUCTS at Tue Jul 29 02:27:31 UTC 2025 -++ jjob_header.sh[46]OPTIND=1 -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[50]env_job=atmos_products -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[49]read -ra configs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[61]shift 4 -++ jjob_header.sh[63][[ -z atmos_products ]] -++ jjob_header.sh[71]export DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601 -++ jjob_header.sh[71]DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601 -++ jjob_header.sh[72][[ YES == \Y\E\S ]] -++ jjob_header.sh[73]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601 -++ jjob_header.sh[75]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601 -++ jjob_header.sh[76]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601 -++ jjob_header.sh[85]export pid=559401 -++ jjob_header.sh[85]pid=559401 -++ jjob_header.sh[86]export pgmout=OUTPUT.559401 -++ jjob_header.sh[86]pgmout=OUTPUT.559401 -++ jjob_header.sh[87]export pgmerr=errfile -++ jjob_header.sh[87]pgmerr=errfile -++ jjob_header.sh[90]export pgm= -++ jjob_header.sh[90]pgm= -++ jjob_header.sh[96]export cycle=t12z -++ jjob_header.sh[96]cycle=t12z -++ jjob_header.sh[97]setpdy.sh -+ setpdy.sh[20]'[' /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601 == /home/mterry ']' -+ setpdy.sh[25][[ ! t12z =~ t??z ]] -+ setpdy.sh[30]case $# in -+ setpdy.sh[31]dates_before_PDY=7 -+ setpdy.sh[32]dates_after_PDY=7 -+ setpdy.sh[50]COMDATEROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+ setpdy.sh[53]'[' -z 20210323 ']' -+ setpdy.sh[57]sed 's/[0-9]\{8\}/20210323/' /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z -sed: can't read /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z: No such file or directory -++ jjob_header.sh[97]true -++ jjob_header.sh[98]source ./PDY -/work2/noaa/global/mterry/global-workflow_forked/ush/jjob_header.sh: line 98: ./PDY: No such file or directory -++ jjob_header.sh[98]true -++ jjob_header.sh[104]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[104]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.base -+++ config.base[6]echo 'BEGIN: config.base' -BEGIN: config.base -+++ config.base[9]export machine=HERCULES -+++ config.base[9]machine=HERCULES -+++ config.base[12]export RUN_ENVIR=emc -+++ config.base[12]RUN_ENVIR=emc -+++ config.base[15]export ACCOUNT=fv3-cpu -+++ config.base[15]ACCOUNT=fv3-cpu -+++ config.base[16]export QUEUE=batch -+++ config.base[16]QUEUE=batch -+++ config.base[17]export QUEUE_SERVICE=batch -+++ config.base[17]QUEUE_SERVICE=batch -+++ config.base[18]export QUEUE_DTN=batch -+++ config.base[18]QUEUE_DTN=batch -+++ config.base[19]export PARTITION_BATCH=hercules -+++ config.base[19]PARTITION_BATCH=hercules -+++ config.base[20]export PARTITION_SERVICE=service -+++ config.base[20]PARTITION_SERVICE=service -+++ config.base[21]export PARTITION_DTN= -+++ config.base[21]PARTITION_DTN= -+++ config.base[22]export RESERVATION= -+++ config.base[22]RESERVATION= -+++ config.base[23]export CLUSTERS= -+++ config.base[23]CLUSTERS= -+++ config.base[24]export CLUSTERS_SERVICE= -+++ config.base[24]CLUSTERS_SERVICE= -+++ config.base[25]export CLUSTERS_DTN= -+++ config.base[25]CLUSTERS_DTN= -+++ config.base[28]export HPSS_PROJECT=emc-global -+++ config.base[28]HPSS_PROJECT=emc-global -+++ config.base[31]export HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[31]HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[32]export EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[32]EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[33]export FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[33]FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[34]export PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[34]PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[35]export SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[35]SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[36]export USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[36]USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[38]export FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[38]FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[39]export FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[39]FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[40]export FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[40]FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[41]export FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[41]FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[42]export FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[42]FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[43]export FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[43]FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[44]export FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[44]FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[45]export FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[45]FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[50]export PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[50]PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[51]export COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[51]COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[52]export COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[52]COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[53]export DMPDIR=/work/noaa/rstprod/dump -+++ config.base[53]DMPDIR=/work/noaa/rstprod/dump -+++ config.base[57]export COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[57]COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[58]export COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[58]COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[59]export COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[59]COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[62]export HOMEDIR=/work2/noaa/global/mterry -+++ config.base[62]HOMEDIR=/work2/noaa/global/mterry -+++ config.base[63]export STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[63]STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]export PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[65]export NOSCRUB=/work2/noaa/global/mterry -+++ config.base[65]NOSCRUB=/work2/noaa/global/mterry -+++ config.base[68]export BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[68]BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[71]export BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[71]BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[74]export DO_PREP_SFC=NO -+++ config.base[74]DO_PREP_SFC=NO -+++ config.base[77]export DO_GOES=NO -+++ config.base[77]DO_GOES=NO -+++ config.base[78]export DO_BUFRSND=NO -+++ config.base[78]DO_BUFRSND=NO -+++ config.base[79]export DO_GEMPAK=NO -+++ config.base[79]DO_GEMPAK=NO -+++ config.base[80]export DO_AWIPS=NO -+++ config.base[80]DO_AWIPS=NO -+++ config.base[81]export DO_NPOESS=NO -+++ config.base[81]DO_NPOESS=NO -+++ config.base[82]export DO_TRACKER=YES -+++ config.base[82]DO_TRACKER=YES -+++ config.base[83]export DO_GENESIS=YES -+++ config.base[83]DO_GENESIS=YES -+++ config.base[84]export DO_GENESIS_FSU=NO -+++ config.base[84]DO_GENESIS_FSU=NO -+++ config.base[85]export DO_VERFOZN=YES -+++ config.base[85]DO_VERFOZN=YES -+++ config.base[86]export DO_VERFRAD=YES -+++ config.base[86]DO_VERFRAD=YES -+++ config.base[87]export DO_VMINMON=YES -+++ config.base[87]DO_VMINMON=YES -+++ config.base[88]export DO_ANLSTAT=NO -+++ config.base[88]DO_ANLSTAT=NO -+++ config.base[91]export MODE=forecast-only -+++ config.base[91]MODE=forecast-only -+++ config.base[92]export DO_TEST_MODE=YES -+++ config.base[92]DO_TEST_MODE=YES -+++ config.base[101]export FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[101]FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[102]export HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[102]HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[103]export HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v1.2.0 -+++ config.base[103]HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v1.2.0 -+++ config.base[106]export NMV=/bin/mv -+++ config.base[106]NMV=/bin/mv -+++ config.base[107]export 'NLN=/bin/ln -sf' -+++ config.base[107]NLN='/bin/ln -sf' -+++ config.base[108]export VERBOSE=YES -+++ config.base[108]VERBOSE=YES -+++ config.base[109]export KEEPDATA=NO -+++ config.base[109]KEEPDATA=NO -+++ config.base[110]export DEBUG_POSTSCRIPT=NO -+++ config.base[110]DEBUG_POSTSCRIPT=NO -+++ config.base[111]export CHGRP_RSTPROD=YES -+++ config.base[111]CHGRP_RSTPROD=YES -+++ config.base[112]export 'CHGRP_CMD=chgrp rstprod' -+++ config.base[112]CHGRP_CMD='chgrp rstprod' -+++ config.base[113]export NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[113]NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[114]export NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[114]NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[117]export BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[117]BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[120]export SDATE=2021032312 -+++ config.base[120]SDATE=2021032312 -+++ config.base[121]export EDATE=2021032312 -+++ config.base[121]EDATE=2021032312 -+++ config.base[122]export EXP_WARM_START=.false. -+++ config.base[122]EXP_WARM_START=.false. -+++ config.base[123]export assim_freq=6 -+++ config.base[123]assim_freq=6 -+++ config.base[124]export PSLOT=C48_S2SW -+++ config.base[124]PSLOT=C48_S2SW -+++ config.base[125]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[125]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[126]export ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[126]ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[127]export DUMP_SUFFIX= -+++ config.base[127]DUMP_SUFFIX= -+++ config.base[128][[ 2021032312 -ge 2019092100 ]] -+++ config.base[128][[ 2021032312 -le 2019110700 ]] -+++ config.base[131]export ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[131]ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[132]export ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[132]ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[133]export FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[133]FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[136]export envir=prod -+++ config.base[136]envir=prod -+++ config.base[137]export NET=gfs -+++ config.base[137]NET=gfs -+++ config.base[138]export RUN=gfs -+++ config.base[138]RUN=gfs -+++ config.base[141]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.com -++++ config.com[4]echo 'BEGIN: config.com' -BEGIN: config.com -++++ config.com[38][[ emc == \n\c\o ]] -++++ config.com[43]COM_OBSPROC_TMPL='${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos' -++++ config.com[44]COM_RTOFS_TMPL='${DMPDIR}' -++++ config.com[45]COM_TCVITAL_TMPL='${DMPDIR}/${RUN}.${YMD}/${HH}/atmos' -++++ config.com[47]declare -rx 'COM_OBS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/obs' -++++ config.com[48]declare -rx COM_OBSPROC_TMPL COM_RTOFS_TMPL -++++ config.com[50]COM_BASE='${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}' -++++ config.com[52]declare -rx 'COM_TOP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}' -++++ config.com[54]declare -rx 'COM_CONF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf' -++++ config.com[55]declare -rx 'COM_OBS_JEDI=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi' -++++ config.com[57]declare -rx 'COM_ATMOS_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input' -++++ config.com[58]declare -rx 'COM_ATMOS_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart' -++++ config.com[59]declare -rx 'COM_ATMOS_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos' -++++ config.com[60]declare -rx 'COM_SNOW_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow' -++++ config.com[61]declare -rx 'COM_SNOW_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon' -++++ config.com[62]declare -rx 'COM_ATMOS_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history' -++++ config.com[63]declare -rx 'COM_ATMOS_MASTER_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master' -++++ config.com[64]declare -rx 'COM_ATMOS_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2' -++++ config.com[65]declare -rx 'COM_ATMOS_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}' -++++ config.com[66]declare -rx 'COM_ATMOS_BUFR_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr' -++++ config.com[67]declare -rx 'COM_ATMOS_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}' -++++ config.com[68]declare -rx 'COM_ATMOS_GENESIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital' -++++ config.com[69]declare -rx 'COM_ATMOS_TRACK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks' -++++ config.com[70]declare -rx 'COM_ATMOS_GOES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim' -++++ config.com[71]declare -rx 'COM_ATMOS_IMAGERY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery' -++++ config.com[72]declare -rx 'COM_ATMOS_OZNMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon' -++++ config.com[73]declare -rx 'COM_ATMOS_RADMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon' -++++ config.com[74]declare -rx 'COM_ATMOS_MINMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon' -++++ config.com[75]declare -rx 'COM_ATMOS_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon' -++++ config.com[76]declare -rx 'COM_ATMOS_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo' -++++ config.com[78]declare -rx 'COM_WAVE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart' -++++ config.com[79]declare -rx 'COM_WAVE_PREP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep' -++++ config.com[80]declare -rx 'COM_WAVE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history' -++++ config.com[81]declare -rx 'COM_WAVE_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded' -++++ config.com[82]declare -rx 'COM_WAVE_GRID_RES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}' -++++ config.com[83]declare -rx 'COM_WAVE_STATION_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station' -++++ config.com[84]declare -rx 'COM_WAVE_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak' -++++ config.com[85]declare -rx 'COM_WAVE_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo' -++++ config.com[87]declare -rx 'COM_OCEAN_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history' -++++ config.com[88]declare -rx 'COM_OCEAN_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart' -++++ config.com[89]declare -rx 'COM_OCEAN_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input' -++++ config.com[90]declare -rx 'COM_OCEAN_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean' -++++ config.com[91]declare -rx 'COM_OCEAN_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon' -++++ config.com[92]declare -rx 'COM_OCEAN_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf' -++++ config.com[93]declare -rx 'COM_OCEAN_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean' -++++ config.com[94]declare -rx 'COM_OCEAN_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf' -++++ config.com[95]declare -rx 'COM_OCEAN_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2' -++++ config.com[96]declare -rx 'COM_OCEAN_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}' -++++ config.com[98]declare -rx 'COM_ICE_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice' -++++ config.com[99]declare -rx 'COM_ICE_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf' -++++ config.com[100]declare -rx 'COM_ICE_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon' -++++ config.com[101]declare -rx 'COM_ICE_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice' -++++ config.com[102]declare -rx 'COM_ICE_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input' -++++ config.com[103]declare -rx 'COM_ICE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history' -++++ config.com[104]declare -rx 'COM_ICE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart' -++++ config.com[105]declare -rx 'COM_ICE_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf' -++++ config.com[106]declare -rx 'COM_ICE_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2' -++++ config.com[107]declare -rx 'COM_ICE_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}' -++++ config.com[109]declare -rx 'COM_CHEM_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history' -++++ config.com[110]declare -rx 'COM_CHEM_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem' -++++ config.com[111]declare -rx 'COM_CHEM_BMAT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix' -++++ config.com[112]declare -rx 'COM_CHEM_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon' -++++ config.com[114]declare -rx 'COM_MED_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart' -+++ config.base[143]export LOGSCRIPT= -+++ config.base[143]LOGSCRIPT= -+++ config.base[145]export 'REDOUT=1>' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.atmos_products -+++ config.atmos_products[6]echo 'BEGIN: config.atmos_products' -BEGIN: config.atmos_products -+++ config.atmos_products[9]. /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources atmos_products -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=atmos_products -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[1019]walltime=00:15:00 -++++ config.resources[1020]ntasks=24 -++++ config.resources[1021]threads_per_task=1 -++++ config.resources[1022]tasks_per_node=24 -++++ config.resources[1023]export is_exclusive=True -++++ config.resources[1023]is_exclusive=True -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n '' ]] -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n '' ]] -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.atmos_products[12]export MAX_TASKS=25 -+++ config.atmos_products[12]MAX_TASKS=25 -+++ config.atmos_products[15]export INTERP_ATMOS_MASTERSH=/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh -+++ config.atmos_products[15]INTERP_ATMOS_MASTERSH=/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh -+++ config.atmos_products[16]export INTERP_ATMOS_SFLUXSH=/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_sflux.sh -+++ config.atmos_products[16]INTERP_ATMOS_SFLUXSH=/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_sflux.sh -+++ config.atmos_products[18][[ gfs == \g\d\a\s ]] -+++ config.atmos_products[24][[ gfs == \g\f\s ]] -+++ config.atmos_products[25]export downset=2 -+++ config.atmos_products[25]downset=2 -+++ config.atmos_products[26]export FHOUT_PGBS=3 -+++ config.atmos_products[26]FHOUT_PGBS=3 -+++ config.atmos_products[27]export FLXGF=YES -+++ config.atmos_products[27]FLXGF=YES -+++ config.atmos_products[28]export WGNE=YES -+++ config.atmos_products[28]WGNE=YES -+++ config.atmos_products[29]export FHMAX_WGNE=180 -+++ config.atmos_products[29]FHMAX_WGNE=180 -+++ config.atmos_products[32]export APCP_MSG=597 -+++ config.atmos_products[32]APCP_MSG=597 -+++ config.atmos_products[35]export paramlista=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.fFFF.paramlist.a.txt -+++ config.atmos_products[35]paramlista=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.fFFF.paramlist.a.txt -+++ config.atmos_products[36]export paramlista_anl=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.anl.paramlist.a.txt -+++ config.atmos_products[36]paramlista_anl=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.anl.paramlist.a.txt -+++ config.atmos_products[37]export paramlista_f000=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.f000.paramlist.a.txt -+++ config.atmos_products[37]paramlista_f000=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.f000.paramlist.a.txt -+++ config.atmos_products[38]export paramlistb=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.fFFF.paramlist.b.txt -+++ config.atmos_products[38]paramlistb=/work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.fFFF.paramlist.b.txt -+++ config.atmos_products[40]echo 'END: config.atmos_products' -END: config.atmos_products -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env atmos_products -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=atmos_products -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 520192000 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 24 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 24 ]] -+++ HERCULES.env[34]max_threads_per_task=3 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 3 ]] -+++ HERCULES.env[40][[ 1 -gt 3 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 24' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[281]export USE_CFP=YES -+++ HERCULES.env[281]USE_CFP=YES -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_ATMOS_PRODUCTS[11]YMD=20210323 -+ JGLOBAL_ATMOS_PRODUCTS[11]HH=12 -+ JGLOBAL_ATMOS_PRODUCTS[11]declare_from_tmpl -rx COMIN_ATMOS_ANALYSIS:COM_ATMOS_ANALYSIS_TMPL COMIN_ATMOS_HISTORY:COM_ATMOS_HISTORY_TMPL COMIN_ATMOS_MASTER:COM_ATMOS_MASTER_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_ATMOS_ANALYSIS=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//analysis/atmos -declare_from_tmpl :: COMIN_ATMOS_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/history -declare_from_tmpl :: COMIN_ATMOS_MASTER=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master -+ JGLOBAL_ATMOS_PRODUCTS[16]for grid in '0p25' '0p50' '1p00' -+ JGLOBAL_ATMOS_PRODUCTS[17]prod_dir=COMOUT_ATMOS_GRIB_0p25 -+ JGLOBAL_ATMOS_PRODUCTS[18]GRID=0p25 -+ JGLOBAL_ATMOS_PRODUCTS[18]YMD=20210323 -+ JGLOBAL_ATMOS_PRODUCTS[18]HH=12 -+ JGLOBAL_ATMOS_PRODUCTS[18]declare_from_tmpl -rx COMOUT_ATMOS_GRIB_0p25:COM_ATMOS_GRIB_GRID_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_ATMOS_GRIB_0p25=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25 -+ JGLOBAL_ATMOS_PRODUCTS[19][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25 ]] -+ JGLOBAL_ATMOS_PRODUCTS[16]for grid in '0p25' '0p50' '1p00' -+ JGLOBAL_ATMOS_PRODUCTS[17]prod_dir=COMOUT_ATMOS_GRIB_0p50 -+ JGLOBAL_ATMOS_PRODUCTS[18]GRID=0p50 -+ JGLOBAL_ATMOS_PRODUCTS[18]YMD=20210323 -+ JGLOBAL_ATMOS_PRODUCTS[18]HH=12 -+ JGLOBAL_ATMOS_PRODUCTS[18]declare_from_tmpl -rx COMOUT_ATMOS_GRIB_0p50:COM_ATMOS_GRIB_GRID_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_ATMOS_GRIB_0p50=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50 -+ JGLOBAL_ATMOS_PRODUCTS[19][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p50 ]] -+ JGLOBAL_ATMOS_PRODUCTS[16]for grid in '0p25' '0p50' '1p00' -+ JGLOBAL_ATMOS_PRODUCTS[17]prod_dir=COMOUT_ATMOS_GRIB_1p00 -+ JGLOBAL_ATMOS_PRODUCTS[18]GRID=1p00 -+ JGLOBAL_ATMOS_PRODUCTS[18]YMD=20210323 -+ JGLOBAL_ATMOS_PRODUCTS[18]HH=12 -+ JGLOBAL_ATMOS_PRODUCTS[18]declare_from_tmpl -rx COMOUT_ATMOS_GRIB_1p00:COM_ATMOS_GRIB_GRID_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_ATMOS_GRIB_1p00=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00 -+ JGLOBAL_ATMOS_PRODUCTS[19][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00 ]] -+ JGLOBAL_ATMOS_PRODUCTS[23]export PREFIX=gfs.t12z. -+ JGLOBAL_ATMOS_PRODUCTS[23]PREFIX=gfs.t12z. -+ JGLOBAL_ATMOS_PRODUCTS[27]/work2/noaa/global/mterry/global-workflow_forked/scripts/exglobal_atmos_products.sh -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ exglobal_atmos_products.sh[4]INTERP_ATMOS_MASTERSH=/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh -+ exglobal_atmos_products.sh[5]INTERP_ATMOS_SFLUXSH=/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_sflux.sh -+ exglobal_atmos_products.sh[8]downset=2 -+ exglobal_atmos_products.sh[9]ntasks_atmos_products=8 -+ exglobal_atmos_products.sh[12]WGNE=YES -+ exglobal_atmos_products.sh[13]FHMAX_WGNE=180 -+ exglobal_atmos_products.sh[15]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601 -+ exglobal_atmos_products.sh[19][[ 26 -le 0 ]] -++ exglobal_atmos_products.sh[30]printf f%03d 26 -+ exglobal_atmos_products.sh[30]fhr3=f026 -+ exglobal_atmos_products.sh[31](( FORECAST_HOUR%FHOUT_PGBS == 0 )) -+ exglobal_atmos_products.sh[40]MASTER_FILE=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.master.grb2f026 -+ exglobal_atmos_products.sh[45]wgrib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.master.grb2f026 -+ exglobal_atmos_products.sh[45]wgrib2 -i -grib tmpfile_f026 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.master.grb2f026 -+ exglobal_atmos_products.sh[45]grep -F -f /work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.fFFF.paramlist.a.txt -1:0:d=2021032312:PRMSL:mean sea level:26 hour fcst: -2:25750:d=2021032312:CLMR:1 hybrid level:26 hour fcst: -3:29476:d=2021032312:ICMR:1 hybrid level:26 hour fcst: -4:36524:d=2021032312:RWMR:1 hybrid level:26 hour fcst: -5:48816:d=2021032312:SNMR:1 hybrid level:26 hour fcst: -6:58979:d=2021032312:GRLE:1 hybrid level:26 hour fcst: -7:61830:d=2021032312:REFD:1 hybrid level:26 hour fcst: -8:87479:d=2021032312:REFD:2 hybrid level:26 hour fcst: -9:113130:d=2021032312:REFC:entire atmosphere (considered as a single layer):26 hour fcst: -10:140473:d=2021032312:VIS:surface:26 hour fcst: -11:172814:d=2021032312:UGRD:planetary boundary layer:26 hour fcst: -12:186817:d=2021032312:VGRD:planetary boundary layer:26 hour fcst: -13:201364:d=2021032312:VRATE:planetary boundary layer:26 hour fcst: -14:216434:d=2021032312:GUST:surface:26 hour fcst: -15:231225:d=2021032312:HGT:0.01 mb:26 hour fcst: -16:254282:d=2021032312:TMP:0.01 mb:26 hour fcst: -17:264808:d=2021032312:RH:0.01 mb:26 hour fcst: -18:271141:d=2021032312:SPFH:0.01 mb:26 hour fcst: -19:294155:d=2021032312:VVEL:0.01 mb:26 hour fcst: -20:320852:d=2021032312:DZDT:0.01 mb:26 hour fcst: -21:348169:d=2021032312:UGRD:0.01 mb:26 hour fcst: -22:359247:d=2021032312:VGRD:0.01 mb:26 hour fcst: -23:369910:d=2021032312:ABSV:0.01 mb:26 hour fcst: -24:385354:d=2021032312:O3MR:0.01 mb:26 hour fcst: -25:399357:d=2021032312:HGT:0.02 mb:26 hour fcst: -26:424604:d=2021032312:TMP:0.02 mb:26 hour fcst: -27:439301:d=2021032312:RH:0.02 mb:26 hour fcst: -28:446712:d=2021032312:SPFH:0.02 mb:26 hour fcst: -29:470738:d=2021032312:VVEL:0.02 mb:26 hour fcst: -30:494829:d=2021032312:DZDT:0.02 mb:26 hour fcst: -31:522305:d=2021032312:UGRD:0.02 mb:26 hour fcst: -32:534793:d=2021032312:VGRD:0.02 mb:26 hour fcst: -33:546699:d=2021032312:ABSV:0.02 mb:26 hour fcst: -34:563631:d=2021032312:O3MR:0.02 mb:26 hour fcst: -35:579187:d=2021032312:HGT:0.04 mb:26 hour fcst: -36:603337:d=2021032312:TMP:0.04 mb:26 hour fcst: -37:620486:d=2021032312:RH:0.04 mb:26 hour fcst: -38:626524:d=2021032312:SPFH:0.04 mb:26 hour fcst: -39:652055:d=2021032312:VVEL:0.04 mb:26 hour fcst: -40:679552:d=2021032312:DZDT:0.04 mb:26 hour fcst: -41:704107:d=2021032312:UGRD:0.04 mb:26 hour fcst: -42:723980:d=2021032312:VGRD:0.04 mb:26 hour fcst: -43:743716:d=2021032312:ABSV:0.04 mb:26 hour fcst: -44:761961:d=2021032312:O3MR:0.04 mb:26 hour fcst: -45:785458:d=2021032312:HGT:0.07 mb:26 hour fcst: -46:809421:d=2021032312:TMP:0.07 mb:26 hour fcst: -47:826363:d=2021032312:RH:0.07 mb:26 hour fcst: -48:835871:d=2021032312:SPFH:0.07 mb:26 hour fcst: -49:859355:d=2021032312:VVEL:0.07 mb:26 hour fcst: -50:888114:d=2021032312:DZDT:0.07 mb:26 hour fcst: -51:913456:d=2021032312:UGRD:0.07 mb:26 hour fcst: -52:933691:d=2021032312:VGRD:0.07 mb:26 hour fcst: -53:953904:d=2021032312:ABSV:0.07 mb:26 hour fcst: -54:972680:d=2021032312:O3MR:0.07 mb:26 hour fcst: -55:992097:d=2021032312:HGT:0.1 mb:26 hour fcst: -56:1013512:d=2021032312:TMP:0.1 mb:26 hour fcst: -57:1029925:d=2021032312:RH:0.1 mb:26 hour fcst: -58:1038011:d=2021032312:SPFH:0.1 mb:26 hour fcst: -59:1059318:d=2021032312:VVEL:0.1 mb:26 hour fcst: -60:1088866:d=2021032312:DZDT:0.1 mb:26 hour fcst: -61:1114475:d=2021032312:UGRD:0.1 mb:26 hour fcst: -62:1134476:d=2021032312:VGRD:0.1 mb:26 hour fcst: -63:1154499:d=2021032312:ABSV:0.1 mb:26 hour fcst: -64:1172993:d=2021032312:O3MR:0.1 mb:26 hour fcst: -65:1196524:d=2021032312:HGT:0.2 mb:26 hour fcst: -66:1217567:d=2021032312:TMP:0.2 mb:26 hour fcst: -67:1233686:d=2021032312:RH:0.2 mb:26 hour fcst: -68:1244656:d=2021032312:SPFH:0.2 mb:26 hour fcst: -69:1264455:d=2021032312:VVEL:0.2 mb:26 hour fcst: -70:1293312:d=2021032312:DZDT:0.2 mb:26 hour fcst: -71:1320768:d=2021032312:UGRD:0.2 mb:26 hour fcst: -72:1340056:d=2021032312:VGRD:0.2 mb:26 hour fcst: -73:1358945:d=2021032312:ABSV:0.2 mb:26 hour fcst: -74:1376593:d=2021032312:O3MR:0.2 mb:26 hour fcst: -75:1402230:d=2021032312:HGT:0.4 mb:26 hour fcst: -76:1423051:d=2021032312:TMP:0.4 mb:26 hour fcst: -77:1438903:d=2021032312:RH:0.4 mb:26 hour fcst: -78:1447906:d=2021032312:SPFH:0.4 mb:26 hour fcst: -79:1467996:d=2021032312:VVEL:0.4 mb:26 hour fcst: -80:1492768:d=2021032312:DZDT:0.4 mb:26 hour fcst: -81:1523367:d=2021032312:UGRD:0.4 mb:26 hour fcst: -82:1542375:d=2021032312:VGRD:0.4 mb:26 hour fcst: -83:1560568:d=2021032312:ABSV:0.4 mb:26 hour fcst: -84:1577318:d=2021032312:O3MR:0.4 mb:26 hour fcst: -85:1602104:d=2021032312:HGT:0.7 mb:26 hour fcst: -86:1620179:d=2021032312:TMP:0.7 mb:26 hour fcst: -87:1635898:d=2021032312:RH:0.7 mb:26 hour fcst: -88:1644805:d=2021032312:SPFH:0.7 mb:26 hour fcst: -89:1664059:d=2021032312:VVEL:0.7 mb:26 hour fcst: -90:1689933:d=2021032312:DZDT:0.7 mb:26 hour fcst: -91:1719194:d=2021032312:UGRD:0.7 mb:26 hour fcst: -92:1737711:d=2021032312:VGRD:0.7 mb:26 hour fcst: -93:1755538:d=2021032312:ABSV:0.7 mb:26 hour fcst: -94:1771947:d=2021032312:O3MR:0.7 mb:26 hour fcst: -95:1792399:d=2021032312:HGT:1 mb:26 hour fcst: -96:1810348:d=2021032312:TMP:1 mb:26 hour fcst: -97:1826153:d=2021032312:RH:1 mb:26 hour fcst: -98:1835931:d=2021032312:SPFH:1 mb:26 hour fcst: -99:1854494:d=2021032312:VVEL:1 mb:26 hour fcst: -100:1881231:d=2021032312:DZDT:1 mb:26 hour fcst: -101:1907188:d=2021032312:UGRD:1 mb:26 hour fcst: -102:1925643:d=2021032312:VGRD:1 mb:26 hour fcst: -103:1943275:d=2021032312:ABSV:1 mb:26 hour fcst: -104:1959564:d=2021032312:O3MR:1 mb:26 hour fcst: -105:1980829:d=2021032312:HGT:2 mb:26 hour fcst: -106:2001034:d=2021032312:TMP:2 mb:26 hour fcst: -107:2017420:d=2021032312:RH:2 mb:26 hour fcst: -108:2025782:d=2021032312:SPFH:2 mb:26 hour fcst: -109:2044810:d=2021032312:VVEL:2 mb:26 hour fcst: -110:2073399:d=2021032312:DZDT:2 mb:26 hour fcst: -111:2100103:d=2021032312:UGRD:2 mb:26 hour fcst: -112:2111591:d=2021032312:VGRD:2 mb:26 hour fcst: -113:2122295:d=2021032312:ABSV:2 mb:26 hour fcst: -114:2139303:d=2021032312:O3MR:2 mb:26 hour fcst: -115:2161814:d=2021032312:HGT:3 mb:26 hour fcst: -116:2181755:d=2021032312:TMP:3 mb:26 hour fcst: -117:2198165:d=2021032312:RH:3 mb:26 hour fcst: -118:2204227:d=2021032312:SPFH:3 mb:26 hour fcst: -119:2225416:d=2021032312:VVEL:3 mb:26 hour fcst: -120:2255450:d=2021032312:DZDT:3 mb:26 hour fcst: -121:2281634:d=2021032312:UGRD:3 mb:26 hour fcst: -122:2293165:d=2021032312:VGRD:3 mb:26 hour fcst: -123:2311127:d=2021032312:ABSV:3 mb:26 hour fcst: -124:2328020:d=2021032312:O3MR:3 mb:26 hour fcst: -125:2351352:d=2021032312:HGT:5 mb:26 hour fcst: -126:2369091:d=2021032312:TMP:5 mb:26 hour fcst: -127:2385725:d=2021032312:RH:5 mb:26 hour fcst: -128:2394902:d=2021032312:SPFH:5 mb:26 hour fcst: -129:2415697:d=2021032312:VVEL:5 mb:26 hour fcst: -130:2444836:d=2021032312:DZDT:5 mb:26 hour fcst: -131:2470824:d=2021032312:UGRD:5 mb:26 hour fcst: -132:2482323:d=2021032312:VGRD:5 mb:26 hour fcst: -133:2501178:d=2021032312:ABSV:5 mb:26 hour fcst: -134:2518790:d=2021032312:O3MR:5 mb:26 hour fcst: -135:2541415:d=2021032312:HGT:7 mb:26 hour fcst: -136:2559073:d=2021032312:TMP:7 mb:26 hour fcst: -137:2575215:d=2021032312:RH:7 mb:26 hour fcst: -138:2581275:d=2021032312:SPFH:7 mb:26 hour fcst: -139:2600668:d=2021032312:VVEL:7 mb:26 hour fcst: -140:2625275:d=2021032312:DZDT:7 mb:26 hour fcst: -141:2650786:d=2021032312:UGRD:7 mb:26 hour fcst: -142:2662180:d=2021032312:VGRD:7 mb:26 hour fcst: -143:2680996:d=2021032312:ABSV:7 mb:26 hour fcst: -144:2698727:d=2021032312:O3MR:7 mb:26 hour fcst: -145:2722318:d=2021032312:HGT:10 mb:26 hour fcst: -146:2742122:d=2021032312:TMP:10 mb:26 hour fcst: -147:2758273:d=2021032312:RH:10 mb:26 hour fcst: -148:2765773:d=2021032312:SPFH:10 mb:26 hour fcst: -149:2784037:d=2021032312:VVEL:10 mb:26 hour fcst: -150:2809556:d=2021032312:DZDT:10 mb:26 hour fcst: -151:2834831:d=2021032312:UGRD:10 mb:26 hour fcst: -152:2846175:d=2021032312:VGRD:10 mb:26 hour fcst: -153:2864695:d=2021032312:ABSV:10 mb:26 hour fcst: -154:2882348:d=2021032312:O3MR:10 mb:26 hour fcst: -155:2901030:d=2021032312:HGT:15 mb:26 hour fcst: -156:2920588:d=2021032312:TMP:15 mb:26 hour fcst: -157:2936689:d=2021032312:RH:15 mb:26 hour fcst: -158:2941958:d=2021032312:SPFH:15 mb:26 hour fcst: -159:2957387:d=2021032312:VVEL:15 mb:26 hour fcst: -160:2983910:d=2021032312:DZDT:15 mb:26 hour fcst: -161:3014231:d=2021032312:UGRD:15 mb:26 hour fcst: -162:3025432:d=2021032312:VGRD:15 mb:26 hour fcst: -163:3043896:d=2021032312:ABSV:15 mb:26 hour fcst: -164:3061531:d=2021032312:O3MR:15 mb:26 hour fcst: -165:3080421:d=2021032312:HGT:20 mb:26 hour fcst: -166:3099777:d=2021032312:TMP:20 mb:26 hour fcst: -167:3115689:d=2021032312:RH:20 mb:26 hour fcst: -168:3121849:d=2021032312:SPFH:20 mb:26 hour fcst: -169:3136082:d=2021032312:VVEL:20 mb:26 hour fcst: -170:3163523:d=2021032312:DZDT:20 mb:26 hour fcst: -171:3193853:d=2021032312:UGRD:20 mb:26 hour fcst: -172:3212511:d=2021032312:VGRD:20 mb:26 hour fcst: -173:3231026:d=2021032312:ABSV:20 mb:26 hour fcst: -174:3248539:d=2021032312:O3MR:20 mb:26 hour fcst: -175:3273277:d=2021032312:HGT:30 mb:26 hour fcst: -176:3292339:d=2021032312:TMP:30 mb:26 hour fcst: -177:3308240:d=2021032312:RH:30 mb:26 hour fcst: -178:3315834:d=2021032312:SPFH:30 mb:26 hour fcst: -179:3331141:d=2021032312:VVEL:30 mb:26 hour fcst: -180:3359465:d=2021032312:DZDT:30 mb:26 hour fcst: -181:3389355:d=2021032312:UGRD:30 mb:26 hour fcst: -182:3407928:d=2021032312:VGRD:30 mb:26 hour fcst: -183:3426434:d=2021032312:ABSV:30 mb:26 hour fcst: -184:3444111:d=2021032312:O3MR:30 mb:26 hour fcst: -185:3470823:d=2021032312:HGT:40 mb:26 hour fcst: -186:3489699:d=2021032312:TMP:40 mb:26 hour fcst: -187:3505992:d=2021032312:RH:40 mb:26 hour fcst: -188:3515450:d=2021032312:SPFH:40 mb:26 hour fcst: -189:3532482:d=2021032312:VVEL:40 mb:26 hour fcst: -190:3561813:d=2021032312:DZDT:40 mb:26 hour fcst: -191:3594033:d=2021032312:UGRD:40 mb:26 hour fcst: -192:3612743:d=2021032312:VGRD:40 mb:26 hour fcst: -193:3631505:d=2021032312:ABSV:40 mb:26 hour fcst: -194:3649361:d=2021032312:O3MR:40 mb:26 hour fcst: -195:3676305:d=2021032312:HGT:50 mb:26 hour fcst: -196:3695240:d=2021032312:TMP:50 mb:26 hour fcst: -197:3711290:d=2021032312:RH:50 mb:26 hour fcst: -198:3721921:d=2021032312:TCDC:50 mb:26 hour fcst: -199:3722100:d=2021032312:SPFH:50 mb:26 hour fcst: -200:3740575:d=2021032312:VVEL:50 mb:26 hour fcst: -201:3770548:d=2021032312:DZDT:50 mb:26 hour fcst: -202:3802685:d=2021032312:UGRD:50 mb:26 hour fcst: -203:3821122:d=2021032312:VGRD:50 mb:26 hour fcst: -204:3839597:d=2021032312:ABSV:50 mb:26 hour fcst: -205:3857267:d=2021032312:CLMR:50 mb:26 hour fcst: -206:3857446:d=2021032312:ICMR:50 mb:26 hour fcst: -207:3857625:d=2021032312:RWMR:50 mb:26 hour fcst: -208:3857804:d=2021032312:SNMR:50 mb:26 hour fcst: -209:3857983:d=2021032312:GRLE:50 mb:26 hour fcst: -210:3858162:d=2021032312:O3MR:50 mb:26 hour fcst: -211:3884529:d=2021032312:HGT:70 mb:26 hour fcst: -212:3903567:d=2021032312:TMP:70 mb:26 hour fcst: -213:3920070:d=2021032312:RH:70 mb:26 hour fcst: -215:3928327:d=2021032312:SPFH:70 mb:26 hour fcst: -216:3949172:d=2021032312:VVEL:70 mb:26 hour fcst: -217:3980380:d=2021032312:DZDT:70 mb:26 hour fcst: -218:4012479:d=2021032312:UGRD:70 mb:26 hour fcst: -219:4031059:d=2021032312:VGRD:70 mb:26 hour fcst: -220:4049823:d=2021032312:ABSV:70 mb:26 hour fcst: -226:4068416:d=2021032312:O3MR:70 mb:26 hour fcst: -227:4094284:d=2021032312:HGT:100 mb:26 hour fcst: -228:4113890:d=2021032312:TMP:100 mb:26 hour fcst: -229:4130735:d=2021032312:RH:100 mb:26 hour fcst: -230:4140853:d=2021032312:TCDC:100 mb:26 hour fcst: -231:4141032:d=2021032312:SPFH:100 mb:26 hour fcst: -232:4166599:d=2021032312:VVEL:100 mb:26 hour fcst: -233:4191480:d=2021032312:DZDT:100 mb:26 hour fcst: -234:4221735:d=2021032312:UGRD:100 mb:26 hour fcst: -235:4241097:d=2021032312:VGRD:100 mb:26 hour fcst: -236:4260517:d=2021032312:ABSV:100 mb:26 hour fcst: -237:4278695:d=2021032312:CLMR:100 mb:26 hour fcst: -238:4278874:d=2021032312:ICMR:100 mb:26 hour fcst: -239:4279785:d=2021032312:RWMR:100 mb:26 hour fcst: -240:4279964:d=2021032312:SNMR:100 mb:26 hour fcst: -241:4282217:d=2021032312:GRLE:100 mb:26 hour fcst: -242:4282396:d=2021032312:O3MR:100 mb:26 hour fcst: -259:4523749:d=2021032312:HGT:150 mb:26 hour fcst: -260:4544140:d=2021032312:TMP:150 mb:26 hour fcst: -261:4561110:d=2021032312:RH:150 mb:26 hour fcst: -262:4572505:d=2021032312:TCDC:150 mb:26 hour fcst: -263:4576543:d=2021032312:SPFH:150 mb:26 hour fcst: -264:4600796:d=2021032312:VVEL:150 mb:26 hour fcst: -265:4628153:d=2021032312:DZDT:150 mb:26 hour fcst: -266:4653747:d=2021032312:UGRD:150 mb:26 hour fcst: -267:4673749:d=2021032312:VGRD:150 mb:26 hour fcst: -268:4693707:d=2021032312:ABSV:150 mb:26 hour fcst: -269:4712655:d=2021032312:CLMR:150 mb:26 hour fcst: -270:4712834:d=2021032312:ICMR:150 mb:26 hour fcst: -271:4718543:d=2021032312:RWMR:150 mb:26 hour fcst: -272:4718722:d=2021032312:SNMR:150 mb:26 hour fcst: -273:4722280:d=2021032312:GRLE:150 mb:26 hour fcst: -274:4722799:d=2021032312:O3MR:150 mb:26 hour fcst: -291:4968240:d=2021032312:HGT:200 mb:26 hour fcst: -292:4989303:d=2021032312:TMP:200 mb:26 hour fcst: -293:5007448:d=2021032312:RH:200 mb:26 hour fcst: -294:5022478:d=2021032312:TCDC:200 mb:26 hour fcst: -295:5029197:d=2021032312:SPFH:200 mb:26 hour fcst: -296:5054952:d=2021032312:VVEL:200 mb:26 hour fcst: -297:5081751:d=2021032312:DZDT:200 mb:26 hour fcst: -298:5108180:d=2021032312:UGRD:200 mb:26 hour fcst: -299:5121568:d=2021032312:VGRD:200 mb:26 hour fcst: -300:5135299:d=2021032312:ABSV:200 mb:26 hour fcst: -301:5155294:d=2021032312:CLMR:200 mb:26 hour fcst: -302:5155473:d=2021032312:ICMR:200 mb:26 hour fcst: -303:5164122:d=2021032312:RWMR:200 mb:26 hour fcst: -304:5164301:d=2021032312:SNMR:200 mb:26 hour fcst: -305:5169753:d=2021032312:GRLE:200 mb:26 hour fcst: -306:5170082:d=2021032312:O3MR:200 mb:26 hour fcst: -323:5426750:d=2021032312:HGT:250 mb:26 hour fcst: -324:5448485:d=2021032312:TMP:250 mb:26 hour fcst: -325:5467319:d=2021032312:RH:250 mb:26 hour fcst: -326:5485141:d=2021032312:TCDC:250 mb:26 hour fcst: -327:5495853:d=2021032312:SPFH:250 mb:26 hour fcst: -328:5523933:d=2021032312:VVEL:250 mb:26 hour fcst: -329:5546667:d=2021032312:DZDT:250 mb:26 hour fcst: -330:5574059:d=2021032312:UGRD:250 mb:26 hour fcst: -331:5588198:d=2021032312:VGRD:250 mb:26 hour fcst: -332:5603054:d=2021032312:ABSV:250 mb:26 hour fcst: -333:5624656:d=2021032312:CLMR:250 mb:26 hour fcst: -334:5624835:d=2021032312:ICMR:250 mb:26 hour fcst: -335:5639137:d=2021032312:RWMR:250 mb:26 hour fcst: -336:5639316:d=2021032312:SNMR:250 mb:26 hour fcst: -337:5648203:d=2021032312:GRLE:250 mb:26 hour fcst: -338:5648509:d=2021032312:O3MR:250 mb:26 hour fcst: -355:5926338:d=2021032312:HGT:300 mb:26 hour fcst: -356:5948657:d=2021032312:TMP:300 mb:26 hour fcst: -357:5967072:d=2021032312:RH:300 mb:26 hour fcst: -358:5986050:d=2021032312:TCDC:300 mb:26 hour fcst: -359:5998834:d=2021032312:SPFH:300 mb:26 hour fcst: -360:6025422:d=2021032312:VVEL:300 mb:26 hour fcst: -361:6049273:d=2021032312:DZDT:300 mb:26 hour fcst: -362:6077293:d=2021032312:UGRD:300 mb:26 hour fcst: -363:6091933:d=2021032312:VGRD:300 mb:26 hour fcst: -364:6107319:d=2021032312:ABSV:300 mb:26 hour fcst: -365:6129649:d=2021032312:CLMR:300 mb:26 hour fcst: -366:6129828:d=2021032312:ICMR:300 mb:26 hour fcst: -367:6145683:d=2021032312:RWMR:300 mb:26 hour fcst: -368:6145862:d=2021032312:SNMR:300 mb:26 hour fcst: -369:6156265:d=2021032312:GRLE:300 mb:26 hour fcst: -370:6156646:d=2021032312:O3MR:300 mb:26 hour fcst: -387:6444168:d=2021032312:HGT:350 mb:26 hour fcst: -388:6466332:d=2021032312:TMP:350 mb:26 hour fcst: -389:6484544:d=2021032312:RH:350 mb:26 hour fcst: -390:6503451:d=2021032312:TCDC:350 mb:26 hour fcst: -391:6516713:d=2021032312:SPFH:350 mb:26 hour fcst: -392:6544864:d=2021032312:VVEL:350 mb:26 hour fcst: -393:6569360:d=2021032312:DZDT:350 mb:26 hour fcst: -394:6597629:d=2021032312:UGRD:350 mb:26 hour fcst: -395:6612212:d=2021032312:VGRD:350 mb:26 hour fcst: -396:6627457:d=2021032312:ABSV:350 mb:26 hour fcst: -397:6649750:d=2021032312:CLMR:350 mb:26 hour fcst: -398:6649929:d=2021032312:ICMR:350 mb:26 hour fcst: -399:6666928:d=2021032312:RWMR:350 mb:26 hour fcst: -400:6667107:d=2021032312:SNMR:350 mb:26 hour fcst: -401:6679061:d=2021032312:GRLE:350 mb:26 hour fcst: -402:6679552:d=2021032312:O3MR:350 mb:26 hour fcst: -419:6965625:d=2021032312:HGT:400 mb:26 hour fcst: -420:6987559:d=2021032312:TMP:400 mb:26 hour fcst: -421:7005938:d=2021032312:RH:400 mb:26 hour fcst: -422:7025140:d=2021032312:TCDC:400 mb:26 hour fcst: -423:7038435:d=2021032312:SPFH:400 mb:26 hour fcst: -424:7066195:d=2021032312:VVEL:400 mb:26 hour fcst: -425:7091008:d=2021032312:DZDT:400 mb:26 hour fcst: -426:7119108:d=2021032312:UGRD:400 mb:26 hour fcst: -427:7141151:d=2021032312:VGRD:400 mb:26 hour fcst: -428:7163810:d=2021032312:ABSV:400 mb:26 hour fcst: -429:7185782:d=2021032312:CLMR:400 mb:26 hour fcst: -430:7186184:d=2021032312:ICMR:400 mb:26 hour fcst: -431:7203261:d=2021032312:RWMR:400 mb:26 hour fcst: -432:7203639:d=2021032312:SNMR:400 mb:26 hour fcst: -433:7217585:d=2021032312:GRLE:400 mb:26 hour fcst: -434:7218625:d=2021032312:O3MR:400 mb:26 hour fcst: -451:7523748:d=2021032312:HGT:450 mb:26 hour fcst: -452:7547667:d=2021032312:TMP:450 mb:26 hour fcst: -453:7566065:d=2021032312:RH:450 mb:26 hour fcst: -454:7585431:d=2021032312:TCDC:450 mb:26 hour fcst: -455:7598027:d=2021032312:SPFH:450 mb:26 hour fcst: -456:7627311:d=2021032312:VVEL:450 mb:26 hour fcst: -457:7652461:d=2021032312:DZDT:450 mb:26 hour fcst: -458:7680479:d=2021032312:UGRD:450 mb:26 hour fcst: -459:7702109:d=2021032312:VGRD:450 mb:26 hour fcst: -460:7724388:d=2021032312:ABSV:450 mb:26 hour fcst: -461:7745816:d=2021032312:CLMR:450 mb:26 hour fcst: -462:7746676:d=2021032312:ICMR:450 mb:26 hour fcst: -463:7763416:d=2021032312:RWMR:450 mb:26 hour fcst: -464:7763936:d=2021032312:SNMR:450 mb:26 hour fcst: -465:7777542:d=2021032312:GRLE:450 mb:26 hour fcst: -466:7778950:d=2021032312:O3MR:450 mb:26 hour fcst: -483:8080613:d=2021032312:HGT:500 mb:26 hour fcst: -484:8104341:d=2021032312:TMP:500 mb:26 hour fcst: -485:8123035:d=2021032312:RH:500 mb:26 hour fcst: -486:8142570:d=2021032312:TCDC:500 mb:26 hour fcst: -487:8154871:d=2021032312:SPFH:500 mb:26 hour fcst: -488:8183126:d=2021032312:VVEL:500 mb:26 hour fcst: -489:8208367:d=2021032312:DZDT:500 mb:26 hour fcst: -490:8236023:d=2021032312:UGRD:500 mb:26 hour fcst: -491:8257495:d=2021032312:VGRD:500 mb:26 hour fcst: -492:8279495:d=2021032312:ABSV:500 mb:26 hour fcst: -493:8300613:d=2021032312:CLMR:500 mb:26 hour fcst: -494:8302241:d=2021032312:ICMR:500 mb:26 hour fcst: -495:8318352:d=2021032312:RWMR:500 mb:26 hour fcst: -496:8319314:d=2021032312:SNMR:500 mb:26 hour fcst: -497:8333595:d=2021032312:GRLE:500 mb:26 hour fcst: -498:8335480:d=2021032312:O3MR:500 mb:26 hour fcst: -515:8638819:d=2021032312:HGT:550 mb:26 hour fcst: -516:8662354:d=2021032312:TMP:550 mb:26 hour fcst: -517:8681201:d=2021032312:RH:550 mb:26 hour fcst: -518:8700962:d=2021032312:TCDC:550 mb:26 hour fcst: -519:8713156:d=2021032312:SPFH:550 mb:26 hour fcst: -520:8742665:d=2021032312:VVEL:550 mb:26 hour fcst: -521:8767996:d=2021032312:DZDT:550 mb:26 hour fcst: -522:8795543:d=2021032312:UGRD:550 mb:26 hour fcst: -523:8816555:d=2021032312:VGRD:550 mb:26 hour fcst: -524:8838386:d=2021032312:ABSV:550 mb:26 hour fcst: -525:8859336:d=2021032312:CLMR:550 mb:26 hour fcst: -526:8862619:d=2021032312:ICMR:550 mb:26 hour fcst: -527:8879314:d=2021032312:RWMR:550 mb:26 hour fcst: -528:8880693:d=2021032312:SNMR:550 mb:26 hour fcst: -529:8895201:d=2021032312:GRLE:550 mb:26 hour fcst: -530:8897124:d=2021032312:O3MR:550 mb:26 hour fcst: -547:9207422:d=2021032312:HGT:600 mb:26 hour fcst: -548:9231574:d=2021032312:TMP:600 mb:26 hour fcst: -549:9250624:d=2021032312:RH:600 mb:26 hour fcst: -550:9270642:d=2021032312:TCDC:600 mb:26 hour fcst: -551:9282720:d=2021032312:SPFH:600 mb:26 hour fcst: -552:9312984:d=2021032312:VVEL:600 mb:26 hour fcst: -553:9338389:d=2021032312:DZDT:600 mb:26 hour fcst: -554:9365804:d=2021032312:UGRD:600 mb:26 hour fcst: -555:9386689:d=2021032312:VGRD:600 mb:26 hour fcst: -556:9408306:d=2021032312:ABSV:600 mb:26 hour fcst: -557:9429112:d=2021032312:CLMR:600 mb:26 hour fcst: -558:9433948:d=2021032312:ICMR:600 mb:26 hour fcst: -559:9447852:d=2021032312:RWMR:600 mb:26 hour fcst: -560:9451573:d=2021032312:SNMR:600 mb:26 hour fcst: -561:9467361:d=2021032312:GRLE:600 mb:26 hour fcst: -562:9469587:d=2021032312:O3MR:600 mb:26 hour fcst: -579:9785513:d=2021032312:HGT:650 mb:26 hour fcst: -580:9809613:d=2021032312:TMP:650 mb:26 hour fcst: -581:9828931:d=2021032312:RH:650 mb:26 hour fcst: -582:9849157:d=2021032312:TCDC:650 mb:26 hour fcst: -583:9860895:d=2021032312:SPFH:650 mb:26 hour fcst: -584:9889603:d=2021032312:VVEL:650 mb:26 hour fcst: -585:9915217:d=2021032312:DZDT:650 mb:26 hour fcst: -586:9942642:d=2021032312:UGRD:650 mb:26 hour fcst: -587:9963739:d=2021032312:VGRD:650 mb:26 hour fcst: -588:9985300:d=2021032312:ABSV:650 mb:26 hour fcst: -589:10006179:d=2021032312:CLMR:650 mb:26 hour fcst: -590:10011208:d=2021032312:ICMR:650 mb:26 hour fcst: -591:10023611:d=2021032312:RWMR:650 mb:26 hour fcst: -592:10028329:d=2021032312:SNMR:650 mb:26 hour fcst: -593:10043207:d=2021032312:GRLE:650 mb:26 hour fcst: -594:10045520:d=2021032312:O3MR:650 mb:26 hour fcst: -611:10362688:d=2021032312:HGT:700 mb:26 hour fcst: -612:10386873:d=2021032312:TMP:700 mb:26 hour fcst: -613:10406645:d=2021032312:RH:700 mb:26 hour fcst: -614:10426905:d=2021032312:TCDC:700 mb:26 hour fcst: -615:10438497:d=2021032312:SPFH:700 mb:26 hour fcst: -616:10467066:d=2021032312:VVEL:700 mb:26 hour fcst: -617:10492786:d=2021032312:DZDT:700 mb:26 hour fcst: -618:10520147:d=2021032312:UGRD:700 mb:26 hour fcst: -619:10541032:d=2021032312:VGRD:700 mb:26 hour fcst: -620:10562631:d=2021032312:ABSV:700 mb:26 hour fcst: -621:10583667:d=2021032312:CLMR:700 mb:26 hour fcst: -622:10590533:d=2021032312:ICMR:700 mb:26 hour fcst: -623:10601811:d=2021032312:RWMR:700 mb:26 hour fcst: -624:10607524:d=2021032312:SNMR:700 mb:26 hour fcst: -625:10621564:d=2021032312:GRLE:700 mb:26 hour fcst: -626:10623441:d=2021032312:O3MR:700 mb:26 hour fcst: -643:10940753:d=2021032312:HGT:750 mb:26 hour fcst: -644:10965158:d=2021032312:TMP:750 mb:26 hour fcst: -645:10985096:d=2021032312:RH:750 mb:26 hour fcst: -646:11005541:d=2021032312:TCDC:750 mb:26 hour fcst: -647:11017764:d=2021032312:SPFH:750 mb:26 hour fcst: -648:11046737:d=2021032312:VVEL:750 mb:26 hour fcst: -649:11072521:d=2021032312:DZDT:750 mb:26 hour fcst: -650:11099764:d=2021032312:UGRD:750 mb:26 hour fcst: -651:11120624:d=2021032312:VGRD:750 mb:26 hour fcst: -652:11142375:d=2021032312:ABSV:750 mb:26 hour fcst: -653:11163520:d=2021032312:CLMR:750 mb:26 hour fcst: -654:11172853:d=2021032312:ICMR:750 mb:26 hour fcst: -655:11184296:d=2021032312:RWMR:750 mb:26 hour fcst: -656:11190169:d=2021032312:SNMR:750 mb:26 hour fcst: -657:11203776:d=2021032312:GRLE:750 mb:26 hour fcst: -658:11205864:d=2021032312:O3MR:750 mb:26 hour fcst: -675:11527815:d=2021032312:HGT:800 mb:26 hour fcst: -676:11552150:d=2021032312:TMP:800 mb:26 hour fcst: -677:11572563:d=2021032312:RH:800 mb:26 hour fcst: -678:11593001:d=2021032312:TCDC:800 mb:26 hour fcst: -679:11606505:d=2021032312:SPFH:800 mb:26 hour fcst: -680:11636036:d=2021032312:VVEL:800 mb:26 hour fcst: -681:11661746:d=2021032312:DZDT:800 mb:26 hour fcst: -682:11688766:d=2021032312:UGRD:800 mb:26 hour fcst: -683:11709710:d=2021032312:VGRD:800 mb:26 hour fcst: -684:11731462:d=2021032312:ABSV:800 mb:26 hour fcst: -685:11752674:d=2021032312:CLMR:800 mb:26 hour fcst: -686:11764224:d=2021032312:ICMR:800 mb:26 hour fcst: -687:11775255:d=2021032312:RWMR:800 mb:26 hour fcst: -688:11783503:d=2021032312:SNMR:800 mb:26 hour fcst: -689:11796360:d=2021032312:GRLE:800 mb:26 hour fcst: -690:11798726:d=2021032312:O3MR:800 mb:26 hour fcst: -707:12125925:d=2021032312:HGT:850 mb:26 hour fcst: -708:12150705:d=2021032312:TMP:850 mb:26 hour fcst: -709:12171582:d=2021032312:RH:850 mb:26 hour fcst: -710:12191840:d=2021032312:TCDC:850 mb:26 hour fcst: -711:12206564:d=2021032312:SPFH:850 mb:26 hour fcst: -712:12236119:d=2021032312:VVEL:850 mb:26 hour fcst: -713:12261637:d=2021032312:DZDT:850 mb:26 hour fcst: -714:12288207:d=2021032312:UGRD:850 mb:26 hour fcst: -715:12309450:d=2021032312:VGRD:850 mb:26 hour fcst: -716:12331275:d=2021032312:ABSV:850 mb:26 hour fcst: -717:12352514:d=2021032312:CLMR:850 mb:26 hour fcst: -718:12365575:d=2021032312:ICMR:850 mb:26 hour fcst: -719:12375405:d=2021032312:RWMR:850 mb:26 hour fcst: -720:12385816:d=2021032312:SNMR:850 mb:26 hour fcst: -721:12397472:d=2021032312:GRLE:850 mb:26 hour fcst: -722:12399807:d=2021032312:O3MR:850 mb:26 hour fcst: -739:12724257:d=2021032312:HGT:900 mb:26 hour fcst: -740:12749315:d=2021032312:TMP:900 mb:26 hour fcst: -741:12770288:d=2021032312:RH:900 mb:26 hour fcst: -742:12790141:d=2021032312:TCDC:900 mb:26 hour fcst: -743:12803008:d=2021032312:SPFH:900 mb:26 hour fcst: -744:12832495:d=2021032312:VVEL:900 mb:26 hour fcst: -745:12857388:d=2021032312:DZDT:900 mb:26 hour fcst: -746:12883468:d=2021032312:UGRD:900 mb:26 hour fcst: -747:12904818:d=2021032312:VGRD:900 mb:26 hour fcst: -748:12926739:d=2021032312:ABSV:900 mb:26 hour fcst: -749:12947963:d=2021032312:CLMR:900 mb:26 hour fcst: -750:12957958:d=2021032312:ICMR:900 mb:26 hour fcst: -751:12966393:d=2021032312:RWMR:900 mb:26 hour fcst: -752:12978396:d=2021032312:SNMR:900 mb:26 hour fcst: -753:12988637:d=2021032312:GRLE:900 mb:26 hour fcst: -754:12991376:d=2021032312:O3MR:900 mb:26 hour fcst: -755:13017420:d=2021032312:HGT:925 mb:26 hour fcst: -756:13042637:d=2021032312:TMP:925 mb:26 hour fcst: -757:13063794:d=2021032312:RH:925 mb:26 hour fcst: -758:13083601:d=2021032312:TCDC:925 mb:26 hour fcst: -759:13095404:d=2021032312:SPFH:925 mb:26 hour fcst: -760:13124947:d=2021032312:VVEL:925 mb:26 hour fcst: -761:13149419:d=2021032312:DZDT:925 mb:26 hour fcst: -762:13175028:d=2021032312:UGRD:925 mb:26 hour fcst: -763:13196527:d=2021032312:VGRD:925 mb:26 hour fcst: -764:13218447:d=2021032312:ABSV:925 mb:26 hour fcst: -765:13239715:d=2021032312:CLMR:925 mb:26 hour fcst: -766:13248229:d=2021032312:ICMR:925 mb:26 hour fcst: -767:13255898:d=2021032312:RWMR:925 mb:26 hour fcst: -768:13268144:d=2021032312:SNMR:925 mb:26 hour fcst: -769:13277739:d=2021032312:GRLE:925 mb:26 hour fcst: -770:13280515:d=2021032312:O3MR:925 mb:26 hour fcst: -771:13306467:d=2021032312:HGT:950 mb:26 hour fcst: -772:13331819:d=2021032312:TMP:950 mb:26 hour fcst: -773:13352863:d=2021032312:RH:950 mb:26 hour fcst: -774:13372384:d=2021032312:TCDC:950 mb:26 hour fcst: -775:13382462:d=2021032312:SPFH:950 mb:26 hour fcst: -776:13409716:d=2021032312:VVEL:950 mb:26 hour fcst: -777:13433559:d=2021032312:DZDT:950 mb:26 hour fcst: -778:13458447:d=2021032312:UGRD:950 mb:26 hour fcst: -779:13479955:d=2021032312:VGRD:950 mb:26 hour fcst: -780:13501895:d=2021032312:ABSV:950 mb:26 hour fcst: -781:13522986:d=2021032312:CLMR:950 mb:26 hour fcst: -782:13529455:d=2021032312:ICMR:950 mb:26 hour fcst: -783:13536577:d=2021032312:RWMR:950 mb:26 hour fcst: -784:13548870:d=2021032312:SNMR:950 mb:26 hour fcst: -785:13557680:d=2021032312:GRLE:950 mb:26 hour fcst: -786:13560228:d=2021032312:O3MR:950 mb:26 hour fcst: -787:13586163:d=2021032312:HINDEX:surface:26 hour fcst: -788:13591263:d=2021032312:HGT:975 mb:26 hour fcst: -789:13616806:d=2021032312:TMP:975 mb:26 hour fcst: -790:13637689:d=2021032312:RH:975 mb:26 hour fcst: -791:13656761:d=2021032312:TCDC:975 mb:26 hour fcst: -792:13664954:d=2021032312:SPFH:975 mb:26 hour fcst: -793:13691833:d=2021032312:VVEL:975 mb:26 hour fcst: -794:13714571:d=2021032312:DZDT:975 mb:26 hour fcst: -795:13738045:d=2021032312:UGRD:975 mb:26 hour fcst: -796:13759443:d=2021032312:VGRD:975 mb:26 hour fcst: -797:13781312:d=2021032312:ABSV:975 mb:26 hour fcst: -798:13802321:d=2021032312:CLMR:975 mb:26 hour fcst: -799:13806885:d=2021032312:ICMR:975 mb:26 hour fcst: -800:13811749:d=2021032312:RWMR:975 mb:26 hour fcst: -801:13823312:d=2021032312:SNMR:975 mb:26 hour fcst: -802:13830420:d=2021032312:GRLE:975 mb:26 hour fcst: -803:13832702:d=2021032312:O3MR:975 mb:26 hour fcst: -804:13858541:d=2021032312:TMP:1000 mb:26 hour fcst: -805:13879336:d=2021032312:RH:1000 mb:26 hour fcst: -806:13898172:d=2021032312:TCDC:1000 mb:26 hour fcst: -807:13902586:d=2021032312:SPFH:1000 mb:26 hour fcst: -808:13929734:d=2021032312:VVEL:1000 mb:26 hour fcst: -809:13950900:d=2021032312:DZDT:1000 mb:26 hour fcst: -810:13972290:d=2021032312:UGRD:1000 mb:26 hour fcst: -811:13993099:d=2021032312:VGRD:1000 mb:26 hour fcst: -812:14014551:d=2021032312:ABSV:1000 mb:26 hour fcst: -813:14035075:d=2021032312:CLMR:1000 mb:26 hour fcst: -814:14036830:d=2021032312:ICMR:1000 mb:26 hour fcst: -815:14039198:d=2021032312:RWMR:1000 mb:26 hour fcst: -816:14048656:d=2021032312:SNMR:1000 mb:26 hour fcst: -817:14051985:d=2021032312:GRLE:1000 mb:26 hour fcst: -818:14053096:d=2021032312:O3MR:1000 mb:26 hour fcst: -819:14078923:d=2021032312:MSLET:mean sea level:26 hour fcst: -820:14105565:d=2021032312:HGT:1000 mb:26 hour fcst: -821:14131301:d=2021032312:SDEN:surface:26 hour fcst: -822:14142290:d=2021032312:REFD:4000 m above ground:26 hour fcst: -823:14155522:d=2021032312:REFD:1000 m above ground:26 hour fcst: -824:14179663:d=2021032312:PRES:surface:26 hour fcst: -825:14205368:d=2021032312:HGT:surface:26 hour fcst: -826:14221438:d=2021032312:TMP:surface:26 hour fcst: -827:14238074:d=2021032312:TSOIL:0-0.1 m below ground:26 hour fcst: -828:14250042:d=2021032312:SOILW:0-0.1 m below ground:26 hour fcst: -829:14258959:d=2021032312:SOILL:0-0.1 m below ground:26 hour fcst: -830:14267759:d=2021032312:TSOIL:0.1-0.4 m below ground:26 hour fcst: -831:14283252:d=2021032312:SOILW:0.1-0.4 m below ground:26 hour fcst: -832:14292266:d=2021032312:SOILL:0.1-0.4 m below ground:26 hour fcst: -833:14301158:d=2021032312:TSOIL:0.4-1 m below ground:26 hour fcst: -834:14315607:d=2021032312:SOILW:0.4-1 m below ground:26 hour fcst: -835:14324740:d=2021032312:SOILL:0.4-1 m below ground:26 hour fcst: -836:14333663:d=2021032312:TSOIL:1-2 m below ground:26 hour fcst: -837:14347953:d=2021032312:SOILW:1-2 m below ground:26 hour fcst: -838:14357010:d=2021032312:SOILL:1-2 m below ground:26 hour fcst: -839:14366087:d=2021032312:CNWAT:surface:26 hour fcst: -840:14371609:d=2021032312:WEASD:surface:26 hour fcst: -841:14384434:d=2021032312:SNOD:surface:26 hour fcst: -842:14398813:d=2021032312:PEVPR:surface:26 hour fcst: -843:14414605:d=2021032312:ICETK:surface:26 hour fcst: -845:14426034:d=2021032312:TMP:2 m above ground:26 hour fcst: -846:14446854:d=2021032312:SPFH:2 m above ground:26 hour fcst: -847:14475668:d=2021032312:DPT:2 m above ground:26 hour fcst: -848:14496930:d=2021032312:RH:2 m above ground:26 hour fcst: -849:14515444:d=2021032312:APTMP:2 m above ground:26 hour fcst: -850:14529843:d=2021032312:TMAX:2 m above ground:24-26 hour max fcst: -851:14550579:d=2021032312:TMIN:2 m above ground:24-26 hour min fcst: -852:14571371:d=2021032312:UGRD:10 m above ground:26 hour fcst: -853:14591916:d=2021032312:VGRD:10 m above ground:26 hour fcst: -854:14612834:d=2021032312:ICEG:10 m above mean sea level:26 hour fcst: -855:14613450:d=2021032312:CPOFP:surface:26 hour fcst: -856:14633796:d=2021032312:CPRAT:surface:26 hour fcst: -857:14655444:d=2021032312:PRATE:surface:26 hour fcst: -858:14682456:d=2021032312:CPRAT:surface:24-26 hour ave fcst: -859:14706099:d=2021032312:PRATE:surface:24-26 hour ave fcst: -860:14731230:d=2021032312:APCP:surface:24-26 hour acc fcst: -861:14738851:d=2021032312:APCP:surface:0-26 hour acc fcst: -862:14752198:d=2021032312:ACPCP:surface:24-26 hour acc fcst: -863:14758312:d=2021032312:ACPCP:surface:0-26 hour acc fcst: -866:14782947:d=2021032312:FROZR:surface:0-26 hour acc fcst: -867:14789334:d=2021032312:FRZR:surface:0-26 hour acc fcst: -868:14791290:d=2021032312:TSNOWP:surface:0-26 hour acc fcst: -869:14799005:d=2021032312:WATR:surface:24-26 hour acc fcst: -870:14811749:d=2021032312:FROZR:surface:24-26 hour acc fcst: -871:14814444:d=2021032312:FRZR:surface:24-26 hour acc fcst: -872:14815168:d=2021032312:TSNOWP:surface:24-26 hour acc fcst: -873:14821730:d=2021032312:CSNOW:surface:26 hour fcst: -874:14823559:d=2021032312:CICEP:surface:26 hour fcst: -875:14823738:d=2021032312:CFRZR:surface:26 hour fcst: -876:14823964:d=2021032312:CRAIN:surface:26 hour fcst: -877:14827478:d=2021032312:LHTFL:surface:24-26 hour ave fcst: -878:14849277:d=2021032312:SHTFL:surface:24-26 hour ave fcst: -879:14870072:d=2021032312:GFLUX:surface:24-26 hour ave fcst: -880:14882137:d=2021032312:UFLX:surface:24-26 hour ave fcst: -881:14898550:d=2021032312:VFLX:surface:24-26 hour ave fcst: -882:14915050:d=2021032312:SFCR:surface:26 hour fcst: -883:14933465:d=2021032312:FRICV:surface:26 hour fcst: -884:14955204:d=2021032312:U-GWD:surface:24-26 hour ave fcst: -885:14970119:d=2021032312:V-GWD:surface:24-26 hour ave fcst: -886:14985084:d=2021032312:VEG:surface:26 hour fcst: -887:14993118:d=2021032312:SOTYP:surface:26 hour fcst: -888:15004167:d=2021032312:WILT:surface:26 hour fcst: -889:15014077:d=2021032312:FLDCP:surface:26 hour fcst: -890:15024283:d=2021032312:SUNSD:surface:26 hour fcst: -891:15034170:d=2021032312:LFTX:surface:26 hour fcst: -892:15047926:d=2021032312:CAPE:surface:26 hour fcst: -893:15063007:d=2021032312:CIN:surface:26 hour fcst: -894:15075951:d=2021032312:PWAT:entire atmosphere (considered as a single layer):26 hour fcst: -895:15102865:d=2021032312:CWAT:entire atmosphere (considered as a single layer):26 hour fcst: -896:15129339:d=2021032312:RH:entire atmosphere (considered as a single layer):26 hour fcst: -897:15144461:d=2021032312:TOZNE:entire atmosphere (considered as a single layer):26 hour fcst: -898:15158776:d=2021032312:LCDC:low cloud layer:26 hour fcst: -899:15176914:d=2021032312:LCDC:low cloud layer:24-26 hour ave fcst: -900:15195610:d=2021032312:MCDC:middle cloud layer:26 hour fcst: -901:15211230:d=2021032312:MCDC:middle cloud layer:24-26 hour ave fcst: -902:15227161:d=2021032312:HCDC:high cloud layer:26 hour fcst: -903:15242011:d=2021032312:HCDC:high cloud layer:24-26 hour ave fcst: -904:15258435:d=2021032312:TCDC:entire atmosphere (considered as a single layer):26 hour fcst: -905:15275486:d=2021032312:TCDC:entire atmosphere (considered as a single layer):24-26 hour ave fcst: -906:15293548:d=2021032312:HGT:cloud ceiling:26 hour fcst: -907:15325528:d=2021032312:PRES:convective cloud bottom level:26 hour fcst: -908:15339166:d=2021032312:PRES:low cloud bottom level:24-26 hour ave fcst: -909:15360775:d=2021032312:PRES:middle cloud bottom level:24-26 hour ave fcst: -910:15377578:d=2021032312:PRES:high cloud bottom level:24-26 hour ave fcst: -912:15415576:d=2021032312:PRES:convective cloud top level:26 hour fcst: -913:15430902:d=2021032312:PRES:low cloud top level:24-26 hour ave fcst: -914:15452420:d=2021032312:PRES:middle cloud top level:24-26 hour ave fcst: -915:15469065:d=2021032312:PRES:high cloud top level:24-26 hour ave fcst: -916:15486996:d=2021032312:TMP:low cloud top level:24-26 hour ave fcst: -917:15503222:d=2021032312:TMP:middle cloud top level:24-26 hour ave fcst: -918:15516524:d=2021032312:TMP:high cloud top level:24-26 hour ave fcst: -919:15531465:d=2021032312:TCDC:convective cloud layer:26 hour fcst: -920:15551909:d=2021032312:TCDC:boundary layer cloud layer:24-26 hour ave fcst: -921:15568046:d=2021032312:CWORK:entire atmosphere (considered as a single layer):24-26 hour ave fcst: -922:15580917:d=2021032312:DSWRF:surface:24-26 hour ave fcst: -925:15630751:d=2021032312:DLWRF:surface:24-26 hour ave fcst: -926:15652559:d=2021032312:USWRF:surface:24-26 hour ave fcst: -927:15670727:d=2021032312:ULWRF:surface:24-26 hour ave fcst: -928:15689610:d=2021032312:USWRF:top of atmosphere:24-26 hour ave fcst: -929:15708616:d=2021032312:ULWRF:top of atmosphere:24-26 hour ave fcst: -930:15729708:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Total Aerosol:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -931:15757200:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Dust Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -932:15775336:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Sea Salt Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -933:15802833:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Sulphate Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -934:15828756:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Particulate Organic Matter Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -935:15854386:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Black Carbon Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -936:15876704:d=2021032312:HLCY:3000-0 m above ground:26 hour fcst: -937:15892563:d=2021032312:HLCY:1000-0 m above ground:26 hour fcst: -938:15914841:d=2021032312:USTM:6000-0 m above ground:26 hour fcst: -939:15937883:d=2021032312:VSTM:6000-0 m above ground:26 hour fcst: -940:15960687:d=2021032312:PRES:tropopause:26 hour fcst: -941:15990419:d=2021032312:ICAHT:tropopause:26 hour fcst: -942:16021219:d=2021032312:HGT:tropopause:26 hour fcst: -943:16049350:d=2021032312:TMP:tropopause:26 hour fcst: -944:16071015:d=2021032312:UGRD:tropopause:26 hour fcst: -945:16086965:d=2021032312:VGRD:tropopause:26 hour fcst: -946:16102873:d=2021032312:VWSH:tropopause:26 hour fcst: -947:16121596:d=2021032312:PRES:max wind:26 hour fcst: -948:16156014:d=2021032312:ICAHT:max wind:26 hour fcst: -949:16191059:d=2021032312:HGT:max wind:26 hour fcst: -950:16226257:d=2021032312:UGRD:max wind:26 hour fcst: -951:16242706:d=2021032312:VGRD:max wind:26 hour fcst: -952:16260130:d=2021032312:TMP:max wind:26 hour fcst: -953:16286786:d=2021032312:UGRD:20 m above ground:26 hour fcst: -954:16307306:d=2021032312:VGRD:20 m above ground:26 hour fcst: -955:16328244:d=2021032312:UGRD:30 m above ground:26 hour fcst: -956:16349191:d=2021032312:VGRD:30 m above ground:26 hour fcst: -957:16370534:d=2021032312:UGRD:40 m above ground:26 hour fcst: -958:16391627:d=2021032312:VGRD:40 m above ground:26 hour fcst: -959:16412885:d=2021032312:UGRD:50 m above ground:26 hour fcst: -960:16433872:d=2021032312:VGRD:50 m above ground:26 hour fcst: -961:16455394:d=2021032312:TMP:80 m above ground:26 hour fcst: -962:16476190:d=2021032312:SPFH:80 m above ground:26 hour fcst: -963:16504935:d=2021032312:PRES:80 m above ground:26 hour fcst: -964:16530595:d=2021032312:UGRD:80 m above ground:26 hour fcst: -965:16551751:d=2021032312:VGRD:80 m above ground:26 hour fcst: -966:16573459:d=2021032312:TMP:100 m above ground:26 hour fcst: -967:16594228:d=2021032312:UGRD:100 m above ground:26 hour fcst: -968:16615625:d=2021032312:VGRD:100 m above ground:26 hour fcst: -981:16881073:d=2021032312:TMP:1829 m above mean sea level:26 hour fcst: -982:16902684:d=2021032312:UGRD:1829 m above mean sea level:26 hour fcst: -983:16924938:d=2021032312:VGRD:1829 m above mean sea level:26 hour fcst: -984:16947799:d=2021032312:TMP:2743 m above mean sea level:26 hour fcst: -985:16969436:d=2021032312:UGRD:2743 m above mean sea level:26 hour fcst: -986:16992132:d=2021032312:VGRD:2743 m above mean sea level:26 hour fcst: -987:17015407:d=2021032312:TMP:3658 m above mean sea level:26 hour fcst: -988:17036754:d=2021032312:UGRD:3658 m above mean sea level:26 hour fcst: -989:17059816:d=2021032312:VGRD:3658 m above mean sea level:26 hour fcst: -993:17152020:d=2021032312:HGT:0C isotherm:26 hour fcst: -994:17173630:d=2021032312:RH:0C isotherm:26 hour fcst: -995:17192918:d=2021032312:HGT:highest tropospheric freezing level:26 hour fcst: -996:17214270:d=2021032312:RH:highest tropospheric freezing level:26 hour fcst: -997:17233497:d=2021032312:TMP:30-0 mb above ground:26 hour fcst: -998:17254321:d=2021032312:RH:30-0 mb above ground:26 hour fcst: -1000:17293609:d=2021032312:SPFH:30-0 mb above ground:26 hour fcst: -1001:17322432:d=2021032312:UGRD:30-0 mb above ground:26 hour fcst: -1002:17343616:d=2021032312:VGRD:30-0 mb above ground:26 hour fcst: -1030:17968767:d=2021032312:4LFTX:surface:26 hour fcst: -1031:17982289:d=2021032312:CAPE:180-0 mb above ground:26 hour fcst: -1032:17996209:d=2021032312:CIN:180-0 mb above ground:26 hour fcst: -1033:18010351:d=2021032312:HPBL:surface:26 hour fcst: -1034:18041426:d=2021032312:RH:0.33-1 sigma layer:26 hour fcst: -1035:18059750:d=2021032312:RH:0.44-1 sigma layer:26 hour fcst: -1036:18077716:d=2021032312:RH:0.72-0.94 sigma layer:26 hour fcst: -1037:18096178:d=2021032312:RH:0.44-0.72 sigma layer:26 hour fcst: -1038:18115282:d=2021032312:TMP:0.995 sigma level:26 hour fcst: -1039:18136160:d=2021032312:POT:0.995 sigma level:26 hour fcst: -1040:18157005:d=2021032312:RH:0.995 sigma level:26 hour fcst: -1041:18174754:d=2021032312:UGRD:0.995 sigma level:26 hour fcst: -1042:18195787:d=2021032312:VGRD:0.995 sigma level:26 hour fcst: -1043:18217332:d=2021032312:VVEL:0.995 sigma level:26 hour fcst: -1044:18238252:d=2021032312:CAPE:90-0 mb above ground:26 hour fcst: -1045:18251041:d=2021032312:CIN:90-0 mb above ground:26 hour fcst: -1046:18265363:d=2021032312:CAPE:255-0 mb above ground:26 hour fcst: -1047:18278695:d=2021032312:CIN:255-0 mb above ground:26 hour fcst: -1048:18290151:d=2021032312:PLPL:255-0 mb above ground:26 hour fcst: -1049:18313683:d=2021032312:LAND:surface:26 hour fcst: -1050:18315624:d=2021032312:ICEC:surface:26 hour fcst: -1051:18318763:d=2021032312:ALBDO:surface:24-26 hour ave fcst: -1052:18334378:d=2021032312:ICETMP:surface:26 hour fcst: -1094:18943105:d=2021032312:UGRD:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -1095:18953761:d=2021032312:VGRD:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -1096:18964453:d=2021032312:TMP:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -1097:18978792:d=2021032312:HGT:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -1098:18996178:d=2021032312:PRES:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -1099:19014458:d=2021032312:VWSH:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -1100:19026353:d=2021032312:UGRD:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -1101:19037195:d=2021032312:VGRD:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -1102:19047965:d=2021032312:TMP:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -1103:19062464:d=2021032312:HGT:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -1104:19079921:d=2021032312:PRES:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -1105:19098226:d=2021032312:VWSH:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -+ exglobal_atmos_products.sh[45]true -+ exglobal_atmos_products.sh[46]export err=0 -+ exglobal_atmos_products.sh[46]err=0 -+ exglobal_atmos_products.sh[47][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[52][[ 2 -eq 2 ]] -+ exglobal_atmos_products.sh[54]wgrib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.master.grb2f026 -+ exglobal_atmos_products.sh[54]wgrib2 -i -grib tmpfileb_f026 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.master.grb2f026 -+ exglobal_atmos_products.sh[54]grep -F -f /work2/noaa/global/mterry/global-workflow_forked/parm/product/gfs.fFFF.paramlist.b.txt -95:1792399:d=2021032312:HGT:1 mb:26 hour fcst: -96:1810348:d=2021032312:TMP:1 mb:26 hour fcst: -97:1826153:d=2021032312:RH:1 mb:26 hour fcst: -101:1907188:d=2021032312:UGRD:1 mb:26 hour fcst: -102:1925643:d=2021032312:VGRD:1 mb:26 hour fcst: -103:1943275:d=2021032312:ABSV:1 mb:26 hour fcst: -104:1959564:d=2021032312:O3MR:1 mb:26 hour fcst: -105:1980829:d=2021032312:HGT:2 mb:26 hour fcst: -106:2001034:d=2021032312:TMP:2 mb:26 hour fcst: -107:2017420:d=2021032312:RH:2 mb:26 hour fcst: -111:2100103:d=2021032312:UGRD:2 mb:26 hour fcst: -112:2111591:d=2021032312:VGRD:2 mb:26 hour fcst: -113:2122295:d=2021032312:ABSV:2 mb:26 hour fcst: -114:2139303:d=2021032312:O3MR:2 mb:26 hour fcst: -115:2161814:d=2021032312:HGT:3 mb:26 hour fcst: -116:2181755:d=2021032312:TMP:3 mb:26 hour fcst: -117:2198165:d=2021032312:RH:3 mb:26 hour fcst: -121:2281634:d=2021032312:UGRD:3 mb:26 hour fcst: -122:2293165:d=2021032312:VGRD:3 mb:26 hour fcst: -123:2311127:d=2021032312:ABSV:3 mb:26 hour fcst: -124:2328020:d=2021032312:O3MR:3 mb:26 hour fcst: -125:2351352:d=2021032312:HGT:5 mb:26 hour fcst: -126:2369091:d=2021032312:TMP:5 mb:26 hour fcst: -127:2385725:d=2021032312:RH:5 mb:26 hour fcst: -131:2470824:d=2021032312:UGRD:5 mb:26 hour fcst: -132:2482323:d=2021032312:VGRD:5 mb:26 hour fcst: -133:2501178:d=2021032312:ABSV:5 mb:26 hour fcst: -134:2518790:d=2021032312:O3MR:5 mb:26 hour fcst: -135:2541415:d=2021032312:HGT:7 mb:26 hour fcst: -136:2559073:d=2021032312:TMP:7 mb:26 hour fcst: -137:2575215:d=2021032312:RH:7 mb:26 hour fcst: -141:2650786:d=2021032312:UGRD:7 mb:26 hour fcst: -142:2662180:d=2021032312:VGRD:7 mb:26 hour fcst: -143:2680996:d=2021032312:ABSV:7 mb:26 hour fcst: -144:2698727:d=2021032312:O3MR:7 mb:26 hour fcst: -214:3928148:d=2021032312:TCDC:70 mb:26 hour fcst: -221:4067521:d=2021032312:CLMR:70 mb:26 hour fcst: -222:4067700:d=2021032312:ICMR:70 mb:26 hour fcst: -223:4067879:d=2021032312:RWMR:70 mb:26 hour fcst: -224:4068058:d=2021032312:SNMR:70 mb:26 hour fcst: -225:4068237:d=2021032312:GRLE:70 mb:26 hour fcst: -243:4306419:d=2021032312:HGT:125 mb:26 hour fcst: -244:4326456:d=2021032312:TMP:125 mb:26 hour fcst: -245:4343393:d=2021032312:RH:125 mb:26 hour fcst: -246:4353873:d=2021032312:TCDC:125 mb:26 hour fcst: -248:4382586:d=2021032312:VVEL:125 mb:26 hour fcst: -249:4408676:d=2021032312:DZDT:125 mb:26 hour fcst: -250:4433753:d=2021032312:UGRD:125 mb:26 hour fcst: -251:4453409:d=2021032312:VGRD:125 mb:26 hour fcst: -252:4473052:d=2021032312:ABSV:125 mb:26 hour fcst: -253:4491489:d=2021032312:CLMR:125 mb:26 hour fcst: -254:4491668:d=2021032312:ICMR:125 mb:26 hour fcst: -255:4496396:d=2021032312:RWMR:125 mb:26 hour fcst: -256:4496575:d=2021032312:SNMR:125 mb:26 hour fcst: -257:4500750:d=2021032312:GRLE:125 mb:26 hour fcst: -275:4744673:d=2021032312:HGT:175 mb:26 hour fcst: -276:4765338:d=2021032312:TMP:175 mb:26 hour fcst: -277:4782884:d=2021032312:RH:175 mb:26 hour fcst: -278:4796057:d=2021032312:TCDC:175 mb:26 hour fcst: -280:4827563:d=2021032312:VVEL:175 mb:26 hour fcst: -281:4855937:d=2021032312:DZDT:175 mb:26 hour fcst: -282:4881993:d=2021032312:UGRD:175 mb:26 hour fcst: -283:4894992:d=2021032312:VGRD:175 mb:26 hour fcst: -284:4915700:d=2021032312:ABSV:175 mb:26 hour fcst: -285:4935107:d=2021032312:CLMR:175 mb:26 hour fcst: -286:4935286:d=2021032312:ICMR:175 mb:26 hour fcst: -287:4941971:d=2021032312:RWMR:175 mb:26 hour fcst: -288:4942150:d=2021032312:SNMR:175 mb:26 hour fcst: -289:4946524:d=2021032312:GRLE:175 mb:26 hour fcst: -307:5190719:d=2021032312:HGT:225 mb:26 hour fcst: -308:5212191:d=2021032312:TMP:225 mb:26 hour fcst: -309:5230804:d=2021032312:RH:225 mb:26 hour fcst: -310:5247480:d=2021032312:TCDC:225 mb:26 hour fcst: -312:5284337:d=2021032312:VVEL:225 mb:26 hour fcst: -313:5311874:d=2021032312:DZDT:225 mb:26 hour fcst: -314:5338620:d=2021032312:UGRD:225 mb:26 hour fcst: -315:5352419:d=2021032312:VGRD:225 mb:26 hour fcst: -316:5366768:d=2021032312:ABSV:225 mb:26 hour fcst: -317:5387588:d=2021032312:CLMR:225 mb:26 hour fcst: -318:5387767:d=2021032312:ICMR:225 mb:26 hour fcst: -319:5399267:d=2021032312:RWMR:225 mb:26 hour fcst: -320:5399446:d=2021032312:SNMR:225 mb:26 hour fcst: -321:5406299:d=2021032312:GRLE:225 mb:26 hour fcst: -339:5673943:d=2021032312:HGT:275 mb:26 hour fcst: -340:5696120:d=2021032312:TMP:275 mb:26 hour fcst: -341:5714930:d=2021032312:RH:275 mb:26 hour fcst: -342:5733370:d=2021032312:TCDC:275 mb:26 hour fcst: -344:5772869:d=2021032312:VVEL:275 mb:26 hour fcst: -345:5796180:d=2021032312:DZDT:275 mb:26 hour fcst: -346:5824034:d=2021032312:UGRD:275 mb:26 hour fcst: -347:5838381:d=2021032312:VGRD:275 mb:26 hour fcst: -348:5853770:d=2021032312:ABSV:275 mb:26 hour fcst: -349:5875850:d=2021032312:CLMR:275 mb:26 hour fcst: -350:5876029:d=2021032312:ICMR:275 mb:26 hour fcst: -351:5890631:d=2021032312:RWMR:275 mb:26 hour fcst: -352:5890810:d=2021032312:SNMR:275 mb:26 hour fcst: -353:5900973:d=2021032312:GRLE:275 mb:26 hour fcst: -371:6183819:d=2021032312:HGT:325 mb:26 hour fcst: -372:6206060:d=2021032312:TMP:325 mb:26 hour fcst: -373:6224402:d=2021032312:RH:325 mb:26 hour fcst: -374:6243339:d=2021032312:TCDC:325 mb:26 hour fcst: -376:6283683:d=2021032312:VVEL:325 mb:26 hour fcst: -377:6308126:d=2021032312:DZDT:325 mb:26 hour fcst: -378:6336337:d=2021032312:UGRD:325 mb:26 hour fcst: -379:6350988:d=2021032312:VGRD:325 mb:26 hour fcst: -380:6366252:d=2021032312:ABSV:325 mb:26 hour fcst: -381:6388517:d=2021032312:CLMR:325 mb:26 hour fcst: -382:6388696:d=2021032312:ICMR:325 mb:26 hour fcst: -383:6405334:d=2021032312:RWMR:325 mb:26 hour fcst: -384:6405513:d=2021032312:SNMR:325 mb:26 hour fcst: -385:6417474:d=2021032312:GRLE:325 mb:26 hour fcst: -403:6705039:d=2021032312:HGT:375 mb:26 hour fcst: -404:6727096:d=2021032312:TMP:375 mb:26 hour fcst: -405:6745421:d=2021032312:RH:375 mb:26 hour fcst: -406:6764585:d=2021032312:TCDC:375 mb:26 hour fcst: -408:6804735:d=2021032312:VVEL:375 mb:26 hour fcst: -409:6829567:d=2021032312:DZDT:375 mb:26 hour fcst: -410:6857667:d=2021032312:UGRD:375 mb:26 hour fcst: -411:6872149:d=2021032312:VGRD:375 mb:26 hour fcst: -412:6887247:d=2021032312:ABSV:375 mb:26 hour fcst: -413:6909384:d=2021032312:CLMR:375 mb:26 hour fcst: -414:6909640:d=2021032312:ICMR:375 mb:26 hour fcst: -415:6926729:d=2021032312:RWMR:375 mb:26 hour fcst: -416:6926986:d=2021032312:SNMR:375 mb:26 hour fcst: -417:6940113:d=2021032312:GRLE:375 mb:26 hour fcst: -435:7243170:d=2021032312:HGT:425 mb:26 hour fcst: -436:7267214:d=2021032312:TMP:425 mb:26 hour fcst: -437:7285658:d=2021032312:RH:425 mb:26 hour fcst: -438:7304930:d=2021032312:TCDC:425 mb:26 hour fcst: -440:7346414:d=2021032312:VVEL:425 mb:26 hour fcst: -441:7371489:d=2021032312:DZDT:425 mb:26 hour fcst: -442:7399674:d=2021032312:UGRD:425 mb:26 hour fcst: -443:7421528:d=2021032312:VGRD:425 mb:26 hour fcst: -444:7444045:d=2021032312:ABSV:425 mb:26 hour fcst: -445:7465654:d=2021032312:CLMR:425 mb:26 hour fcst: -446:7466308:d=2021032312:ICMR:425 mb:26 hour fcst: -447:7483083:d=2021032312:RWMR:425 mb:26 hour fcst: -448:7483560:d=2021032312:SNMR:425 mb:26 hour fcst: -449:7498103:d=2021032312:GRLE:425 mb:26 hour fcst: -467:7803055:d=2021032312:HGT:475 mb:26 hour fcst: -468:7826901:d=2021032312:TMP:475 mb:26 hour fcst: -469:7845377:d=2021032312:RH:475 mb:26 hour fcst: -470:7865035:d=2021032312:TCDC:475 mb:26 hour fcst: -472:7905273:d=2021032312:VVEL:475 mb:26 hour fcst: -473:7930280:d=2021032312:DZDT:475 mb:26 hour fcst: -474:7958235:d=2021032312:UGRD:475 mb:26 hour fcst: -475:7979813:d=2021032312:VGRD:475 mb:26 hour fcst: -476:8001970:d=2021032312:ABSV:475 mb:26 hour fcst: -477:8023251:d=2021032312:CLMR:475 mb:26 hour fcst: -478:8024433:d=2021032312:ICMR:475 mb:26 hour fcst: -479:8040788:d=2021032312:RWMR:475 mb:26 hour fcst: -480:8041471:d=2021032312:SNMR:475 mb:26 hour fcst: -481:8055433:d=2021032312:GRLE:475 mb:26 hour fcst: -499:8358931:d=2021032312:HGT:525 mb:26 hour fcst: -500:8382568:d=2021032312:TMP:525 mb:26 hour fcst: -501:8401361:d=2021032312:RH:525 mb:26 hour fcst: -502:8421037:d=2021032312:TCDC:525 mb:26 hour fcst: -504:8462106:d=2021032312:VVEL:525 mb:26 hour fcst: -505:8487209:d=2021032312:DZDT:525 mb:26 hour fcst: -506:8514805:d=2021032312:UGRD:525 mb:26 hour fcst: -507:8535985:d=2021032312:VGRD:525 mb:26 hour fcst: -508:8557928:d=2021032312:ABSV:525 mb:26 hour fcst: -509:8578918:d=2021032312:CLMR:525 mb:26 hour fcst: -510:8581234:d=2021032312:ICMR:525 mb:26 hour fcst: -511:8598465:d=2021032312:RWMR:525 mb:26 hour fcst: -512:8599594:d=2021032312:SNMR:525 mb:26 hour fcst: -513:8614009:d=2021032312:GRLE:525 mb:26 hour fcst: -531:8919951:d=2021032312:HGT:575 mb:26 hour fcst: -532:8943448:d=2021032312:TMP:575 mb:26 hour fcst: -533:8962393:d=2021032312:RH:575 mb:26 hour fcst: -534:8982264:d=2021032312:TCDC:575 mb:26 hour fcst: -536:9024226:d=2021032312:VVEL:575 mb:26 hour fcst: -537:9049442:d=2021032312:DZDT:575 mb:26 hour fcst: -538:9076890:d=2021032312:UGRD:575 mb:26 hour fcst: -539:9097836:d=2021032312:VGRD:575 mb:26 hour fcst: -540:9119309:d=2021032312:ABSV:575 mb:26 hour fcst: -541:9140152:d=2021032312:CLMR:575 mb:26 hour fcst: -542:9144411:d=2021032312:ICMR:575 mb:26 hour fcst: -543:9160404:d=2021032312:RWMR:575 mb:26 hour fcst: -544:9162759:d=2021032312:SNMR:575 mb:26 hour fcst: -545:9177317:d=2021032312:GRLE:575 mb:26 hour fcst: -563:9497729:d=2021032312:HGT:625 mb:26 hour fcst: -564:9522025:d=2021032312:TMP:625 mb:26 hour fcst: -565:9541170:d=2021032312:RH:625 mb:26 hour fcst: -566:9561257:d=2021032312:TCDC:625 mb:26 hour fcst: -568:9601375:d=2021032312:VVEL:625 mb:26 hour fcst: -569:9626856:d=2021032312:DZDT:625 mb:26 hour fcst: -570:9654266:d=2021032312:UGRD:625 mb:26 hour fcst: -571:9675267:d=2021032312:VGRD:625 mb:26 hour fcst: -572:9696919:d=2021032312:ABSV:625 mb:26 hour fcst: -573:9717701:d=2021032312:CLMR:625 mb:26 hour fcst: -574:9722759:d=2021032312:ICMR:625 mb:26 hour fcst: -575:9735794:d=2021032312:RWMR:625 mb:26 hour fcst: -576:9740033:d=2021032312:SNMR:625 mb:26 hour fcst: -577:9755301:d=2021032312:GRLE:625 mb:26 hour fcst: -595:10073043:d=2021032312:HGT:675 mb:26 hour fcst: -596:10097397:d=2021032312:TMP:675 mb:26 hour fcst: -597:10116897:d=2021032312:RH:675 mb:26 hour fcst: -598:10137168:d=2021032312:TCDC:675 mb:26 hour fcst: -600:10179430:d=2021032312:VVEL:675 mb:26 hour fcst: -601:10205088:d=2021032312:DZDT:675 mb:26 hour fcst: -602:10232489:d=2021032312:UGRD:675 mb:26 hour fcst: -603:10253543:d=2021032312:VGRD:675 mb:26 hour fcst: -604:10275203:d=2021032312:ABSV:675 mb:26 hour fcst: -605:10296295:d=2021032312:CLMR:675 mb:26 hour fcst: -606:10301884:d=2021032312:ICMR:675 mb:26 hour fcst: -607:10313634:d=2021032312:RWMR:675 mb:26 hour fcst: -608:10318710:d=2021032312:SNMR:675 mb:26 hour fcst: -609:10333211:d=2021032312:GRLE:675 mb:26 hour fcst: -627:10650737:d=2021032312:HGT:725 mb:26 hour fcst: -628:10674886:d=2021032312:TMP:725 mb:26 hour fcst: -629:10694635:d=2021032312:RH:725 mb:26 hour fcst: -630:10714949:d=2021032312:TCDC:725 mb:26 hour fcst: -632:10755628:d=2021032312:VVEL:725 mb:26 hour fcst: -633:10781404:d=2021032312:DZDT:725 mb:26 hour fcst: -634:10808716:d=2021032312:UGRD:725 mb:26 hour fcst: -635:10829530:d=2021032312:VGRD:725 mb:26 hour fcst: -636:10851148:d=2021032312:ABSV:725 mb:26 hour fcst: -637:10872225:d=2021032312:CLMR:725 mb:26 hour fcst: -638:10880223:d=2021032312:ICMR:725 mb:26 hour fcst: -639:10891925:d=2021032312:RWMR:725 mb:26 hour fcst: -640:10897659:d=2021032312:SNMR:725 mb:26 hour fcst: -641:10911502:d=2021032312:GRLE:725 mb:26 hour fcst: -659:11233105:d=2021032312:HGT:775 mb:26 hour fcst: -660:11257564:d=2021032312:TMP:775 mb:26 hour fcst: -661:11277660:d=2021032312:RH:775 mb:26 hour fcst: -662:11298093:d=2021032312:TCDC:775 mb:26 hour fcst: -664:11339994:d=2021032312:VVEL:775 mb:26 hour fcst: -665:11365750:d=2021032312:DZDT:775 mb:26 hour fcst: -666:11392865:d=2021032312:UGRD:775 mb:26 hour fcst: -667:11413729:d=2021032312:VGRD:775 mb:26 hour fcst: -668:11435456:d=2021032312:ABSV:775 mb:26 hour fcst: -669:11456629:d=2021032312:CLMR:775 mb:26 hour fcst: -670:11466899:d=2021032312:ICMR:775 mb:26 hour fcst: -671:11478165:d=2021032312:RWMR:775 mb:26 hour fcst: -672:11484933:d=2021032312:SNMR:775 mb:26 hour fcst: -673:11498169:d=2021032312:GRLE:775 mb:26 hour fcst: -691:11825779:d=2021032312:HGT:825 mb:26 hour fcst: -692:11850374:d=2021032312:TMP:825 mb:26 hour fcst: -693:11871017:d=2021032312:RH:825 mb:26 hour fcst: -694:11891403:d=2021032312:TCDC:825 mb:26 hour fcst: -696:11935030:d=2021032312:VVEL:825 mb:26 hour fcst: -697:11960677:d=2021032312:DZDT:825 mb:26 hour fcst: -698:11987588:d=2021032312:UGRD:825 mb:26 hour fcst: -699:12008817:d=2021032312:VGRD:825 mb:26 hour fcst: -700:12030672:d=2021032312:ABSV:825 mb:26 hour fcst: -701:12051927:d=2021032312:CLMR:825 mb:26 hour fcst: -702:12064593:d=2021032312:ICMR:825 mb:26 hour fcst: -703:12074925:d=2021032312:RWMR:825 mb:26 hour fcst: -704:12084406:d=2021032312:SNMR:825 mb:26 hour fcst: -705:12096672:d=2021032312:GRLE:825 mb:26 hour fcst: -723:12426367:d=2021032312:HGT:875 mb:26 hour fcst: -724:12451096:d=2021032312:TMP:875 mb:26 hour fcst: -725:12472026:d=2021032312:RH:875 mb:26 hour fcst: -726:12492019:d=2021032312:TCDC:875 mb:26 hour fcst: -728:12535757:d=2021032312:VVEL:875 mb:26 hour fcst: -729:12561101:d=2021032312:DZDT:875 mb:26 hour fcst: -730:12587484:d=2021032312:UGRD:875 mb:26 hour fcst: -731:12608787:d=2021032312:VGRD:875 mb:26 hour fcst: -732:12630638:d=2021032312:ABSV:875 mb:26 hour fcst: -733:12651886:d=2021032312:CLMR:875 mb:26 hour fcst: -734:12664113:d=2021032312:ICMR:875 mb:26 hour fcst: -735:12673274:d=2021032312:RWMR:875 mb:26 hour fcst: -736:12684598:d=2021032312:SNMR:875 mb:26 hour fcst: -737:12695597:d=2021032312:GRLE:875 mb:26 hour fcst: -829:14258959:d=2021032312:SOILL:0-0.1 m below ground:26 hour fcst: -832:14292266:d=2021032312:SOILL:0.1-0.4 m below ground:26 hour fcst: -835:14324740:d=2021032312:SOILL:0.4-1 m below ground:26 hour fcst: -838:14357010:d=2021032312:SOILL:1-2 m below ground:26 hour fcst: -839:14366087:d=2021032312:CNWAT:surface:26 hour fcst: -843:14414605:d=2021032312:ICETK:surface:26 hour fcst: -923:15599601:d=2021032312:DUVB:surface:24-26 hour ave fcst: -924:15617031:d=2021032312:CDUVB:surface:24-26 hour ave fcst: -969:16637430:d=2021032312:TMP:305 m above mean sea level:26 hour fcst: -970:16655635:d=2021032312:UGRD:305 m above mean sea level:26 hour fcst: -971:16674798:d=2021032312:VGRD:305 m above mean sea level:26 hour fcst: -972:16694441:d=2021032312:TMP:457 m above mean sea level:26 hour fcst: -973:16713849:d=2021032312:UGRD:457 m above mean sea level:26 hour fcst: -974:16734091:d=2021032312:VGRD:457 m above mean sea level:26 hour fcst: -975:16754654:d=2021032312:TMP:610 m above mean sea level:26 hour fcst: -976:16774993:d=2021032312:UGRD:610 m above mean sea level:26 hour fcst: -977:16795814:d=2021032312:VGRD:610 m above mean sea level:26 hour fcst: -978:16817010:d=2021032312:TMP:914 m above mean sea level:26 hour fcst: -979:16837910:d=2021032312:UGRD:914 m above mean sea level:26 hour fcst: -980:16859267:d=2021032312:VGRD:914 m above mean sea level:26 hour fcst: -990:17083623:d=2021032312:TMP:4572 m above mean sea level:26 hour fcst: -991:17104820:d=2021032312:UGRD:4572 m above mean sea level:26 hour fcst: -992:17128023:d=2021032312:VGRD:4572 m above mean sea level:26 hour fcst: -1005:17406650:d=2021032312:TMP:60-30 mb above ground:26 hour fcst: -1006:17427521:d=2021032312:RH:60-30 mb above ground:26 hour fcst: -1007:17446158:d=2021032312:SPFH:60-30 mb above ground:26 hour fcst: -1008:17474936:d=2021032312:UGRD:60-30 mb above ground:26 hour fcst: -1009:17496545:d=2021032312:VGRD:60-30 mb above ground:26 hour fcst: -1010:17518700:d=2021032312:TMP:90-60 mb above ground:26 hour fcst: -1011:17539569:d=2021032312:RH:90-60 mb above ground:26 hour fcst: -1012:17558634:d=2021032312:SPFH:90-60 mb above ground:26 hour fcst: -1013:17587539:d=2021032312:UGRD:90-60 mb above ground:26 hour fcst: -1014:17608938:d=2021032312:VGRD:90-60 mb above ground:26 hour fcst: -1015:17630946:d=2021032312:TMP:120-90 mb above ground:26 hour fcst: -1016:17651872:d=2021032312:RH:120-90 mb above ground:26 hour fcst: -1017:17671210:d=2021032312:SPFH:120-90 mb above ground:26 hour fcst: -1018:17700168:d=2021032312:UGRD:120-90 mb above ground:26 hour fcst: -1019:17721403:d=2021032312:VGRD:120-90 mb above ground:26 hour fcst: -1020:17743170:d=2021032312:TMP:150-120 mb above ground:26 hour fcst: -1021:17764077:d=2021032312:RH:150-120 mb above ground:26 hour fcst: -1022:17783708:d=2021032312:SPFH:150-120 mb above ground:26 hour fcst: -1023:17812754:d=2021032312:UGRD:150-120 mb above ground:26 hour fcst: -1024:17833793:d=2021032312:VGRD:150-120 mb above ground:26 hour fcst: -1025:17855477:d=2021032312:TMP:180-150 mb above ground:26 hour fcst: -1026:17876622:d=2021032312:RH:180-150 mb above ground:26 hour fcst: -1027:17896628:d=2021032312:SPFH:180-150 mb above ground:26 hour fcst: -1028:17926039:d=2021032312:UGRD:180-150 mb above ground:26 hour fcst: -1029:17947093:d=2021032312:VGRD:180-150 mb above ground:26 hour fcst: -1058:18429656:d=2021032312:UGRD:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -1059:18440311:d=2021032312:VGRD:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -1060:18450855:d=2021032312:TMP:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -1061:18465523:d=2021032312:HGT:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -1062:18482265:d=2021032312:PRES:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -1063:18499151:d=2021032312:VWSH:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -1064:18510148:d=2021032312:UGRD:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -1065:18521211:d=2021032312:VGRD:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -1066:18535662:d=2021032312:TMP:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -1067:18551038:d=2021032312:HGT:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -1068:18568723:d=2021032312:PRES:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -1069:18586446:d=2021032312:VWSH:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -1070:18597794:d=2021032312:UGRD:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -1071:18609080:d=2021032312:VGRD:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -1072:18620362:d=2021032312:TMP:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -1073:18636179:d=2021032312:HGT:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -1074:18654828:d=2021032312:PRES:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -1075:18673434:d=2021032312:VWSH:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -1076:18685697:d=2021032312:UGRD:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -1077:18697009:d=2021032312:VGRD:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -1078:18708331:d=2021032312:TMP:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -1079:18724139:d=2021032312:HGT:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -1080:18742651:d=2021032312:PRES:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -1081:18761155:d=2021032312:VWSH:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -1082:18773506:d=2021032312:UGRD:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -1083:18784485:d=2021032312:VGRD:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -1084:18795478:d=2021032312:TMP:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -1085:18810427:d=2021032312:HGT:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -1086:18828353:d=2021032312:PRES:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -1087:18846052:d=2021032312:VWSH:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -1088:18858158:d=2021032312:UGRD:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -1089:18869215:d=2021032312:VGRD:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -1090:18880248:d=2021032312:TMP:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -1091:18895325:d=2021032312:HGT:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -1092:18913163:d=2021032312:PRES:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -1093:18930805:d=2021032312:VWSH:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -+ exglobal_atmos_products.sh[54]true -+ exglobal_atmos_products.sh[55]export err=0 -+ exglobal_atmos_products.sh[55]err=0 -+ exglobal_atmos_products.sh[56][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[62]grid_string=0p25 -+ exglobal_atmos_products.sh[63][[ '' == \Y\E\S ]] -+ exglobal_atmos_products.sh[66]echo 'Supplemental product generation is disable for fhr = f026' -Supplemental product generation is disable for fhr = f026 -+ exglobal_atmos_products.sh[67]PGBS=NO -+ exglobal_atmos_products.sh[70]IFS=: -+ exglobal_atmos_products.sh[70]read -ra grids -+ exglobal_atmos_products.sh[154](( nset=1 )) -+ exglobal_atmos_products.sh[154](( nset <= downset )) -+ exglobal_atmos_products.sh[74]echo 'Begin processing nset = 1' -Begin processing nset = 1 -+ exglobal_atmos_products.sh[77]nproc=24 -+ exglobal_atmos_products.sh[80][[ 1 == 1 ]] -+ exglobal_atmos_products.sh[81]grp= -+ exglobal_atmos_products.sh[87]tmpfile=tmpfile_f026 -++ exglobal_atmos_products.sh[90]wgrib2 tmpfile_f026 -++ exglobal_atmos_products.sh[90]wc -l -+ exglobal_atmos_products.sh[90]ncount=753 -+ exglobal_atmos_products.sh[91][[ 24 -gt 753 ]] -+ exglobal_atmos_products.sh[95]inv=31 -+ exglobal_atmos_products.sh[96]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601/poescript -+ exglobal_atmos_products.sh[98]last=0 -+ exglobal_atmos_products.sh[134](( iproc = 1 )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=1 -+ exglobal_atmos_products.sh[101]last=31 -+ exglobal_atmos_products.sh[102][[ 31 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 31 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -31:522305:d=2021032312:UGRD:0.02 mb:26 hour fcst: -+ exglobal_atmos_products.sh[112]rc=0 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 0 == 0 ]] -+ exglobal_atmos_products.sh[115]last=32 -+ exglobal_atmos_products.sh[117][[ 1 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 1:32 -grib tmpfile_f026_1 -1:0:d=2021032312:PRMSL:mean sea level:26 hour fcst: -2:25750:d=2021032312:CLMR:1 hybrid level:26 hour fcst: -3:29476:d=2021032312:ICMR:1 hybrid level:26 hour fcst: -4:36524:d=2021032312:RWMR:1 hybrid level:26 hour fcst: -5:48816:d=2021032312:SNMR:1 hybrid level:26 hour fcst: -6:58979:d=2021032312:GRLE:1 hybrid level:26 hour fcst: -7:61830:d=2021032312:REFD:1 hybrid level:26 hour fcst: -8:87479:d=2021032312:REFD:2 hybrid level:26 hour fcst: -9:113130:d=2021032312:REFC:entire atmosphere (considered as a single layer):26 hour fcst: -10:140473:d=2021032312:VIS:surface:26 hour fcst: -11:172814:d=2021032312:UGRD:planetary boundary layer:26 hour fcst: -12:186817:d=2021032312:VGRD:planetary boundary layer:26 hour fcst: -13:201364:d=2021032312:VRATE:planetary boundary layer:26 hour fcst: -14:216434:d=2021032312:GUST:surface:26 hour fcst: -15:231225:d=2021032312:HGT:0.01 mb:26 hour fcst: -16:254282:d=2021032312:TMP:0.01 mb:26 hour fcst: -17:264808:d=2021032312:RH:0.01 mb:26 hour fcst: -18:271141:d=2021032312:SPFH:0.01 mb:26 hour fcst: -19:294155:d=2021032312:VVEL:0.01 mb:26 hour fcst: -20:320852:d=2021032312:DZDT:0.01 mb:26 hour fcst: -21:348169:d=2021032312:UGRD:0.01 mb:26 hour fcst: -22:359247:d=2021032312:VGRD:0.01 mb:26 hour fcst: -23:369910:d=2021032312:ABSV:0.01 mb:26 hour fcst: -24:385354:d=2021032312:O3MR:0.01 mb:26 hour fcst: -25:399357:d=2021032312:HGT:0.02 mb:26 hour fcst: -26:424604:d=2021032312:TMP:0.02 mb:26 hour fcst: -27:439301:d=2021032312:RH:0.02 mb:26 hour fcst: -28:446712:d=2021032312:SPFH:0.02 mb:26 hour fcst: -29:470738:d=2021032312:VVEL:0.02 mb:26 hour fcst: -30:494829:d=2021032312:DZDT:0.02 mb:26 hour fcst: -31:522305:d=2021032312:UGRD:0.02 mb:26 hour fcst: -32:534793:d=2021032312:VGRD:0.02 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_1 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_1 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_1 pgb2file_f026_1 0p25' -+ exglobal_atmos_products.sh[133][[ 32 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=33 -+ exglobal_atmos_products.sh[101]last=63 -+ exglobal_atmos_products.sh[102][[ 63 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 63 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 2 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 33:63 -grib tmpfile_f026_2 -33:546699:d=2021032312:ABSV:0.02 mb:26 hour fcst: -34:563631:d=2021032312:O3MR:0.02 mb:26 hour fcst: -35:579187:d=2021032312:HGT:0.04 mb:26 hour fcst: -36:603337:d=2021032312:TMP:0.04 mb:26 hour fcst: -37:620486:d=2021032312:RH:0.04 mb:26 hour fcst: -38:626524:d=2021032312:SPFH:0.04 mb:26 hour fcst: -39:652055:d=2021032312:VVEL:0.04 mb:26 hour fcst: -40:679552:d=2021032312:DZDT:0.04 mb:26 hour fcst: -41:704107:d=2021032312:UGRD:0.04 mb:26 hour fcst: -42:723980:d=2021032312:VGRD:0.04 mb:26 hour fcst: -43:743716:d=2021032312:ABSV:0.04 mb:26 hour fcst: -44:761961:d=2021032312:O3MR:0.04 mb:26 hour fcst: -45:785458:d=2021032312:HGT:0.07 mb:26 hour fcst: -46:809421:d=2021032312:TMP:0.07 mb:26 hour fcst: -47:826363:d=2021032312:RH:0.07 mb:26 hour fcst: -48:835871:d=2021032312:SPFH:0.07 mb:26 hour fcst: -49:859355:d=2021032312:VVEL:0.07 mb:26 hour fcst: -50:888114:d=2021032312:DZDT:0.07 mb:26 hour fcst: -51:913456:d=2021032312:UGRD:0.07 mb:26 hour fcst: -52:933691:d=2021032312:VGRD:0.07 mb:26 hour fcst: -53:953904:d=2021032312:ABSV:0.07 mb:26 hour fcst: -54:972680:d=2021032312:O3MR:0.07 mb:26 hour fcst: -55:992097:d=2021032312:HGT:0.1 mb:26 hour fcst: -56:1013512:d=2021032312:TMP:0.1 mb:26 hour fcst: -57:1029925:d=2021032312:RH:0.1 mb:26 hour fcst: -58:1038011:d=2021032312:SPFH:0.1 mb:26 hour fcst: -59:1059318:d=2021032312:VVEL:0.1 mb:26 hour fcst: -60:1088866:d=2021032312:DZDT:0.1 mb:26 hour fcst: -61:1114475:d=2021032312:UGRD:0.1 mb:26 hour fcst: -62:1134476:d=2021032312:VGRD:0.1 mb:26 hour fcst: -63:1154499:d=2021032312:ABSV:0.1 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_2 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_2 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_2 pgb2file_f026_2 0p25' -+ exglobal_atmos_products.sh[133][[ 63 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=64 -+ exglobal_atmos_products.sh[101]last=94 -+ exglobal_atmos_products.sh[102][[ 94 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 94 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 3 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 64:94 -grib tmpfile_f026_3 -64:1172993:d=2021032312:O3MR:0.1 mb:26 hour fcst: -65:1196524:d=2021032312:HGT:0.2 mb:26 hour fcst: -66:1217567:d=2021032312:TMP:0.2 mb:26 hour fcst: -67:1233686:d=2021032312:RH:0.2 mb:26 hour fcst: -68:1244656:d=2021032312:SPFH:0.2 mb:26 hour fcst: -69:1264455:d=2021032312:VVEL:0.2 mb:26 hour fcst: -70:1293312:d=2021032312:DZDT:0.2 mb:26 hour fcst: -71:1320768:d=2021032312:UGRD:0.2 mb:26 hour fcst: -72:1340056:d=2021032312:VGRD:0.2 mb:26 hour fcst: -73:1358945:d=2021032312:ABSV:0.2 mb:26 hour fcst: -74:1376593:d=2021032312:O3MR:0.2 mb:26 hour fcst: -75:1402230:d=2021032312:HGT:0.4 mb:26 hour fcst: -76:1423051:d=2021032312:TMP:0.4 mb:26 hour fcst: -77:1438903:d=2021032312:RH:0.4 mb:26 hour fcst: -78:1447906:d=2021032312:SPFH:0.4 mb:26 hour fcst: -79:1467996:d=2021032312:VVEL:0.4 mb:26 hour fcst: -80:1492768:d=2021032312:DZDT:0.4 mb:26 hour fcst: -81:1523367:d=2021032312:UGRD:0.4 mb:26 hour fcst: -82:1542375:d=2021032312:VGRD:0.4 mb:26 hour fcst: -83:1560568:d=2021032312:ABSV:0.4 mb:26 hour fcst: -84:1577318:d=2021032312:O3MR:0.4 mb:26 hour fcst: -85:1602104:d=2021032312:HGT:0.7 mb:26 hour fcst: -86:1620179:d=2021032312:TMP:0.7 mb:26 hour fcst: -87:1635898:d=2021032312:RH:0.7 mb:26 hour fcst: -88:1644805:d=2021032312:SPFH:0.7 mb:26 hour fcst: -89:1664059:d=2021032312:VVEL:0.7 mb:26 hour fcst: -90:1689933:d=2021032312:DZDT:0.7 mb:26 hour fcst: -91:1719194:d=2021032312:UGRD:0.7 mb:26 hour fcst: -92:1737711:d=2021032312:VGRD:0.7 mb:26 hour fcst: -93:1755538:d=2021032312:ABSV:0.7 mb:26 hour fcst: -94:1771947:d=2021032312:O3MR:0.7 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_3 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_3 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_3 pgb2file_f026_3 0p25' -+ exglobal_atmos_products.sh[133][[ 94 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=95 -+ exglobal_atmos_products.sh[101]last=125 -+ exglobal_atmos_products.sh[102][[ 125 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 125 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 4 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 95:125 -grib tmpfile_f026_4 -95:1792399:d=2021032312:HGT:1 mb:26 hour fcst: -96:1810348:d=2021032312:TMP:1 mb:26 hour fcst: -97:1826153:d=2021032312:RH:1 mb:26 hour fcst: -98:1835931:d=2021032312:SPFH:1 mb:26 hour fcst: -99:1854494:d=2021032312:VVEL:1 mb:26 hour fcst: -100:1881231:d=2021032312:DZDT:1 mb:26 hour fcst: -101:1907188:d=2021032312:UGRD:1 mb:26 hour fcst: -102:1925643:d=2021032312:VGRD:1 mb:26 hour fcst: -103:1943275:d=2021032312:ABSV:1 mb:26 hour fcst: -104:1959564:d=2021032312:O3MR:1 mb:26 hour fcst: -105:1980829:d=2021032312:HGT:2 mb:26 hour fcst: -106:2001034:d=2021032312:TMP:2 mb:26 hour fcst: -107:2017420:d=2021032312:RH:2 mb:26 hour fcst: -108:2025782:d=2021032312:SPFH:2 mb:26 hour fcst: -109:2044810:d=2021032312:VVEL:2 mb:26 hour fcst: -110:2073399:d=2021032312:DZDT:2 mb:26 hour fcst: -111:2100103:d=2021032312:UGRD:2 mb:26 hour fcst: -112:2111591:d=2021032312:VGRD:2 mb:26 hour fcst: -113:2122295:d=2021032312:ABSV:2 mb:26 hour fcst: -114:2139303:d=2021032312:O3MR:2 mb:26 hour fcst: -115:2161814:d=2021032312:HGT:3 mb:26 hour fcst: -116:2181755:d=2021032312:TMP:3 mb:26 hour fcst: -117:2198165:d=2021032312:RH:3 mb:26 hour fcst: -118:2204227:d=2021032312:SPFH:3 mb:26 hour fcst: -119:2225416:d=2021032312:VVEL:3 mb:26 hour fcst: -120:2255450:d=2021032312:DZDT:3 mb:26 hour fcst: -121:2281634:d=2021032312:UGRD:3 mb:26 hour fcst: -122:2293165:d=2021032312:VGRD:3 mb:26 hour fcst: -123:2311127:d=2021032312:ABSV:3 mb:26 hour fcst: -124:2328020:d=2021032312:O3MR:3 mb:26 hour fcst: -125:2351352:d=2021032312:HGT:5 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_4 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_4 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_4 pgb2file_f026_4 0p25' -+ exglobal_atmos_products.sh[133][[ 125 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=126 -+ exglobal_atmos_products.sh[101]last=156 -+ exglobal_atmos_products.sh[102][[ 156 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 156 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 5 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 126:156 -grib tmpfile_f026_5 -126:2369091:d=2021032312:TMP:5 mb:26 hour fcst: -127:2385725:d=2021032312:RH:5 mb:26 hour fcst: -128:2394902:d=2021032312:SPFH:5 mb:26 hour fcst: -129:2415697:d=2021032312:VVEL:5 mb:26 hour fcst: -130:2444836:d=2021032312:DZDT:5 mb:26 hour fcst: -131:2470824:d=2021032312:UGRD:5 mb:26 hour fcst: -132:2482323:d=2021032312:VGRD:5 mb:26 hour fcst: -133:2501178:d=2021032312:ABSV:5 mb:26 hour fcst: -134:2518790:d=2021032312:O3MR:5 mb:26 hour fcst: -135:2541415:d=2021032312:HGT:7 mb:26 hour fcst: -136:2559073:d=2021032312:TMP:7 mb:26 hour fcst: -137:2575215:d=2021032312:RH:7 mb:26 hour fcst: -138:2581275:d=2021032312:SPFH:7 mb:26 hour fcst: -139:2600668:d=2021032312:VVEL:7 mb:26 hour fcst: -140:2625275:d=2021032312:DZDT:7 mb:26 hour fcst: -141:2650786:d=2021032312:UGRD:7 mb:26 hour fcst: -142:2662180:d=2021032312:VGRD:7 mb:26 hour fcst: -143:2680996:d=2021032312:ABSV:7 mb:26 hour fcst: -144:2698727:d=2021032312:O3MR:7 mb:26 hour fcst: -145:2722318:d=2021032312:HGT:10 mb:26 hour fcst: -146:2742122:d=2021032312:TMP:10 mb:26 hour fcst: -147:2758273:d=2021032312:RH:10 mb:26 hour fcst: -148:2765773:d=2021032312:SPFH:10 mb:26 hour fcst: -149:2784037:d=2021032312:VVEL:10 mb:26 hour fcst: -150:2809556:d=2021032312:DZDT:10 mb:26 hour fcst: -151:2834831:d=2021032312:UGRD:10 mb:26 hour fcst: -152:2846175:d=2021032312:VGRD:10 mb:26 hour fcst: -153:2864695:d=2021032312:ABSV:10 mb:26 hour fcst: -154:2882348:d=2021032312:O3MR:10 mb:26 hour fcst: -155:2901030:d=2021032312:HGT:15 mb:26 hour fcst: -156:2920588:d=2021032312:TMP:15 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_5 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_5 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_5 pgb2file_f026_5 0p25' -+ exglobal_atmos_products.sh[133][[ 156 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=157 -+ exglobal_atmos_products.sh[101]last=187 -+ exglobal_atmos_products.sh[102][[ 187 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 187 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 6 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 157:187 -grib tmpfile_f026_6 -157:2936689:d=2021032312:RH:15 mb:26 hour fcst: -158:2941958:d=2021032312:SPFH:15 mb:26 hour fcst: -159:2957387:d=2021032312:VVEL:15 mb:26 hour fcst: -160:2983910:d=2021032312:DZDT:15 mb:26 hour fcst: -161:3014231:d=2021032312:UGRD:15 mb:26 hour fcst: -162:3025432:d=2021032312:VGRD:15 mb:26 hour fcst: -163:3043896:d=2021032312:ABSV:15 mb:26 hour fcst: -164:3061531:d=2021032312:O3MR:15 mb:26 hour fcst: -165:3080421:d=2021032312:HGT:20 mb:26 hour fcst: -166:3099777:d=2021032312:TMP:20 mb:26 hour fcst: -167:3115689:d=2021032312:RH:20 mb:26 hour fcst: -168:3121849:d=2021032312:SPFH:20 mb:26 hour fcst: -169:3136082:d=2021032312:VVEL:20 mb:26 hour fcst: -170:3163523:d=2021032312:DZDT:20 mb:26 hour fcst: -171:3193853:d=2021032312:UGRD:20 mb:26 hour fcst: -172:3212511:d=2021032312:VGRD:20 mb:26 hour fcst: -173:3231026:d=2021032312:ABSV:20 mb:26 hour fcst: -174:3248539:d=2021032312:O3MR:20 mb:26 hour fcst: -175:3273277:d=2021032312:HGT:30 mb:26 hour fcst: -176:3292339:d=2021032312:TMP:30 mb:26 hour fcst: -177:3308240:d=2021032312:RH:30 mb:26 hour fcst: -178:3315834:d=2021032312:SPFH:30 mb:26 hour fcst: -179:3331141:d=2021032312:VVEL:30 mb:26 hour fcst: -180:3359465:d=2021032312:DZDT:30 mb:26 hour fcst: -181:3389355:d=2021032312:UGRD:30 mb:26 hour fcst: -182:3407928:d=2021032312:VGRD:30 mb:26 hour fcst: -183:3426434:d=2021032312:ABSV:30 mb:26 hour fcst: -184:3444111:d=2021032312:O3MR:30 mb:26 hour fcst: -185:3470823:d=2021032312:HGT:40 mb:26 hour fcst: -186:3489699:d=2021032312:TMP:40 mb:26 hour fcst: -187:3505992:d=2021032312:RH:40 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_6 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_6 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_6 pgb2file_f026_6 0p25' -+ exglobal_atmos_products.sh[133][[ 187 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=188 -+ exglobal_atmos_products.sh[101]last=218 -+ exglobal_atmos_products.sh[102][[ 218 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 218 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 7 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 188:218 -grib tmpfile_f026_7 -188:3515450:d=2021032312:SPFH:40 mb:26 hour fcst: -189:3532482:d=2021032312:VVEL:40 mb:26 hour fcst: -190:3561813:d=2021032312:DZDT:40 mb:26 hour fcst: -191:3594033:d=2021032312:UGRD:40 mb:26 hour fcst: -192:3612743:d=2021032312:VGRD:40 mb:26 hour fcst: -193:3631505:d=2021032312:ABSV:40 mb:26 hour fcst: -194:3649361:d=2021032312:O3MR:40 mb:26 hour fcst: -195:3676305:d=2021032312:HGT:50 mb:26 hour fcst: -196:3695240:d=2021032312:TMP:50 mb:26 hour fcst: -197:3711290:d=2021032312:RH:50 mb:26 hour fcst: -198:3721921:d=2021032312:TCDC:50 mb:26 hour fcst: -199:3722100:d=2021032312:SPFH:50 mb:26 hour fcst: -200:3740575:d=2021032312:VVEL:50 mb:26 hour fcst: -201:3770548:d=2021032312:DZDT:50 mb:26 hour fcst: -202:3802685:d=2021032312:UGRD:50 mb:26 hour fcst: -203:3821122:d=2021032312:VGRD:50 mb:26 hour fcst: -204:3839597:d=2021032312:ABSV:50 mb:26 hour fcst: -205:3857267:d=2021032312:CLMR:50 mb:26 hour fcst: -206:3857446:d=2021032312:ICMR:50 mb:26 hour fcst: -207:3857625:d=2021032312:RWMR:50 mb:26 hour fcst: -208:3857804:d=2021032312:SNMR:50 mb:26 hour fcst: -209:3857983:d=2021032312:GRLE:50 mb:26 hour fcst: -210:3858162:d=2021032312:O3MR:50 mb:26 hour fcst: -211:3884529:d=2021032312:HGT:70 mb:26 hour fcst: -212:3903567:d=2021032312:TMP:70 mb:26 hour fcst: -213:3920070:d=2021032312:RH:70 mb:26 hour fcst: -214:3928148:d=2021032312:SPFH:70 mb:26 hour fcst: -215:3948993:d=2021032312:VVEL:70 mb:26 hour fcst: -216:3980201:d=2021032312:DZDT:70 mb:26 hour fcst: -217:4012300:d=2021032312:UGRD:70 mb:26 hour fcst: -218:4030880:d=2021032312:VGRD:70 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_7 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_7 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_7 pgb2file_f026_7 0p25' -+ exglobal_atmos_products.sh[133][[ 218 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=219 -+ exglobal_atmos_products.sh[101]last=249 -+ exglobal_atmos_products.sh[102][[ 249 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 249 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 8 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 219:249 -grib tmpfile_f026_8 -219:4049644:d=2021032312:ABSV:70 mb:26 hour fcst: -220:4067342:d=2021032312:O3MR:70 mb:26 hour fcst: -221:4093210:d=2021032312:HGT:100 mb:26 hour fcst: -222:4112816:d=2021032312:TMP:100 mb:26 hour fcst: -223:4129661:d=2021032312:RH:100 mb:26 hour fcst: -224:4139779:d=2021032312:TCDC:100 mb:26 hour fcst: -225:4139958:d=2021032312:SPFH:100 mb:26 hour fcst: -226:4165525:d=2021032312:VVEL:100 mb:26 hour fcst: -227:4190406:d=2021032312:DZDT:100 mb:26 hour fcst: -228:4220661:d=2021032312:UGRD:100 mb:26 hour fcst: -229:4240023:d=2021032312:VGRD:100 mb:26 hour fcst: -230:4259443:d=2021032312:ABSV:100 mb:26 hour fcst: -231:4277621:d=2021032312:CLMR:100 mb:26 hour fcst: -232:4277800:d=2021032312:ICMR:100 mb:26 hour fcst: -233:4278711:d=2021032312:RWMR:100 mb:26 hour fcst: -234:4278890:d=2021032312:SNMR:100 mb:26 hour fcst: -235:4281143:d=2021032312:GRLE:100 mb:26 hour fcst: -236:4281322:d=2021032312:O3MR:100 mb:26 hour fcst: -237:4305345:d=2021032312:HGT:150 mb:26 hour fcst: -238:4325736:d=2021032312:TMP:150 mb:26 hour fcst: -239:4342706:d=2021032312:RH:150 mb:26 hour fcst: -240:4354101:d=2021032312:TCDC:150 mb:26 hour fcst: -241:4358139:d=2021032312:SPFH:150 mb:26 hour fcst: -242:4382392:d=2021032312:VVEL:150 mb:26 hour fcst: -243:4409749:d=2021032312:DZDT:150 mb:26 hour fcst: -244:4435343:d=2021032312:UGRD:150 mb:26 hour fcst: -245:4455345:d=2021032312:VGRD:150 mb:26 hour fcst: -246:4475303:d=2021032312:ABSV:150 mb:26 hour fcst: -247:4494251:d=2021032312:CLMR:150 mb:26 hour fcst: -248:4494430:d=2021032312:ICMR:150 mb:26 hour fcst: -249:4500139:d=2021032312:RWMR:150 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_8 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_8 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_8 pgb2file_f026_8 0p25' -+ exglobal_atmos_products.sh[133][[ 249 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=250 -+ exglobal_atmos_products.sh[101]last=280 -+ exglobal_atmos_products.sh[102][[ 280 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 280 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 9 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 250:280 -grib tmpfile_f026_9 -250:4500318:d=2021032312:SNMR:150 mb:26 hour fcst: -251:4503876:d=2021032312:GRLE:150 mb:26 hour fcst: -252:4504395:d=2021032312:O3MR:150 mb:26 hour fcst: -253:4526269:d=2021032312:HGT:200 mb:26 hour fcst: -254:4547332:d=2021032312:TMP:200 mb:26 hour fcst: -255:4565477:d=2021032312:RH:200 mb:26 hour fcst: -256:4580507:d=2021032312:TCDC:200 mb:26 hour fcst: -257:4587226:d=2021032312:SPFH:200 mb:26 hour fcst: -258:4612981:d=2021032312:VVEL:200 mb:26 hour fcst: -259:4639780:d=2021032312:DZDT:200 mb:26 hour fcst: -260:4666209:d=2021032312:UGRD:200 mb:26 hour fcst: -261:4679597:d=2021032312:VGRD:200 mb:26 hour fcst: -262:4693328:d=2021032312:ABSV:200 mb:26 hour fcst: -263:4713323:d=2021032312:CLMR:200 mb:26 hour fcst: -264:4713502:d=2021032312:ICMR:200 mb:26 hour fcst: -265:4722151:d=2021032312:RWMR:200 mb:26 hour fcst: -266:4722330:d=2021032312:SNMR:200 mb:26 hour fcst: -267:4727782:d=2021032312:GRLE:200 mb:26 hour fcst: -268:4728111:d=2021032312:O3MR:200 mb:26 hour fcst: -269:4748748:d=2021032312:HGT:250 mb:26 hour fcst: -270:4770483:d=2021032312:TMP:250 mb:26 hour fcst: -271:4789317:d=2021032312:RH:250 mb:26 hour fcst: -272:4807139:d=2021032312:TCDC:250 mb:26 hour fcst: -273:4817851:d=2021032312:SPFH:250 mb:26 hour fcst: -274:4845931:d=2021032312:VVEL:250 mb:26 hour fcst: -275:4868665:d=2021032312:DZDT:250 mb:26 hour fcst: -276:4896057:d=2021032312:UGRD:250 mb:26 hour fcst: -277:4910196:d=2021032312:VGRD:250 mb:26 hour fcst: -278:4925052:d=2021032312:ABSV:250 mb:26 hour fcst: -279:4946654:d=2021032312:CLMR:250 mb:26 hour fcst: -280:4946833:d=2021032312:ICMR:250 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_9 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_9 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_9 pgb2file_f026_9 0p25' -+ exglobal_atmos_products.sh[133][[ 280 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=281 -+ exglobal_atmos_products.sh[101]last=311 -+ exglobal_atmos_products.sh[102][[ 311 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 311 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 10 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 281:311 -grib tmpfile_f026_10 -281:4961135:d=2021032312:RWMR:250 mb:26 hour fcst: -282:4961314:d=2021032312:SNMR:250 mb:26 hour fcst: -283:4970201:d=2021032312:GRLE:250 mb:26 hour fcst: -284:4970507:d=2021032312:O3MR:250 mb:26 hour fcst: -285:4995941:d=2021032312:HGT:300 mb:26 hour fcst: -286:5018260:d=2021032312:TMP:300 mb:26 hour fcst: -287:5036675:d=2021032312:RH:300 mb:26 hour fcst: -288:5055653:d=2021032312:TCDC:300 mb:26 hour fcst: -289:5068437:d=2021032312:SPFH:300 mb:26 hour fcst: -290:5095025:d=2021032312:VVEL:300 mb:26 hour fcst: -291:5118876:d=2021032312:DZDT:300 mb:26 hour fcst: -292:5146896:d=2021032312:UGRD:300 mb:26 hour fcst: -293:5161536:d=2021032312:VGRD:300 mb:26 hour fcst: -294:5176922:d=2021032312:ABSV:300 mb:26 hour fcst: -295:5199252:d=2021032312:CLMR:300 mb:26 hour fcst: -296:5199431:d=2021032312:ICMR:300 mb:26 hour fcst: -297:5215286:d=2021032312:RWMR:300 mb:26 hour fcst: -298:5215465:d=2021032312:SNMR:300 mb:26 hour fcst: -299:5225868:d=2021032312:GRLE:300 mb:26 hour fcst: -300:5226249:d=2021032312:O3MR:300 mb:26 hour fcst: -301:5253422:d=2021032312:HGT:350 mb:26 hour fcst: -302:5275586:d=2021032312:TMP:350 mb:26 hour fcst: -303:5293798:d=2021032312:RH:350 mb:26 hour fcst: -304:5312705:d=2021032312:TCDC:350 mb:26 hour fcst: -305:5325967:d=2021032312:SPFH:350 mb:26 hour fcst: -306:5354118:d=2021032312:VVEL:350 mb:26 hour fcst: -307:5378614:d=2021032312:DZDT:350 mb:26 hour fcst: -308:5406883:d=2021032312:UGRD:350 mb:26 hour fcst: -309:5421466:d=2021032312:VGRD:350 mb:26 hour fcst: -310:5436711:d=2021032312:ABSV:350 mb:26 hour fcst: -311:5459004:d=2021032312:CLMR:350 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_10 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_10 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_10 pgb2file_f026_10 0p25' -+ exglobal_atmos_products.sh[133][[ 311 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=312 -+ exglobal_atmos_products.sh[101]last=342 -+ exglobal_atmos_products.sh[102][[ 342 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 342 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 11 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 312:342 -grib tmpfile_f026_11 -312:5459183:d=2021032312:ICMR:350 mb:26 hour fcst: -313:5476182:d=2021032312:RWMR:350 mb:26 hour fcst: -314:5476361:d=2021032312:SNMR:350 mb:26 hour fcst: -315:5488315:d=2021032312:GRLE:350 mb:26 hour fcst: -316:5488806:d=2021032312:O3MR:350 mb:26 hour fcst: -317:5514293:d=2021032312:HGT:400 mb:26 hour fcst: -318:5536227:d=2021032312:TMP:400 mb:26 hour fcst: -319:5554606:d=2021032312:RH:400 mb:26 hour fcst: -320:5573808:d=2021032312:TCDC:400 mb:26 hour fcst: -321:5587103:d=2021032312:SPFH:400 mb:26 hour fcst: -322:5614863:d=2021032312:VVEL:400 mb:26 hour fcst: -323:5639676:d=2021032312:DZDT:400 mb:26 hour fcst: -324:5667776:d=2021032312:UGRD:400 mb:26 hour fcst: -325:5689819:d=2021032312:VGRD:400 mb:26 hour fcst: -326:5712478:d=2021032312:ABSV:400 mb:26 hour fcst: -327:5734450:d=2021032312:CLMR:400 mb:26 hour fcst: -328:5734852:d=2021032312:ICMR:400 mb:26 hour fcst: -329:5751929:d=2021032312:RWMR:400 mb:26 hour fcst: -330:5752307:d=2021032312:SNMR:400 mb:26 hour fcst: -331:5766253:d=2021032312:GRLE:400 mb:26 hour fcst: -332:5767293:d=2021032312:O3MR:400 mb:26 hour fcst: -333:5791838:d=2021032312:HGT:450 mb:26 hour fcst: -334:5815757:d=2021032312:TMP:450 mb:26 hour fcst: -335:5834155:d=2021032312:RH:450 mb:26 hour fcst: -336:5853521:d=2021032312:TCDC:450 mb:26 hour fcst: -337:5866117:d=2021032312:SPFH:450 mb:26 hour fcst: -338:5895401:d=2021032312:VVEL:450 mb:26 hour fcst: -339:5920551:d=2021032312:DZDT:450 mb:26 hour fcst: -340:5948569:d=2021032312:UGRD:450 mb:26 hour fcst: -341:5970199:d=2021032312:VGRD:450 mb:26 hour fcst: -342:5992478:d=2021032312:ABSV:450 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_11 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_11 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_11 pgb2file_f026_11 0p25' -+ exglobal_atmos_products.sh[133][[ 342 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=343 -+ exglobal_atmos_products.sh[101]last=373 -+ exglobal_atmos_products.sh[102][[ 373 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 373 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 12 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 343:373 -grib tmpfile_f026_12 -343:6013906:d=2021032312:CLMR:450 mb:26 hour fcst: -344:6014766:d=2021032312:ICMR:450 mb:26 hour fcst: -345:6031506:d=2021032312:RWMR:450 mb:26 hour fcst: -346:6032026:d=2021032312:SNMR:450 mb:26 hour fcst: -347:6045632:d=2021032312:GRLE:450 mb:26 hour fcst: -348:6047040:d=2021032312:O3MR:450 mb:26 hour fcst: -349:6071145:d=2021032312:HGT:500 mb:26 hour fcst: -350:6094873:d=2021032312:TMP:500 mb:26 hour fcst: -351:6113567:d=2021032312:RH:500 mb:26 hour fcst: -352:6133102:d=2021032312:TCDC:500 mb:26 hour fcst: -353:6145403:d=2021032312:SPFH:500 mb:26 hour fcst: -354:6173658:d=2021032312:VVEL:500 mb:26 hour fcst: -355:6198899:d=2021032312:DZDT:500 mb:26 hour fcst: -356:6226555:d=2021032312:UGRD:500 mb:26 hour fcst: -357:6248027:d=2021032312:VGRD:500 mb:26 hour fcst: -358:6270027:d=2021032312:ABSV:500 mb:26 hour fcst: -359:6291145:d=2021032312:CLMR:500 mb:26 hour fcst: -360:6292773:d=2021032312:ICMR:500 mb:26 hour fcst: -361:6308884:d=2021032312:RWMR:500 mb:26 hour fcst: -362:6309846:d=2021032312:SNMR:500 mb:26 hour fcst: -363:6324127:d=2021032312:GRLE:500 mb:26 hour fcst: -364:6326012:d=2021032312:O3MR:500 mb:26 hour fcst: -365:6349463:d=2021032312:HGT:550 mb:26 hour fcst: -366:6372998:d=2021032312:TMP:550 mb:26 hour fcst: -367:6391845:d=2021032312:RH:550 mb:26 hour fcst: -368:6411606:d=2021032312:TCDC:550 mb:26 hour fcst: -369:6423800:d=2021032312:SPFH:550 mb:26 hour fcst: -370:6453309:d=2021032312:VVEL:550 mb:26 hour fcst: -371:6478640:d=2021032312:DZDT:550 mb:26 hour fcst: -372:6506187:d=2021032312:UGRD:550 mb:26 hour fcst: -373:6527199:d=2021032312:VGRD:550 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_12 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_12 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_12 pgb2file_f026_12 0p25' -+ exglobal_atmos_products.sh[133][[ 373 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=374 -+ exglobal_atmos_products.sh[101]last=404 -+ exglobal_atmos_products.sh[102][[ 404 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 404 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -404:7078031:d=2021032312:UGRD:650 mb:26 hour fcst: -+ exglobal_atmos_products.sh[112]rc=0 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 0 == 0 ]] -+ exglobal_atmos_products.sh[115]last=405 -+ exglobal_atmos_products.sh[117][[ 13 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 374:405 -grib tmpfile_f026_13 -374:6549030:d=2021032312:ABSV:550 mb:26 hour fcst: -375:6569980:d=2021032312:CLMR:550 mb:26 hour fcst: -376:6573263:d=2021032312:ICMR:550 mb:26 hour fcst: -377:6589958:d=2021032312:RWMR:550 mb:26 hour fcst: -378:6591337:d=2021032312:SNMR:550 mb:26 hour fcst: -379:6605845:d=2021032312:GRLE:550 mb:26 hour fcst: -380:6607768:d=2021032312:O3MR:550 mb:26 hour fcst: -381:6630595:d=2021032312:HGT:600 mb:26 hour fcst: -382:6654747:d=2021032312:TMP:600 mb:26 hour fcst: -383:6673797:d=2021032312:RH:600 mb:26 hour fcst: -384:6693815:d=2021032312:TCDC:600 mb:26 hour fcst: -385:6705893:d=2021032312:SPFH:600 mb:26 hour fcst: -386:6736157:d=2021032312:VVEL:600 mb:26 hour fcst: -387:6761562:d=2021032312:DZDT:600 mb:26 hour fcst: -388:6788977:d=2021032312:UGRD:600 mb:26 hour fcst: -389:6809862:d=2021032312:VGRD:600 mb:26 hour fcst: -390:6831479:d=2021032312:ABSV:600 mb:26 hour fcst: -391:6852285:d=2021032312:CLMR:600 mb:26 hour fcst: -392:6857121:d=2021032312:ICMR:600 mb:26 hour fcst: -393:6871025:d=2021032312:RWMR:600 mb:26 hour fcst: -394:6874746:d=2021032312:SNMR:600 mb:26 hour fcst: -395:6890534:d=2021032312:GRLE:600 mb:26 hour fcst: -396:6892760:d=2021032312:O3MR:600 mb:26 hour fcst: -397:6920902:d=2021032312:HGT:650 mb:26 hour fcst: -398:6945002:d=2021032312:TMP:650 mb:26 hour fcst: -399:6964320:d=2021032312:RH:650 mb:26 hour fcst: -400:6984546:d=2021032312:TCDC:650 mb:26 hour fcst: -401:6996284:d=2021032312:SPFH:650 mb:26 hour fcst: -402:7024992:d=2021032312:VVEL:650 mb:26 hour fcst: -403:7050606:d=2021032312:DZDT:650 mb:26 hour fcst: -404:7078031:d=2021032312:UGRD:650 mb:26 hour fcst: -405:7099128:d=2021032312:VGRD:650 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_13 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_13 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_13 pgb2file_f026_13 0p25' -+ exglobal_atmos_products.sh[133][[ 405 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=406 -+ exglobal_atmos_products.sh[101]last=436 -+ exglobal_atmos_products.sh[102][[ 436 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 436 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -436:7655492:d=2021032312:UGRD:750 mb:26 hour fcst: -+ exglobal_atmos_products.sh[112]rc=0 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 0 == 0 ]] -+ exglobal_atmos_products.sh[115]last=437 -+ exglobal_atmos_products.sh[117][[ 14 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 406:437 -grib tmpfile_f026_14 -406:7120689:d=2021032312:ABSV:650 mb:26 hour fcst: -407:7141568:d=2021032312:CLMR:650 mb:26 hour fcst: -408:7146597:d=2021032312:ICMR:650 mb:26 hour fcst: -409:7159000:d=2021032312:RWMR:650 mb:26 hour fcst: -410:7163718:d=2021032312:SNMR:650 mb:26 hour fcst: -411:7178596:d=2021032312:GRLE:650 mb:26 hour fcst: -412:7180909:d=2021032312:O3MR:650 mb:26 hour fcst: -413:7208432:d=2021032312:HGT:700 mb:26 hour fcst: -414:7232617:d=2021032312:TMP:700 mb:26 hour fcst: -415:7252389:d=2021032312:RH:700 mb:26 hour fcst: -416:7272649:d=2021032312:TCDC:700 mb:26 hour fcst: -417:7284241:d=2021032312:SPFH:700 mb:26 hour fcst: -418:7312810:d=2021032312:VVEL:700 mb:26 hour fcst: -419:7338530:d=2021032312:DZDT:700 mb:26 hour fcst: -420:7365891:d=2021032312:UGRD:700 mb:26 hour fcst: -421:7386776:d=2021032312:VGRD:700 mb:26 hour fcst: -422:7408375:d=2021032312:ABSV:700 mb:26 hour fcst: -423:7429411:d=2021032312:CLMR:700 mb:26 hour fcst: -424:7436277:d=2021032312:ICMR:700 mb:26 hour fcst: -425:7447555:d=2021032312:RWMR:700 mb:26 hour fcst: -426:7453268:d=2021032312:SNMR:700 mb:26 hour fcst: -427:7467308:d=2021032312:GRLE:700 mb:26 hour fcst: -428:7469185:d=2021032312:O3MR:700 mb:26 hour fcst: -429:7496481:d=2021032312:HGT:750 mb:26 hour fcst: -430:7520886:d=2021032312:TMP:750 mb:26 hour fcst: -431:7540824:d=2021032312:RH:750 mb:26 hour fcst: -432:7561269:d=2021032312:TCDC:750 mb:26 hour fcst: -433:7573492:d=2021032312:SPFH:750 mb:26 hour fcst: -434:7602465:d=2021032312:VVEL:750 mb:26 hour fcst: -435:7628249:d=2021032312:DZDT:750 mb:26 hour fcst: -436:7655492:d=2021032312:UGRD:750 mb:26 hour fcst: -437:7676352:d=2021032312:VGRD:750 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_14 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_14 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_14 pgb2file_f026_14 0p25' -+ exglobal_atmos_products.sh[133][[ 437 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=438 -+ exglobal_atmos_products.sh[101]last=468 -+ exglobal_atmos_products.sh[102][[ 468 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 468 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -468:8249079:d=2021032312:UGRD:850 mb:26 hour fcst: -+ exglobal_atmos_products.sh[112]rc=0 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 0 == 0 ]] -+ exglobal_atmos_products.sh[115]last=469 -+ exglobal_atmos_products.sh[117][[ 15 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 438:469 -grib tmpfile_f026_15 -438:7698103:d=2021032312:ABSV:750 mb:26 hour fcst: -439:7719248:d=2021032312:CLMR:750 mb:26 hour fcst: -440:7728581:d=2021032312:ICMR:750 mb:26 hour fcst: -441:7740024:d=2021032312:RWMR:750 mb:26 hour fcst: -442:7745897:d=2021032312:SNMR:750 mb:26 hour fcst: -443:7759504:d=2021032312:GRLE:750 mb:26 hour fcst: -444:7761592:d=2021032312:O3MR:750 mb:26 hour fcst: -445:7788833:d=2021032312:HGT:800 mb:26 hour fcst: -446:7813168:d=2021032312:TMP:800 mb:26 hour fcst: -447:7833581:d=2021032312:RH:800 mb:26 hour fcst: -448:7854019:d=2021032312:TCDC:800 mb:26 hour fcst: -449:7867523:d=2021032312:SPFH:800 mb:26 hour fcst: -450:7897054:d=2021032312:VVEL:800 mb:26 hour fcst: -451:7922764:d=2021032312:DZDT:800 mb:26 hour fcst: -452:7949784:d=2021032312:UGRD:800 mb:26 hour fcst: -453:7970728:d=2021032312:VGRD:800 mb:26 hour fcst: -454:7992480:d=2021032312:ABSV:800 mb:26 hour fcst: -455:8013692:d=2021032312:CLMR:800 mb:26 hour fcst: -456:8025242:d=2021032312:ICMR:800 mb:26 hour fcst: -457:8036273:d=2021032312:RWMR:800 mb:26 hour fcst: -458:8044521:d=2021032312:SNMR:800 mb:26 hour fcst: -459:8057378:d=2021032312:GRLE:800 mb:26 hour fcst: -460:8059744:d=2021032312:O3MR:800 mb:26 hour fcst: -461:8086797:d=2021032312:HGT:850 mb:26 hour fcst: -462:8111577:d=2021032312:TMP:850 mb:26 hour fcst: -463:8132454:d=2021032312:RH:850 mb:26 hour fcst: -464:8152712:d=2021032312:TCDC:850 mb:26 hour fcst: -465:8167436:d=2021032312:SPFH:850 mb:26 hour fcst: -466:8196991:d=2021032312:VVEL:850 mb:26 hour fcst: -467:8222509:d=2021032312:DZDT:850 mb:26 hour fcst: -468:8249079:d=2021032312:UGRD:850 mb:26 hour fcst: -469:8270322:d=2021032312:VGRD:850 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_15 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_15 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_15 pgb2file_f026_15 0p25' -+ exglobal_atmos_products.sh[133][[ 469 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=470 -+ exglobal_atmos_products.sh[101]last=500 -+ exglobal_atmos_products.sh[102][[ 500 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 500 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -500:8838010:d=2021032312:UGRD:925 mb:26 hour fcst: -+ exglobal_atmos_products.sh[112]rc=0 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 0 == 0 ]] -+ exglobal_atmos_products.sh[115]last=501 -+ exglobal_atmos_products.sh[117][[ 16 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 470:501 -grib tmpfile_f026_16 -470:8292147:d=2021032312:ABSV:850 mb:26 hour fcst: -471:8313386:d=2021032312:CLMR:850 mb:26 hour fcst: -472:8326447:d=2021032312:ICMR:850 mb:26 hour fcst: -473:8336277:d=2021032312:RWMR:850 mb:26 hour fcst: -474:8346688:d=2021032312:SNMR:850 mb:26 hour fcst: -475:8358344:d=2021032312:GRLE:850 mb:26 hour fcst: -476:8360679:d=2021032312:O3MR:850 mb:26 hour fcst: -477:8387239:d=2021032312:HGT:900 mb:26 hour fcst: -478:8412297:d=2021032312:TMP:900 mb:26 hour fcst: -479:8433270:d=2021032312:RH:900 mb:26 hour fcst: -480:8453123:d=2021032312:TCDC:900 mb:26 hour fcst: -481:8465990:d=2021032312:SPFH:900 mb:26 hour fcst: -482:8495477:d=2021032312:VVEL:900 mb:26 hour fcst: -483:8520370:d=2021032312:DZDT:900 mb:26 hour fcst: -484:8546450:d=2021032312:UGRD:900 mb:26 hour fcst: -485:8567800:d=2021032312:VGRD:900 mb:26 hour fcst: -486:8589721:d=2021032312:ABSV:900 mb:26 hour fcst: -487:8610945:d=2021032312:CLMR:900 mb:26 hour fcst: -488:8620940:d=2021032312:ICMR:900 mb:26 hour fcst: -489:8629375:d=2021032312:RWMR:900 mb:26 hour fcst: -490:8641378:d=2021032312:SNMR:900 mb:26 hour fcst: -491:8651619:d=2021032312:GRLE:900 mb:26 hour fcst: -492:8654358:d=2021032312:O3MR:900 mb:26 hour fcst: -493:8680402:d=2021032312:HGT:925 mb:26 hour fcst: -494:8705619:d=2021032312:TMP:925 mb:26 hour fcst: -495:8726776:d=2021032312:RH:925 mb:26 hour fcst: -496:8746583:d=2021032312:TCDC:925 mb:26 hour fcst: -497:8758386:d=2021032312:SPFH:925 mb:26 hour fcst: -498:8787929:d=2021032312:VVEL:925 mb:26 hour fcst: -499:8812401:d=2021032312:DZDT:925 mb:26 hour fcst: -500:8838010:d=2021032312:UGRD:925 mb:26 hour fcst: -501:8859509:d=2021032312:VGRD:925 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_16 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_16 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_16 pgb2file_f026_16 0p25' -+ exglobal_atmos_products.sh[133][[ 501 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=502 -+ exglobal_atmos_products.sh[101]last=532 -+ exglobal_atmos_products.sh[102][[ 532 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 532 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 17 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 502:532 -grib tmpfile_f026_17 -502:8881429:d=2021032312:ABSV:925 mb:26 hour fcst: -503:8902697:d=2021032312:CLMR:925 mb:26 hour fcst: -504:8911211:d=2021032312:ICMR:925 mb:26 hour fcst: -505:8918880:d=2021032312:RWMR:925 mb:26 hour fcst: -506:8931126:d=2021032312:SNMR:925 mb:26 hour fcst: -507:8940721:d=2021032312:GRLE:925 mb:26 hour fcst: -508:8943497:d=2021032312:O3MR:925 mb:26 hour fcst: -509:8969449:d=2021032312:HGT:950 mb:26 hour fcst: -510:8994801:d=2021032312:TMP:950 mb:26 hour fcst: -511:9015845:d=2021032312:RH:950 mb:26 hour fcst: -512:9035366:d=2021032312:TCDC:950 mb:26 hour fcst: -513:9045444:d=2021032312:SPFH:950 mb:26 hour fcst: -514:9072698:d=2021032312:VVEL:950 mb:26 hour fcst: -515:9096541:d=2021032312:DZDT:950 mb:26 hour fcst: -516:9121429:d=2021032312:UGRD:950 mb:26 hour fcst: -517:9142937:d=2021032312:VGRD:950 mb:26 hour fcst: -518:9164877:d=2021032312:ABSV:950 mb:26 hour fcst: -519:9185968:d=2021032312:CLMR:950 mb:26 hour fcst: -520:9192437:d=2021032312:ICMR:950 mb:26 hour fcst: -521:9199559:d=2021032312:RWMR:950 mb:26 hour fcst: -522:9211852:d=2021032312:SNMR:950 mb:26 hour fcst: -523:9220662:d=2021032312:GRLE:950 mb:26 hour fcst: -524:9223210:d=2021032312:O3MR:950 mb:26 hour fcst: -525:9249145:d=2021032312:HINDEX:surface:26 hour fcst: -526:9254245:d=2021032312:HGT:975 mb:26 hour fcst: -527:9279788:d=2021032312:TMP:975 mb:26 hour fcst: -528:9300671:d=2021032312:RH:975 mb:26 hour fcst: -529:9319743:d=2021032312:TCDC:975 mb:26 hour fcst: -530:9327936:d=2021032312:SPFH:975 mb:26 hour fcst: -531:9354815:d=2021032312:VVEL:975 mb:26 hour fcst: -532:9377553:d=2021032312:DZDT:975 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_17 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_17 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_17 pgb2file_f026_17 0p25' -+ exglobal_atmos_products.sh[133][[ 532 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=533 -+ exglobal_atmos_products.sh[101]last=563 -+ exglobal_atmos_products.sh[102][[ 563 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 563 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 18 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 533:563 -grib tmpfile_f026_18 -533:9401027:d=2021032312:UGRD:975 mb:26 hour fcst: -534:9422425:d=2021032312:VGRD:975 mb:26 hour fcst: -535:9444294:d=2021032312:ABSV:975 mb:26 hour fcst: -536:9465303:d=2021032312:CLMR:975 mb:26 hour fcst: -537:9469867:d=2021032312:ICMR:975 mb:26 hour fcst: -538:9474731:d=2021032312:RWMR:975 mb:26 hour fcst: -539:9486294:d=2021032312:SNMR:975 mb:26 hour fcst: -540:9493402:d=2021032312:GRLE:975 mb:26 hour fcst: -541:9495684:d=2021032312:O3MR:975 mb:26 hour fcst: -542:9521523:d=2021032312:TMP:1000 mb:26 hour fcst: -543:9542318:d=2021032312:RH:1000 mb:26 hour fcst: -544:9561154:d=2021032312:TCDC:1000 mb:26 hour fcst: -545:9565568:d=2021032312:SPFH:1000 mb:26 hour fcst: -546:9592716:d=2021032312:VVEL:1000 mb:26 hour fcst: -547:9613882:d=2021032312:DZDT:1000 mb:26 hour fcst: -548:9635272:d=2021032312:UGRD:1000 mb:26 hour fcst: -549:9656081:d=2021032312:VGRD:1000 mb:26 hour fcst: -550:9677533:d=2021032312:ABSV:1000 mb:26 hour fcst: -551:9698057:d=2021032312:CLMR:1000 mb:26 hour fcst: -552:9699812:d=2021032312:ICMR:1000 mb:26 hour fcst: -553:9702180:d=2021032312:RWMR:1000 mb:26 hour fcst: -554:9711638:d=2021032312:SNMR:1000 mb:26 hour fcst: -555:9714967:d=2021032312:GRLE:1000 mb:26 hour fcst: -556:9716078:d=2021032312:O3MR:1000 mb:26 hour fcst: -557:9741905:d=2021032312:MSLET:mean sea level:26 hour fcst: -558:9768547:d=2021032312:HGT:1000 mb:26 hour fcst: -559:9794283:d=2021032312:SDEN:surface:26 hour fcst: -560:9805272:d=2021032312:REFD:4000 m above ground:26 hour fcst: -561:9818504:d=2021032312:REFD:1000 m above ground:26 hour fcst: -562:9842645:d=2021032312:PRES:surface:26 hour fcst: -563:9868350:d=2021032312:HGT:surface:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_18 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_18 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_18 pgb2file_f026_18 0p25' -+ exglobal_atmos_products.sh[133][[ 563 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=564 -+ exglobal_atmos_products.sh[101]last=594 -+ exglobal_atmos_products.sh[102][[ 594 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 594 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 19 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 564:594 -grib tmpfile_f026_19 -564:9884420:d=2021032312:TMP:surface:26 hour fcst: -565:9901056:d=2021032312:TSOIL:0-0.1 m below ground:26 hour fcst: -566:9913024:d=2021032312:SOILW:0-0.1 m below ground:26 hour fcst: -567:9921941:d=2021032312:SOILL:0-0.1 m below ground:26 hour fcst: -568:9930741:d=2021032312:TSOIL:0.1-0.4 m below ground:26 hour fcst: -569:9946234:d=2021032312:SOILW:0.1-0.4 m below ground:26 hour fcst: -570:9955248:d=2021032312:SOILL:0.1-0.4 m below ground:26 hour fcst: -571:9964140:d=2021032312:TSOIL:0.4-1 m below ground:26 hour fcst: -572:9978589:d=2021032312:SOILW:0.4-1 m below ground:26 hour fcst: -573:9987722:d=2021032312:SOILL:0.4-1 m below ground:26 hour fcst: -574:9996645:d=2021032312:TSOIL:1-2 m below ground:26 hour fcst: -575:10010935:d=2021032312:SOILW:1-2 m below ground:26 hour fcst: -576:10019992:d=2021032312:SOILL:1-2 m below ground:26 hour fcst: -577:10029069:d=2021032312:CNWAT:surface:26 hour fcst: -578:10034591:d=2021032312:WEASD:surface:26 hour fcst: -579:10047416:d=2021032312:SNOD:surface:26 hour fcst: -580:10061795:d=2021032312:PEVPR:surface:26 hour fcst: -581:10077587:d=2021032312:ICETK:surface:26 hour fcst: -582:10080995:d=2021032312:TMP:2 m above ground:26 hour fcst: -583:10101815:d=2021032312:SPFH:2 m above ground:26 hour fcst: -584:10130629:d=2021032312:DPT:2 m above ground:26 hour fcst: -585:10151891:d=2021032312:RH:2 m above ground:26 hour fcst: -586:10170405:d=2021032312:APTMP:2 m above ground:26 hour fcst: -587:10184804:d=2021032312:TMAX:2 m above ground:24-26 hour max fcst: -588:10205540:d=2021032312:TMIN:2 m above ground:24-26 hour min fcst: -589:10226332:d=2021032312:UGRD:10 m above ground:26 hour fcst: -590:10246877:d=2021032312:VGRD:10 m above ground:26 hour fcst: -591:10267795:d=2021032312:ICEG:10 m above mean sea level:26 hour fcst: -592:10268411:d=2021032312:CPOFP:surface:26 hour fcst: -593:10288757:d=2021032312:CPRAT:surface:26 hour fcst: -594:10310405:d=2021032312:PRATE:surface:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_19 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_19 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_19 pgb2file_f026_19 0p25' -+ exglobal_atmos_products.sh[133][[ 594 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=595 -+ exglobal_atmos_products.sh[101]last=625 -+ exglobal_atmos_products.sh[102][[ 625 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 625 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 20 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 595:625 -grib tmpfile_f026_20 -595:10337417:d=2021032312:CPRAT:surface:24-26 hour ave fcst: -596:10361060:d=2021032312:PRATE:surface:24-26 hour ave fcst: -597:10386191:d=2021032312:APCP:surface:24-26 hour acc fcst: -598:10393812:d=2021032312:APCP:surface:0-26 hour acc fcst: -599:10407159:d=2021032312:ACPCP:surface:24-26 hour acc fcst: -600:10413273:d=2021032312:ACPCP:surface:0-26 hour acc fcst: -601:10424419:d=2021032312:FROZR:surface:0-26 hour acc fcst: -602:10430806:d=2021032312:FRZR:surface:0-26 hour acc fcst: -603:10432762:d=2021032312:TSNOWP:surface:0-26 hour acc fcst: -604:10440477:d=2021032312:WATR:surface:24-26 hour acc fcst: -605:10453221:d=2021032312:FROZR:surface:24-26 hour acc fcst: -606:10455916:d=2021032312:FRZR:surface:24-26 hour acc fcst: -607:10456640:d=2021032312:TSNOWP:surface:24-26 hour acc fcst: -608:10463202:d=2021032312:CSNOW:surface:26 hour fcst: -609:10465031:d=2021032312:CICEP:surface:26 hour fcst: -610:10465210:d=2021032312:CFRZR:surface:26 hour fcst: -611:10465436:d=2021032312:CRAIN:surface:26 hour fcst: -612:10468950:d=2021032312:LHTFL:surface:24-26 hour ave fcst: -613:10490749:d=2021032312:SHTFL:surface:24-26 hour ave fcst: -614:10511544:d=2021032312:GFLUX:surface:24-26 hour ave fcst: -615:10523609:d=2021032312:UFLX:surface:24-26 hour ave fcst: -616:10540022:d=2021032312:VFLX:surface:24-26 hour ave fcst: -617:10556522:d=2021032312:SFCR:surface:26 hour fcst: -618:10574937:d=2021032312:FRICV:surface:26 hour fcst: -619:10596676:d=2021032312:U-GWD:surface:24-26 hour ave fcst: -620:10611591:d=2021032312:V-GWD:surface:24-26 hour ave fcst: -621:10626556:d=2021032312:VEG:surface:26 hour fcst: -622:10634590:d=2021032312:SOTYP:surface:26 hour fcst: -623:10645639:d=2021032312:WILT:surface:26 hour fcst: -624:10655549:d=2021032312:FLDCP:surface:26 hour fcst: -625:10665755:d=2021032312:SUNSD:surface:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_20 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_20 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_20 pgb2file_f026_20 0p25' -+ exglobal_atmos_products.sh[133][[ 625 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=626 -+ exglobal_atmos_products.sh[101]last=656 -+ exglobal_atmos_products.sh[102][[ 656 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 656 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 21 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 626:656 -grib tmpfile_f026_21 -626:10675642:d=2021032312:LFTX:surface:26 hour fcst: -627:10689398:d=2021032312:CAPE:surface:26 hour fcst: -628:10704479:d=2021032312:CIN:surface:26 hour fcst: -629:10717423:d=2021032312:PWAT:entire atmosphere (considered as a single layer):26 hour fcst: -630:10744337:d=2021032312:CWAT:entire atmosphere (considered as a single layer):26 hour fcst: -631:10770811:d=2021032312:RH:entire atmosphere (considered as a single layer):26 hour fcst: -632:10785933:d=2021032312:TOZNE:entire atmosphere (considered as a single layer):26 hour fcst: -633:10800248:d=2021032312:LCDC:low cloud layer:26 hour fcst: -634:10818386:d=2021032312:LCDC:low cloud layer:24-26 hour ave fcst: -635:10837082:d=2021032312:MCDC:middle cloud layer:26 hour fcst: -636:10852702:d=2021032312:MCDC:middle cloud layer:24-26 hour ave fcst: -637:10868633:d=2021032312:HCDC:high cloud layer:26 hour fcst: -638:10883483:d=2021032312:HCDC:high cloud layer:24-26 hour ave fcst: -639:10899907:d=2021032312:TCDC:entire atmosphere (considered as a single layer):26 hour fcst: -640:10916958:d=2021032312:TCDC:entire atmosphere (considered as a single layer):24-26 hour ave fcst: -641:10935020:d=2021032312:HGT:cloud ceiling:26 hour fcst: -642:10967000:d=2021032312:PRES:convective cloud bottom level:26 hour fcst: -643:10980638:d=2021032312:PRES:low cloud bottom level:24-26 hour ave fcst: -644:11002247:d=2021032312:PRES:middle cloud bottom level:24-26 hour ave fcst: -645:11019050:d=2021032312:PRES:high cloud bottom level:24-26 hour ave fcst: -646:11036486:d=2021032312:PRES:convective cloud top level:26 hour fcst: -647:11051812:d=2021032312:PRES:low cloud top level:24-26 hour ave fcst: -648:11073330:d=2021032312:PRES:middle cloud top level:24-26 hour ave fcst: -649:11089975:d=2021032312:PRES:high cloud top level:24-26 hour ave fcst: -650:11107906:d=2021032312:TMP:low cloud top level:24-26 hour ave fcst: -651:11124132:d=2021032312:TMP:middle cloud top level:24-26 hour ave fcst: -652:11137434:d=2021032312:TMP:high cloud top level:24-26 hour ave fcst: -653:11152375:d=2021032312:TCDC:convective cloud layer:26 hour fcst: -654:11172819:d=2021032312:TCDC:boundary layer cloud layer:24-26 hour ave fcst: -655:11188956:d=2021032312:CWORK:entire atmosphere (considered as a single layer):24-26 hour ave fcst: -656:11201827:d=2021032312:DSWRF:surface:24-26 hour ave fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_21 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_21 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_21 pgb2file_f026_21 0p25' -+ exglobal_atmos_products.sh[133][[ 656 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=657 -+ exglobal_atmos_products.sh[101]last=687 -+ exglobal_atmos_products.sh[102][[ 687 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 687 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -687:11918004:d=2021032312:UGRD:30 m above ground:26 hour fcst: -+ exglobal_atmos_products.sh[112]rc=0 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 0 == 0 ]] -+ exglobal_atmos_products.sh[115]last=688 -+ exglobal_atmos_products.sh[117][[ 22 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 657:688 -grib tmpfile_f026_22 -657:11220511:d=2021032312:DLWRF:surface:24-26 hour ave fcst: -658:11242319:d=2021032312:USWRF:surface:24-26 hour ave fcst: -659:11260487:d=2021032312:ULWRF:surface:24-26 hour ave fcst: -660:11279370:d=2021032312:USWRF:top of atmosphere:24-26 hour ave fcst: -661:11298376:d=2021032312:ULWRF:top of atmosphere:24-26 hour ave fcst: -662:11319468:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Total Aerosol:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -663:11346960:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Dust Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -664:11365096:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Sea Salt Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -665:11392593:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Sulphate Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -666:11418516:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Particulate Organic Matter Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -667:11444146:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Black Carbon Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -668:11466464:d=2021032312:HLCY:3000-0 m above ground:26 hour fcst: -669:11482323:d=2021032312:HLCY:1000-0 m above ground:26 hour fcst: -670:11504601:d=2021032312:USTM:6000-0 m above ground:26 hour fcst: -671:11527643:d=2021032312:VSTM:6000-0 m above ground:26 hour fcst: -672:11550447:d=2021032312:PRES:tropopause:26 hour fcst: -673:11580179:d=2021032312:ICAHT:tropopause:26 hour fcst: -674:11610979:d=2021032312:HGT:tropopause:26 hour fcst: -675:11639110:d=2021032312:TMP:tropopause:26 hour fcst: -676:11660775:d=2021032312:UGRD:tropopause:26 hour fcst: -677:11676725:d=2021032312:VGRD:tropopause:26 hour fcst: -678:11692633:d=2021032312:VWSH:tropopause:26 hour fcst: -679:11711356:d=2021032312:PRES:max wind:26 hour fcst: -680:11745774:d=2021032312:ICAHT:max wind:26 hour fcst: -681:11780819:d=2021032312:HGT:max wind:26 hour fcst: -682:11816017:d=2021032312:UGRD:max wind:26 hour fcst: -683:11832466:d=2021032312:VGRD:max wind:26 hour fcst: -684:11849890:d=2021032312:TMP:max wind:26 hour fcst: -685:11876546:d=2021032312:UGRD:20 m above ground:26 hour fcst: -686:11897066:d=2021032312:VGRD:20 m above ground:26 hour fcst: -687:11918004:d=2021032312:UGRD:30 m above ground:26 hour fcst: -688:11938951:d=2021032312:VGRD:30 m above ground:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_22 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_22 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_22 pgb2file_f026_22 0p25' -+ exglobal_atmos_products.sh[133][[ 688 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=689 -+ exglobal_atmos_products.sh[101]last=719 -+ exglobal_atmos_products.sh[102][[ 719 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 719 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 23 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 689:719 -grib tmpfile_f026_23 -689:11960294:d=2021032312:UGRD:40 m above ground:26 hour fcst: -690:11981387:d=2021032312:VGRD:40 m above ground:26 hour fcst: -691:12002645:d=2021032312:UGRD:50 m above ground:26 hour fcst: -692:12023632:d=2021032312:VGRD:50 m above ground:26 hour fcst: -693:12045154:d=2021032312:TMP:80 m above ground:26 hour fcst: -694:12065950:d=2021032312:SPFH:80 m above ground:26 hour fcst: -695:12094695:d=2021032312:PRES:80 m above ground:26 hour fcst: -696:12120355:d=2021032312:UGRD:80 m above ground:26 hour fcst: -697:12141511:d=2021032312:VGRD:80 m above ground:26 hour fcst: -698:12163219:d=2021032312:TMP:100 m above ground:26 hour fcst: -699:12183988:d=2021032312:UGRD:100 m above ground:26 hour fcst: -700:12205385:d=2021032312:VGRD:100 m above ground:26 hour fcst: -701:12227190:d=2021032312:TMP:1829 m above mean sea level:26 hour fcst: -702:12248801:d=2021032312:UGRD:1829 m above mean sea level:26 hour fcst: -703:12271055:d=2021032312:VGRD:1829 m above mean sea level:26 hour fcst: -704:12293916:d=2021032312:TMP:2743 m above mean sea level:26 hour fcst: -705:12315553:d=2021032312:UGRD:2743 m above mean sea level:26 hour fcst: -706:12338249:d=2021032312:VGRD:2743 m above mean sea level:26 hour fcst: -707:12361524:d=2021032312:TMP:3658 m above mean sea level:26 hour fcst: -708:12382871:d=2021032312:UGRD:3658 m above mean sea level:26 hour fcst: -709:12405933:d=2021032312:VGRD:3658 m above mean sea level:26 hour fcst: -710:12429740:d=2021032312:HGT:0C isotherm:26 hour fcst: -711:12451350:d=2021032312:RH:0C isotherm:26 hour fcst: -712:12470638:d=2021032312:HGT:highest tropospheric freezing level:26 hour fcst: -713:12491990:d=2021032312:RH:highest tropospheric freezing level:26 hour fcst: -714:12511217:d=2021032312:TMP:30-0 mb above ground:26 hour fcst: -715:12532041:d=2021032312:RH:30-0 mb above ground:26 hour fcst: -716:12550154:d=2021032312:SPFH:30-0 mb above ground:26 hour fcst: -717:12578977:d=2021032312:UGRD:30-0 mb above ground:26 hour fcst: -718:12600161:d=2021032312:VGRD:30-0 mb above ground:26 hour fcst: -719:12621960:d=2021032312:4LFTX:surface:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_23 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_23 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_23 pgb2file_f026_23 0p25' -+ exglobal_atmos_products.sh[133][[ 719 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=720 -+ exglobal_atmos_products.sh[101]last=750 -+ exglobal_atmos_products.sh[102][[ 750 -gt 753 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 750 tmpfile_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 24 -eq 24 ]] -+ exglobal_atmos_products.sh[118]last=753 -+ exglobal_atmos_products.sh[122]wgrib2 tmpfile_f026 -for 720:753 -grib tmpfile_f026_24 -720:12635482:d=2021032312:CAPE:180-0 mb above ground:26 hour fcst: -721:12649402:d=2021032312:CIN:180-0 mb above ground:26 hour fcst: -722:12663544:d=2021032312:HPBL:surface:26 hour fcst: -723:12694619:d=2021032312:RH:0.33-1 sigma layer:26 hour fcst: -724:12712943:d=2021032312:RH:0.44-1 sigma layer:26 hour fcst: -725:12730909:d=2021032312:RH:0.72-0.94 sigma layer:26 hour fcst: -726:12749371:d=2021032312:RH:0.44-0.72 sigma layer:26 hour fcst: -727:12768475:d=2021032312:TMP:0.995 sigma level:26 hour fcst: -728:12789353:d=2021032312:POT:0.995 sigma level:26 hour fcst: -729:12810198:d=2021032312:RH:0.995 sigma level:26 hour fcst: -730:12827947:d=2021032312:UGRD:0.995 sigma level:26 hour fcst: -731:12848980:d=2021032312:VGRD:0.995 sigma level:26 hour fcst: -732:12870525:d=2021032312:VVEL:0.995 sigma level:26 hour fcst: -733:12891445:d=2021032312:CAPE:90-0 mb above ground:26 hour fcst: -734:12904234:d=2021032312:CIN:90-0 mb above ground:26 hour fcst: -735:12918556:d=2021032312:CAPE:255-0 mb above ground:26 hour fcst: -736:12931888:d=2021032312:CIN:255-0 mb above ground:26 hour fcst: -737:12943344:d=2021032312:PLPL:255-0 mb above ground:26 hour fcst: -738:12966876:d=2021032312:LAND:surface:26 hour fcst: -739:12968817:d=2021032312:ICEC:surface:26 hour fcst: -740:12971956:d=2021032312:ALBDO:surface:24-26 hour ave fcst: -741:12987571:d=2021032312:ICETMP:surface:26 hour fcst: -742:12994294:d=2021032312:UGRD:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -743:13004950:d=2021032312:VGRD:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -744:13015642:d=2021032312:TMP:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -745:13029981:d=2021032312:HGT:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -746:13047367:d=2021032312:PRES:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -747:13065647:d=2021032312:VWSH:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -748:13077542:d=2021032312:UGRD:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -749:13088384:d=2021032312:VGRD:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -750:13099154:d=2021032312:TMP:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -751:13113653:d=2021032312:HGT:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -752:13131110:d=2021032312:PRES:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -753:13149415:d=2021032312:VWSH:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfile_f026_24 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2file_f026_24 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_24 pgb2file_f026_24 0p25' -+ exglobal_atmos_products.sh[133][[ 753 -eq 753 ]] -+ exglobal_atmos_products.sh[134](( pproc = iproc+1 )) -+ exglobal_atmos_products.sh[134](( pproc < nproc )) -+ exglobal_atmos_products.sh[137]break -+ exglobal_atmos_products.sh[142]/work2/noaa/global/mterry/global-workflow_forked/ush/run_mpmd.sh /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601/poescript -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ run_mpmd.sh[31]source /work2/noaa/global/mterry/global-workflow_forked/ush/preamble.sh -++ preamble.sh[20]set +x -Begin run_mpmd.sh at Tue Jul 29 02:27:55 UTC 2025 -+ run_mpmd.sh[33]cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601/poescript -+ run_mpmd.sh[36][[ YES != \Y\E\S ]] -+ run_mpmd.sh[46]export OMP_NUM_THREADS=1 -+ run_mpmd.sh[46]OMP_NUM_THREADS=1 -++ run_mpmd.sh[49]wc -l -+ run_mpmd.sh[49]nprocs=24 -+ run_mpmd.sh[52]mpmd_cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601/mpmd_cmdfile -+ run_mpmd.sh[53][[ -s /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601/mpmd_cmdfile ]] -+ run_mpmd.sh[55]cat - INFO: Executing MPMD job, STDOUT redirected for each process separately - INFO: On failure, logs for each job will be available in /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601/mpmd.proc_num.out - INFO: The proc_num corresponds to the line in '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601/mpmd_cmdfile' -+ run_mpmd.sh[61][[ srun -l --export=ALL --hint=nomultithread =~ ^srun.* ]] -+ run_mpmd.sh[65]nm=0 -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '0 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_1 pgb2file_f026_1 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '1 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_2 pgb2file_f026_2 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '2 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_3 pgb2file_f026_3 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '3 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_4 pgb2file_f026_4 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '4 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_5 pgb2file_f026_5 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '5 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_6 pgb2file_f026_6 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '6 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_7 pgb2file_f026_7 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '7 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_8 pgb2file_f026_8 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '8 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_9 pgb2file_f026_9 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '9 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_10 pgb2file_f026_10 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '10 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_11 pgb2file_f026_11 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '11 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_12 pgb2file_f026_12 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '12 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_13 pgb2file_f026_13 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '13 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_14 pgb2file_f026_14 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '14 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_15 pgb2file_f026_15 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '15 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_16 pgb2file_f026_16 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '16 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_17 pgb2file_f026_17 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '17 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_18 pgb2file_f026_18 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '18 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_19 pgb2file_f026_19 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '19 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_20 pgb2file_f026_20 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '20 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_21 pgb2file_f026_21 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '21 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_22 pgb2file_f026_22 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '22 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_23 pgb2file_f026_23 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '23 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfile_f026_24 pgb2file_f026_24 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[72]set +e -+ run_mpmd.sh[74]srun -l --export=ALL --hint=nomultithread --multi-prog --output=mpmd.%j.%t.out -n 24 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601/mpmd_cmdfile -+ run_mpmd.sh[75]err=0 -+ run_mpmd.sh[76]set_strict -+ preamble.sh[35][[ YES == \Y\E\S ]] -+ preamble.sh[37]set -eu -+ run_mpmd.sh[103][[ 0 -eq 0 ]] -+ run_mpmd.sh[104]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601/mpmd_cmdfile -++ run_mpmd.sh[105]find . -name 'mpmd.*.out' -+ run_mpmd.sh[105]out_files='./mpmd.5951061.17.out -./mpmd.5951061.0.out -./mpmd.5951061.6.out -./mpmd.5951061.15.out -./mpmd.5951061.9.out -./mpmd.5951061.2.out -./mpmd.5951061.3.out -./mpmd.5951061.23.out -./mpmd.5951061.4.out -./mpmd.5951061.22.out -./mpmd.5951061.1.out -./mpmd.5951061.7.out -./mpmd.5951061.21.out -./mpmd.5951061.12.out -./mpmd.5951061.19.out -./mpmd.5951061.8.out -./mpmd.5951061.16.out -./mpmd.5951061.18.out -./mpmd.5951061.13.out -./mpmd.5951061.10.out -./mpmd.5951061.20.out -./mpmd.5951061.14.out -./mpmd.5951061.5.out -./mpmd.5951061.11.out' -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.17.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.17.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.0.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.0.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.6.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.6.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.15.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.15.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.9.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.9.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.2.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.2.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.3.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.3.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.23.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.23.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.4.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.4.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.22.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.22.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.1.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.1.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.7.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.7.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.21.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.21.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.12.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.12.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.19.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.19.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.8.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.8.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.16.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.16.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.18.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.18.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.13.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.13.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.10.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.10.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.20.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.20.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.14.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.14.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.5.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.5.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.11.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.11.out -+ run_mpmd.sh[110]cat mpmd.out -17: + bash[8]'[' -z '' ']' -17: + bash[9]case "$-" in -17: + bash[12]__lmod_vx=x -17: + bash[16]'[' -n x ']' -17: + bash[16]set +x -17: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -17: Shell debugging restarted -17: + bash[224]unset __lmod_vx -17: + interp_atmos_master.sh[7]input_file=tmpfile_f026_18 -17: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_18 -17: + interp_atmos_master.sh[9]grid_string=0p25 -17: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -17: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -17: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -17: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -17: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -17: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -17: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -17: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -17: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -17: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -17: + interp_atmos_master.sh[31]IFS=: -17: + interp_atmos_master.sh[31]read -ra grids -17: + interp_atmos_master.sh[33]output_grids= -17: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -17: + interp_atmos_master.sh[35]gridopt=grid0p25 -17: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_18_0p25' -17: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_18 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_18_0p25 -17: 1:0:d=2021032312:UGRD:975 mb:26 hour fcst: -17: 2:21398:d=2021032312:VGRD:975 mb:26 hour fcst: -17: 3:43267:d=2021032312:ABSV:975 mb:26 hour fcst: -17: 4:64276:d=2021032312:CLMR:975 mb:26 hour fcst: -17: 5:68840:d=2021032312:ICMR:975 mb:26 hour fcst: -17: 6:73704:d=2021032312:RWMR:975 mb:26 hour fcst: -17: 7:85267:d=2021032312:SNMR:975 mb:26 hour fcst: -17: 8:92375:d=2021032312:GRLE:975 mb:26 hour fcst: -17: 9:94657:d=2021032312:O3MR:975 mb:26 hour fcst: -17: 10:120496:d=2021032312:TMP:1000 mb:26 hour fcst: -17: 11:141291:d=2021032312:RH:1000 mb:26 hour fcst: -17: 12:160127:d=2021032312:TCDC:1000 mb:26 hour fcst: -17: 13:164541:d=2021032312:SPFH:1000 mb:26 hour fcst: -17: 14:191689:d=2021032312:VVEL:1000 mb:26 hour fcst: -17: 15:212855:d=2021032312:DZDT:1000 mb:26 hour fcst: -17: 16:234245:d=2021032312:UGRD:1000 mb:26 hour fcst: -17: 17:255054:d=2021032312:VGRD:1000 mb:26 hour fcst: -17: 18:276506:d=2021032312:ABSV:1000 mb:26 hour fcst: -17: 19:297030:d=2021032312:CLMR:1000 mb:26 hour fcst: -17: 20:298785:d=2021032312:ICMR:1000 mb:26 hour fcst: -17: 21:301153:d=2021032312:RWMR:1000 mb:26 hour fcst: -17: 22:310611:d=2021032312:SNMR:1000 mb:26 hour fcst: -17: 23:313940:d=2021032312:GRLE:1000 mb:26 hour fcst: -17: 24:315051:d=2021032312:O3MR:1000 mb:26 hour fcst: -17: 25:340878:d=2021032312:MSLET:mean sea level:26 hour fcst: -17: 26:367520:d=2021032312:HGT:1000 mb:26 hour fcst: -17: 27:393256:d=2021032312:SDEN:surface:26 hour fcst: -17: 28:404245:d=2021032312:REFD:4000 m above ground:26 hour fcst: -17: 29:417477:d=2021032312:REFD:1000 m above ground:26 hour fcst: -17: 30:441618:d=2021032312:PRES:surface:26 hour fcst: -17: 31:467323:d=2021032312:HGT:surface:26 hour fcst: -17: + interp_atmos_master.sh[47]export err=0 -17: + interp_atmos_master.sh[47]err=0 -17: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -17: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -17: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_18_0p25 -17: + product_functions.sh[5]local filename=pgb2file_f026_18_0p25 -17: + product_functions.sh[6]wgrib2 pgb2file_f026_18_0p25 -not_if :RH: -grib pgb2file_f026_18_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_18_0p25.new -17: 1:0:d=2021032312:UGRD:975 mb:26 hour fcst: -17: 2:518480:d=2021032312:VGRD:975 mb:26 hour fcst: -17: 3:1055204:d=2021032312:ABSV:975 mb:26 hour fcst: -17: 4:1555470:d=2021032312:CLMR:975 mb:26 hour fcst: -17: 5:1655533:d=2021032312:ICMR:975 mb:26 hour fcst: -17: 6:1760283:d=2021032312:RWMR:975 mb:26 hour fcst: -17: 7:2001522:d=2021032312:SNMR:975 mb:26 hour fcst: -17: 8:2159882:d=2021032312:GRLE:975 mb:26 hour fcst: -17: 9:2201839:d=2021032312:O3MR:975 mb:26 hour fcst: -17: 10:2899964:d=2021032312:TMP:1000 mb:26 hour fcst: -17: 11:3383156:d=2021032312:RH:1000 mb:26 hour fcst: -17: 12:3803575:d=2021032312:TCDC:1000 mb:26 hour fcst: -17: 13:3883468:d=2021032312:SPFH:1000 mb:26 hour fcst: -17: 14:4614575:d=2021032312:VVEL:1000 mb:26 hour fcst: -17: 15:5130057:d=2021032312:DZDT:1000 mb:26 hour fcst: -17: 16:5644035:d=2021032312:UGRD:1000 mb:26 hour fcst: -17: 17:6150495:d=2021032312:VGRD:1000 mb:26 hour fcst: -17: 18:6672040:d=2021032312:ABSV:1000 mb:26 hour fcst: -17: 19:7156361:d=2021032312:CLMR:1000 mb:26 hour fcst: -17: 20:7193522:d=2021032312:ICMR:1000 mb:26 hour fcst: -17: 21:7243100:d=2021032312:RWMR:1000 mb:26 hour fcst: -17: 22:7436711:d=2021032312:SNMR:1000 mb:26 hour fcst: -17: 23:7505319:d=2021032312:GRLE:1000 mb:26 hour fcst: -17: 24:7524126:d=2021032312:O3MR:1000 mb:26 hour fcst: -17: 25:8220620:d=2021032312:MSLET:mean sea level:26 hour fcst: -17: 26:8942047:d=2021032312:HGT:1000 mb:26 hour fcst: -17: 27:9630043:d=2021032312:SDEN:surface:26 hour fcst: -17: 28:9969368:d=2021032312:REFD:4000 m above ground:26 hour fcst: -17: 29:10247221:d=2021032312:REFD:1000 m above ground:26 hour fcst: -17: 30:10818497:d=2021032312:PRES:surface:26 hour fcst: -17: 31:11463684:d=2021032312:HGT:surface:26 hour fcst: -17: + product_functions.sh[10]rc=0 -17: + product_functions.sh[11](( rc == 0 )) -17: + product_functions.sh[11]mv pgb2file_f026_18_0p25.new pgb2file_f026_18_0p25 -17: + product_functions.sh[12]return 0 -17: + interp_atmos_master.sh[56]export err=0 -17: + interp_atmos_master.sh[56]err=0 -17: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -17: ++ interp_atmos_master.sh[62]wc -l -17: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_18_0p25 -match 'LAND|ICEC' -17: + interp_atmos_master.sh[62]var_count=0 -17: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -17: + interp_atmos_master.sh[73]exit 0 - 0: + bash[8]'[' -z '' ']' - 0: + bash[9]case "$-" in - 0: + bash[12]__lmod_vx=x - 0: + bash[16]'[' -n x ']' - 0: + bash[16]set +x - 0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 0: Shell debugging restarted - 0: + bash[224]unset __lmod_vx - 0: + interp_atmos_master.sh[7]input_file=tmpfile_f026_1 - 0: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_1 - 0: + interp_atmos_master.sh[9]grid_string=0p25 - 0: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 0: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 0: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 0: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 0: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 0: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 0: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 0: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 0: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 0: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 0: + interp_atmos_master.sh[31]IFS=: - 0: + interp_atmos_master.sh[31]read -ra grids - 0: + interp_atmos_master.sh[33]output_grids= - 0: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 0: + interp_atmos_master.sh[35]gridopt=grid0p25 - 0: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_1_0p25' - 0: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_1 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_1_0p25 - 0: 1:0:d=2021032312:PRMSL:mean sea level:26 hour fcst: - 0: 2:25750:d=2021032312:CLMR:1 hybrid level:26 hour fcst: - 0: 3:29476:d=2021032312:ICMR:1 hybrid level:26 hour fcst: - 0: 4:36524:d=2021032312:RWMR:1 hybrid level:26 hour fcst: - 0: 5:48816:d=2021032312:SNMR:1 hybrid level:26 hour fcst: - 0: 6:58979:d=2021032312:GRLE:1 hybrid level:26 hour fcst: - 0: 7:61830:d=2021032312:REFD:1 hybrid level:26 hour fcst: - 0: 8:87479:d=2021032312:REFD:2 hybrid level:26 hour fcst: - 0: 9:113130:d=2021032312:REFC:entire atmosphere (considered as a single layer):26 hour fcst: - 0: 10:140473:d=2021032312:VIS:surface:26 hour fcst: - 0: 11:172814:d=2021032312:UGRD:planetary boundary layer:26 hour fcst: - 0: 12:186817:d=2021032312:VGRD:planetary boundary layer:26 hour fcst: - 0: 13:201364:d=2021032312:VRATE:planetary boundary layer:26 hour fcst: - 0: 14:216434:d=2021032312:GUST:surface:26 hour fcst: - 0: 15:231225:d=2021032312:HGT:0.01 mb:26 hour fcst: - 0: 16:254282:d=2021032312:TMP:0.01 mb:26 hour fcst: - 0: 17:264808:d=2021032312:RH:0.01 mb:26 hour fcst: - 0: 18:271141:d=2021032312:SPFH:0.01 mb:26 hour fcst: - 0: 19:294155:d=2021032312:VVEL:0.01 mb:26 hour fcst: - 0: 20:320852:d=2021032312:DZDT:0.01 mb:26 hour fcst: - 0: 21:348169:d=2021032312:UGRD:0.01 mb:26 hour fcst: - 0: 22:359247:d=2021032312:VGRD:0.01 mb:26 hour fcst: - 0: 23:369910:d=2021032312:ABSV:0.01 mb:26 hour fcst: - 0: 24:385354:d=2021032312:O3MR:0.01 mb:26 hour fcst: - 0: 25:399357:d=2021032312:HGT:0.02 mb:26 hour fcst: - 0: 26:424604:d=2021032312:TMP:0.02 mb:26 hour fcst: - 0: 27:439301:d=2021032312:RH:0.02 mb:26 hour fcst: - 0: 28:446712:d=2021032312:SPFH:0.02 mb:26 hour fcst: - 0: 29:470738:d=2021032312:VVEL:0.02 mb:26 hour fcst: - 0: 30:494829:d=2021032312:DZDT:0.02 mb:26 hour fcst: - 0: 31:522305:d=2021032312:UGRD:0.02 mb:26 hour fcst: - 0: 32:534793:d=2021032312:VGRD:0.02 mb:26 hour fcst: - 0: + interp_atmos_master.sh[47]export err=0 - 0: + interp_atmos_master.sh[47]err=0 - 0: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 0: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 0: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_1_0p25 - 0: + product_functions.sh[5]local filename=pgb2file_f026_1_0p25 - 0: + product_functions.sh[6]wgrib2 pgb2file_f026_1_0p25 -not_if :RH: -grib pgb2file_f026_1_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_1_0p25.new - 0: 1:0:d=2021032312:PRMSL:mean sea level:26 hour fcst: - 0: 2:689643:d=2021032312:CLMR:1 hybrid level:26 hour fcst: - 0: 3:776149:d=2021032312:ICMR:1 hybrid level:26 hour fcst: - 0: 4:932616:d=2021032312:RWMR:1 hybrid level:26 hour fcst: - 0: 5:1184244:d=2021032312:SNMR:1 hybrid level:26 hour fcst: - 0: 6:1417802:d=2021032312:GRLE:1 hybrid level:26 hour fcst: - 0: 7:1472836:d=2021032312:REFD:1 hybrid level:26 hour fcst: - 0: 8:2073872:d=2021032312:REFD:2 hybrid level:26 hour fcst: - 0: 9:2675245:d=2021032312:REFC:entire atmosphere (considered as a single layer):26 hour fcst: - 0: 10:3331149:d=2021032312:VIS:surface:26 hour fcst: - 0: 11:4212899:d=2021032312:UGRD:planetary boundary layer:26 hour fcst: - 0: 12:4512997:d=2021032312:VGRD:planetary boundary layer:26 hour fcst: - 0: 13:4823419:d=2021032312:VRATE:planetary boundary layer:26 hour fcst: - 0: 14:5139044:d=2021032312:GUST:surface:26 hour fcst: - 0: 15:5448965:d=2021032312:HGT:0.01 mb:26 hour fcst: - 0: 16:6053053:d=2021032312:TMP:0.01 mb:26 hour fcst: - 0: 17:6321605:d=2021032312:RH:0.01 mb:26 hour fcst: - 0: 18:6503414:d=2021032312:SPFH:0.01 mb:26 hour fcst: - 0: 19:7104370:d=2021032312:VVEL:0.01 mb:26 hour fcst: - 0: 20:7844292:d=2021032312:DZDT:0.01 mb:26 hour fcst: - 0: 21:8628900:d=2021032312:UGRD:0.01 mb:26 hour fcst: - 0: 22:8904592:d=2021032312:VGRD:0.01 mb:26 hour fcst: - 0: 23:9172866:d=2021032312:ABSV:0.01 mb:26 hour fcst: - 0: 24:9498841:d=2021032312:O3MR:0.01 mb:26 hour fcst: - 0: 25:9800437:d=2021032312:HGT:0.02 mb:26 hour fcst: - 0: 26:10490109:d=2021032312:TMP:0.02 mb:26 hour fcst: - 0: 27:10831925:d=2021032312:RH:0.02 mb:26 hour fcst: - 0: 28:11011834:d=2021032312:SPFH:0.02 mb:26 hour fcst: - 0: 29:11658841:d=2021032312:VVEL:0.02 mb:26 hour fcst: - 0: 30:12297663:d=2021032312:DZDT:0.02 mb:26 hour fcst: - 0: 31:13093049:d=2021032312:UGRD:0.02 mb:26 hour fcst: - 0: 32:13386177:d=2021032312:VGRD:0.02 mb:26 hour fcst: - 0: + product_functions.sh[10]rc=0 - 0: + product_functions.sh[11](( rc == 0 )) - 0: + product_functions.sh[11]mv pgb2file_f026_1_0p25.new pgb2file_f026_1_0p25 - 0: + product_functions.sh[12]return 0 - 0: + interp_atmos_master.sh[56]export err=0 - 0: + interp_atmos_master.sh[56]err=0 - 0: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 0: ++ interp_atmos_master.sh[62]wc -l - 0: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_1_0p25 -match 'LAND|ICEC' - 0: + interp_atmos_master.sh[62]var_count=0 - 0: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 0: + interp_atmos_master.sh[73]exit 0 - 6: + bash[8]'[' -z '' ']' - 6: + bash[9]case "$-" in - 6: + bash[12]__lmod_vx=x - 6: + bash[16]'[' -n x ']' - 6: + bash[16]set +x - 6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 6: Shell debugging restarted - 6: + bash[224]unset __lmod_vx - 6: + interp_atmos_master.sh[7]input_file=tmpfile_f026_7 - 6: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_7 - 6: + interp_atmos_master.sh[9]grid_string=0p25 - 6: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 6: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 6: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 6: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 6: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 6: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 6: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 6: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 6: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 6: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 6: + interp_atmos_master.sh[31]IFS=: - 6: + interp_atmos_master.sh[31]read -ra grids - 6: + interp_atmos_master.sh[33]output_grids= - 6: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 6: + interp_atmos_master.sh[35]gridopt=grid0p25 - 6: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_7_0p25' - 6: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_7 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_7_0p25 - 6: 1:0:d=2021032312:SPFH:40 mb:26 hour fcst: - 6: 2:17032:d=2021032312:VVEL:40 mb:26 hour fcst: - 6: 3:46363:d=2021032312:DZDT:40 mb:26 hour fcst: - 6: 4:78583:d=2021032312:UGRD:40 mb:26 hour fcst: - 6: 5:97293:d=2021032312:VGRD:40 mb:26 hour fcst: - 6: 6:116055:d=2021032312:ABSV:40 mb:26 hour fcst: - 6: 7:133911:d=2021032312:O3MR:40 mb:26 hour fcst: - 6: 8:160855:d=2021032312:HGT:50 mb:26 hour fcst: - 6: 9:179790:d=2021032312:TMP:50 mb:26 hour fcst: - 6: 10:195840:d=2021032312:RH:50 mb:26 hour fcst: - 6: 11:206471:d=2021032312:TCDC:50 mb:26 hour fcst: - 6: 12:206650:d=2021032312:SPFH:50 mb:26 hour fcst: - 6: 13:225125:d=2021032312:VVEL:50 mb:26 hour fcst: - 6: 14:255098:d=2021032312:DZDT:50 mb:26 hour fcst: - 6: 15:287235:d=2021032312:UGRD:50 mb:26 hour fcst: - 6: 16:305672:d=2021032312:VGRD:50 mb:26 hour fcst: - 6: 17:324147:d=2021032312:ABSV:50 mb:26 hour fcst: - 6: 18:341817:d=2021032312:CLMR:50 mb:26 hour fcst: - 6: 19:341996:d=2021032312:ICMR:50 mb:26 hour fcst: - 6: 20:342175:d=2021032312:RWMR:50 mb:26 hour fcst: - 6: 21:342354:d=2021032312:SNMR:50 mb:26 hour fcst: - 6: 22:342533:d=2021032312:GRLE:50 mb:26 hour fcst: - 6: 23:342712:d=2021032312:O3MR:50 mb:26 hour fcst: - 6: 24:369079:d=2021032312:HGT:70 mb:26 hour fcst: - 6: 25:388117:d=2021032312:TMP:70 mb:26 hour fcst: - 6: 26:404620:d=2021032312:RH:70 mb:26 hour fcst: - 6: 27:412698:d=2021032312:SPFH:70 mb:26 hour fcst: - 6: 28:433543:d=2021032312:VVEL:70 mb:26 hour fcst: - 6: 29:464751:d=2021032312:DZDT:70 mb:26 hour fcst: - 6: 30:496850:d=2021032312:UGRD:70 mb:26 hour fcst: - 6: 31:515430:d=2021032312:VGRD:70 mb:26 hour fcst: - 6: + interp_atmos_master.sh[47]export err=0 - 6: + interp_atmos_master.sh[47]err=0 - 6: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 6: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 6: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_7_0p25 - 6: + product_functions.sh[5]local filename=pgb2file_f026_7_0p25 - 6: + product_functions.sh[6]wgrib2 pgb2file_f026_7_0p25 -not_if :RH: -grib pgb2file_f026_7_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_7_0p25.new - 6: 1:0:d=2021032312:SPFH:40 mb:26 hour fcst: - 6: 2:415340:d=2021032312:VVEL:40 mb:26 hour fcst: - 6: 3:1236913:d=2021032312:DZDT:40 mb:26 hour fcst: - 6: 4:2212833:d=2021032312:UGRD:40 mb:26 hour fcst: - 6: 5:2654710:d=2021032312:VGRD:40 mb:26 hour fcst: - 6: 6:3093119:d=2021032312:ABSV:40 mb:26 hour fcst: - 6: 7:3495570:d=2021032312:O3MR:40 mb:26 hour fcst: - 6: 8:4229333:d=2021032312:HGT:50 mb:26 hour fcst: - 6: 9:4668165:d=2021032312:TMP:50 mb:26 hour fcst: - 6: 10:5022997:d=2021032312:RH:50 mb:26 hour fcst: - 6: 11:5278995:d=2021032312:TCDC:50 mb:26 hour fcst: - 6: 12:5279174:d=2021032312:SPFH:50 mb:26 hour fcst: - 6: 13:5732104:d=2021032312:VVEL:50 mb:26 hour fcst: - 6: 14:6567723:d=2021032312:DZDT:50 mb:26 hour fcst: - 6: 15:7550626:d=2021032312:UGRD:50 mb:26 hour fcst: - 6: 16:7983083:d=2021032312:VGRD:50 mb:26 hour fcst: - 6: 17:8412749:d=2021032312:ABSV:50 mb:26 hour fcst: - 6: 18:8805877:d=2021032312:CLMR:50 mb:26 hour fcst: - 6: 19:8806056:d=2021032312:ICMR:50 mb:26 hour fcst: - 6: 20:8806235:d=2021032312:RWMR:50 mb:26 hour fcst: - 6: 21:8806414:d=2021032312:SNMR:50 mb:26 hour fcst: - 6: 22:8806593:d=2021032312:GRLE:50 mb:26 hour fcst: - 6: 23:8806772:d=2021032312:O3MR:50 mb:26 hour fcst: - 6: 24:9523770:d=2021032312:HGT:70 mb:26 hour fcst: - 6: 25:9967052:d=2021032312:TMP:70 mb:26 hour fcst: - 6: 26:10335693:d=2021032312:RH:70 mb:26 hour fcst: - 6: 27:10524284:d=2021032312:SPFH:70 mb:26 hour fcst: - 6: 28:11046993:d=2021032312:VVEL:70 mb:26 hour fcst: - 6: 29:11915992:d=2021032312:DZDT:70 mb:26 hour fcst: - 6: 30:12885491:d=2021032312:UGRD:70 mb:26 hour fcst: - 6: 31:13329530:d=2021032312:VGRD:70 mb:26 hour fcst: - 6: + product_functions.sh[10]rc=0 - 6: + product_functions.sh[11](( rc == 0 )) - 6: + product_functions.sh[11]mv pgb2file_f026_7_0p25.new pgb2file_f026_7_0p25 - 6: + product_functions.sh[12]return 0 - 6: + interp_atmos_master.sh[56]export err=0 - 6: + interp_atmos_master.sh[56]err=0 - 6: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 6: ++ interp_atmos_master.sh[62]wc -l - 6: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_7_0p25 -match 'LAND|ICEC' - 6: + interp_atmos_master.sh[62]var_count=0 - 6: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 6: + interp_atmos_master.sh[73]exit 0 -15: + bash[8]'[' -z '' ']' -15: + bash[9]case "$-" in -15: + bash[12]__lmod_vx=x -15: + bash[16]'[' -n x ']' -15: + bash[16]set +x -15: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -15: Shell debugging restarted -15: + bash[224]unset __lmod_vx -15: + interp_atmos_master.sh[7]input_file=tmpfile_f026_16 -15: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_16 -15: + interp_atmos_master.sh[9]grid_string=0p25 -15: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -15: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -15: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -15: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -15: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -15: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -15: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -15: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -15: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -15: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -15: + interp_atmos_master.sh[31]IFS=: -15: + interp_atmos_master.sh[31]read -ra grids -15: + interp_atmos_master.sh[33]output_grids= -15: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -15: + interp_atmos_master.sh[35]gridopt=grid0p25 -15: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_16_0p25' -15: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_16 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_16_0p25 -15: 1:0:d=2021032312:ABSV:850 mb:26 hour fcst: -15: 2:21239:d=2021032312:CLMR:850 mb:26 hour fcst: -15: 3:34300:d=2021032312:ICMR:850 mb:26 hour fcst: -15: 4:44130:d=2021032312:RWMR:850 mb:26 hour fcst: -15: 5:54541:d=2021032312:SNMR:850 mb:26 hour fcst: -15: 6:66197:d=2021032312:GRLE:850 mb:26 hour fcst: -15: 7:68532:d=2021032312:O3MR:850 mb:26 hour fcst: -15: 8:95092:d=2021032312:HGT:900 mb:26 hour fcst: -15: 9:120150:d=2021032312:TMP:900 mb:26 hour fcst: -15: 10:141123:d=2021032312:RH:900 mb:26 hour fcst: -15: 11:160976:d=2021032312:TCDC:900 mb:26 hour fcst: -15: 12:173843:d=2021032312:SPFH:900 mb:26 hour fcst: -15: 13:203330:d=2021032312:VVEL:900 mb:26 hour fcst: -15: 14:228223:d=2021032312:DZDT:900 mb:26 hour fcst: -15: 15:254303:d=2021032312:UGRD:900 mb:26 hour fcst: -15: 16:275653:d=2021032312:VGRD:900 mb:26 hour fcst: -15: 17:297574:d=2021032312:ABSV:900 mb:26 hour fcst: -15: 18:318798:d=2021032312:CLMR:900 mb:26 hour fcst: -15: 19:328793:d=2021032312:ICMR:900 mb:26 hour fcst: -15: 20:337228:d=2021032312:RWMR:900 mb:26 hour fcst: -15: 21:349231:d=2021032312:SNMR:900 mb:26 hour fcst: -15: 22:359472:d=2021032312:GRLE:900 mb:26 hour fcst: -15: 23:362211:d=2021032312:O3MR:900 mb:26 hour fcst: -15: 24:388255:d=2021032312:HGT:925 mb:26 hour fcst: -15: 25:413472:d=2021032312:TMP:925 mb:26 hour fcst: -15: 26:434629:d=2021032312:RH:925 mb:26 hour fcst: -15: 27:454436:d=2021032312:TCDC:925 mb:26 hour fcst: -15: 28:466239:d=2021032312:SPFH:925 mb:26 hour fcst: -15: 29:495782:d=2021032312:VVEL:925 mb:26 hour fcst: -15: 30:520254:d=2021032312:DZDT:925 mb:26 hour fcst: -15: 31:545863:d=2021032312:UGRD:925 mb:26 hour fcst: -15: 32:567362:d=2021032312:VGRD:925 mb:26 hour fcst: -15: + interp_atmos_master.sh[47]export err=0 -15: + interp_atmos_master.sh[47]err=0 -15: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -15: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -15: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_16_0p25 -15: + product_functions.sh[5]local filename=pgb2file_f026_16_0p25 -15: + product_functions.sh[6]wgrib2 pgb2file_f026_16_0p25 -not_if :RH: -grib pgb2file_f026_16_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_16_0p25.new -15: 1:0:d=2021032312:ABSV:850 mb:26 hour fcst: -15: 2:506650:d=2021032312:CLMR:850 mb:26 hour fcst: -15: 3:788123:d=2021032312:ICMR:850 mb:26 hour fcst: -15: 4:1024766:d=2021032312:RWMR:850 mb:26 hour fcst: -15: 5:1225304:d=2021032312:SNMR:850 mb:26 hour fcst: -15: 6:1490149:d=2021032312:GRLE:850 mb:26 hour fcst: -15: 7:1531509:d=2021032312:O3MR:850 mb:26 hour fcst: -15: 8:2255608:d=2021032312:HGT:900 mb:26 hour fcst: -15: 9:2905031:d=2021032312:TMP:900 mb:26 hour fcst: -15: 10:3400711:d=2021032312:RH:900 mb:26 hour fcst: -15: 11:3854950:d=2021032312:TCDC:900 mb:26 hour fcst: -15: 12:4116598:d=2021032312:SPFH:900 mb:26 hour fcst: -15: 13:4921798:d=2021032312:VVEL:900 mb:26 hour fcst: -15: 14:5583638:d=2021032312:DZDT:900 mb:26 hour fcst: -15: 15:6286112:d=2021032312:UGRD:900 mb:26 hour fcst: -15: 16:6802356:d=2021032312:VGRD:900 mb:26 hour fcst: -15: 17:7346104:d=2021032312:ABSV:900 mb:26 hour fcst: -15: 18:7851105:d=2021032312:CLMR:900 mb:26 hour fcst: -15: 19:8073390:d=2021032312:ICMR:900 mb:26 hour fcst: -15: 20:8272702:d=2021032312:RWMR:900 mb:26 hour fcst: -15: 21:8511885:d=2021032312:SNMR:900 mb:26 hour fcst: -15: 22:8748814:d=2021032312:GRLE:900 mb:26 hour fcst: -15: 23:8800582:d=2021032312:O3MR:900 mb:26 hour fcst: -15: 24:9507890:d=2021032312:HGT:925 mb:26 hour fcst: -15: 25:10163796:d=2021032312:TMP:925 mb:26 hour fcst: -15: 26:10664189:d=2021032312:RH:925 mb:26 hour fcst: -15: 27:11117015:d=2021032312:TCDC:925 mb:26 hour fcst: -15: 28:11352124:d=2021032312:SPFH:925 mb:26 hour fcst: -15: 29:12177184:d=2021032312:VVEL:925 mb:26 hour fcst: -15: 30:12823459:d=2021032312:DZDT:925 mb:26 hour fcst: -15: 31:13505223:d=2021032312:UGRD:925 mb:26 hour fcst: -15: 32:14024262:d=2021032312:VGRD:925 mb:26 hour fcst: -15: + product_functions.sh[10]rc=0 -15: + product_functions.sh[11](( rc == 0 )) -15: + product_functions.sh[11]mv pgb2file_f026_16_0p25.new pgb2file_f026_16_0p25 -15: + product_functions.sh[12]return 0 -15: + interp_atmos_master.sh[56]export err=0 -15: + interp_atmos_master.sh[56]err=0 -15: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -15: ++ interp_atmos_master.sh[62]wc -l -15: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_16_0p25 -match 'LAND|ICEC' -15: + interp_atmos_master.sh[62]var_count=0 -15: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -15: + interp_atmos_master.sh[73]exit 0 - 9: + bash[8]'[' -z '' ']' - 9: + bash[9]case "$-" in - 9: + bash[12]__lmod_vx=x - 9: + bash[16]'[' -n x ']' - 9: + bash[16]set +x - 9: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 9: Shell debugging restarted - 9: + bash[224]unset __lmod_vx - 9: + interp_atmos_master.sh[7]input_file=tmpfile_f026_10 - 9: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_10 - 9: + interp_atmos_master.sh[9]grid_string=0p25 - 9: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 9: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 9: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 9: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 9: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 9: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 9: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 9: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 9: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 9: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 9: + interp_atmos_master.sh[31]IFS=: - 9: + interp_atmos_master.sh[31]read -ra grids - 9: + interp_atmos_master.sh[33]output_grids= - 9: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 9: + interp_atmos_master.sh[35]gridopt=grid0p25 - 9: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_10_0p25' - 9: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_10 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_10_0p25 - 9: 1:0:d=2021032312:RWMR:250 mb:26 hour fcst: - 9: 2:179:d=2021032312:SNMR:250 mb:26 hour fcst: - 9: 3:9066:d=2021032312:GRLE:250 mb:26 hour fcst: - 9: 4:9372:d=2021032312:O3MR:250 mb:26 hour fcst: - 9: 5:34806:d=2021032312:HGT:300 mb:26 hour fcst: - 9: 6:57125:d=2021032312:TMP:300 mb:26 hour fcst: - 9: 7:75540:d=2021032312:RH:300 mb:26 hour fcst: - 9: 8:94518:d=2021032312:TCDC:300 mb:26 hour fcst: - 9: 9:107302:d=2021032312:SPFH:300 mb:26 hour fcst: - 9: 10:133890:d=2021032312:VVEL:300 mb:26 hour fcst: - 9: 11:157741:d=2021032312:DZDT:300 mb:26 hour fcst: - 9: 12:185761:d=2021032312:UGRD:300 mb:26 hour fcst: - 9: 13:200401:d=2021032312:VGRD:300 mb:26 hour fcst: - 9: 14:215787:d=2021032312:ABSV:300 mb:26 hour fcst: - 9: 15:238117:d=2021032312:CLMR:300 mb:26 hour fcst: - 9: 16:238296:d=2021032312:ICMR:300 mb:26 hour fcst: - 9: 17:254151:d=2021032312:RWMR:300 mb:26 hour fcst: - 9: 18:254330:d=2021032312:SNMR:300 mb:26 hour fcst: - 9: 19:264733:d=2021032312:GRLE:300 mb:26 hour fcst: - 9: 20:265114:d=2021032312:O3MR:300 mb:26 hour fcst: - 9: 21:292287:d=2021032312:HGT:350 mb:26 hour fcst: - 9: 22:314451:d=2021032312:TMP:350 mb:26 hour fcst: - 9: 23:332663:d=2021032312:RH:350 mb:26 hour fcst: - 9: 24:351570:d=2021032312:TCDC:350 mb:26 hour fcst: - 9: 25:364832:d=2021032312:SPFH:350 mb:26 hour fcst: - 9: 26:392983:d=2021032312:VVEL:350 mb:26 hour fcst: - 9: 27:417479:d=2021032312:DZDT:350 mb:26 hour fcst: - 9: 28:445748:d=2021032312:UGRD:350 mb:26 hour fcst: - 9: 29:460331:d=2021032312:VGRD:350 mb:26 hour fcst: - 9: 30:475576:d=2021032312:ABSV:350 mb:26 hour fcst: - 9: 31:497869:d=2021032312:CLMR:350 mb:26 hour fcst: - 9: + interp_atmos_master.sh[47]export err=0 - 9: + interp_atmos_master.sh[47]err=0 - 9: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 9: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 9: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_10_0p25 - 9: + product_functions.sh[5]local filename=pgb2file_f026_10_0p25 - 9: + product_functions.sh[6]wgrib2 pgb2file_f026_10_0p25 -not_if :RH: -grib pgb2file_f026_10_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_10_0p25.new - 9: 1:0:d=2021032312:RWMR:250 mb:26 hour fcst: - 9: 2:179:d=2021032312:SNMR:250 mb:26 hour fcst: - 9: 3:173914:d=2021032312:GRLE:250 mb:26 hour fcst: - 9: 4:175844:d=2021032312:O3MR:250 mb:26 hour fcst: - 9: 5:840889:d=2021032312:HGT:300 mb:26 hour fcst: - 9: 6:1395472:d=2021032312:TMP:300 mb:26 hour fcst: - 9: 7:1807726:d=2021032312:RH:300 mb:26 hour fcst: - 9: 8:2232055:d=2021032312:TCDC:300 mb:26 hour fcst: - 9: 9:2468713:d=2021032312:SPFH:300 mb:26 hour fcst: - 9: 10:3175885:d=2021032312:VVEL:300 mb:26 hour fcst: - 9: 11:3807518:d=2021032312:DZDT:300 mb:26 hour fcst: - 9: 12:4593627:d=2021032312:UGRD:300 mb:26 hour fcst: - 9: 13:4910715:d=2021032312:VGRD:300 mb:26 hour fcst: - 9: 14:5244499:d=2021032312:ABSV:300 mb:26 hour fcst: - 9: 15:5796003:d=2021032312:CLMR:300 mb:26 hour fcst: - 9: 16:5796182:d=2021032312:ICMR:300 mb:26 hour fcst: - 9: 17:6138526:d=2021032312:RWMR:300 mb:26 hour fcst: - 9: 18:6138705:d=2021032312:SNMR:300 mb:26 hour fcst: - 9: 19:6343423:d=2021032312:GRLE:300 mb:26 hour fcst: - 9: 20:6347258:d=2021032312:O3MR:300 mb:26 hour fcst: - 9: 21:7066063:d=2021032312:HGT:350 mb:26 hour fcst: - 9: 22:7612923:d=2021032312:TMP:350 mb:26 hour fcst: - 9: 23:8019215:d=2021032312:RH:350 mb:26 hour fcst: - 9: 24:8448857:d=2021032312:TCDC:350 mb:26 hour fcst: - 9: 25:8690767:d=2021032312:SPFH:350 mb:26 hour fcst: - 9: 26:9437419:d=2021032312:VVEL:350 mb:26 hour fcst: - 9: 27:10090358:d=2021032312:DZDT:350 mb:26 hour fcst: - 9: 28:10884930:d=2021032312:UGRD:350 mb:26 hour fcst: - 9: 29:11196920:d=2021032312:VGRD:350 mb:26 hour fcst: - 9: 30:11527906:d=2021032312:ABSV:350 mb:26 hour fcst: - 9: 31:12074898:d=2021032312:CLMR:350 mb:26 hour fcst: - 9: + product_functions.sh[10]rc=0 - 9: + product_functions.sh[11](( rc == 0 )) - 9: + product_functions.sh[11]mv pgb2file_f026_10_0p25.new pgb2file_f026_10_0p25 - 9: + product_functions.sh[12]return 0 - 9: + interp_atmos_master.sh[56]export err=0 - 9: + interp_atmos_master.sh[56]err=0 - 9: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 9: ++ interp_atmos_master.sh[62]wc -l - 9: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_10_0p25 -match 'LAND|ICEC' - 9: + interp_atmos_master.sh[62]var_count=0 - 9: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 9: + interp_atmos_master.sh[73]exit 0 - 2: + bash[8]'[' -z '' ']' - 2: + bash[9]case "$-" in - 2: + bash[12]__lmod_vx=x - 2: + bash[16]'[' -n x ']' - 2: + bash[16]set +x - 2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 2: Shell debugging restarted - 2: + bash[224]unset __lmod_vx - 2: + interp_atmos_master.sh[7]input_file=tmpfile_f026_3 - 2: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_3 - 2: + interp_atmos_master.sh[9]grid_string=0p25 - 2: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 2: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 2: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 2: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 2: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 2: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 2: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 2: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 2: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 2: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 2: + interp_atmos_master.sh[31]IFS=: - 2: + interp_atmos_master.sh[31]read -ra grids - 2: + interp_atmos_master.sh[33]output_grids= - 2: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 2: + interp_atmos_master.sh[35]gridopt=grid0p25 - 2: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_3_0p25' - 2: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_3 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_3_0p25 - 2: 1:0:d=2021032312:O3MR:0.1 mb:26 hour fcst: - 2: 2:23531:d=2021032312:HGT:0.2 mb:26 hour fcst: - 2: 3:44574:d=2021032312:TMP:0.2 mb:26 hour fcst: - 2: 4:60693:d=2021032312:RH:0.2 mb:26 hour fcst: - 2: 5:71663:d=2021032312:SPFH:0.2 mb:26 hour fcst: - 2: 6:91462:d=2021032312:VVEL:0.2 mb:26 hour fcst: - 2: 7:120319:d=2021032312:DZDT:0.2 mb:26 hour fcst: - 2: 8:147775:d=2021032312:UGRD:0.2 mb:26 hour fcst: - 2: 9:167063:d=2021032312:VGRD:0.2 mb:26 hour fcst: - 2: 10:185952:d=2021032312:ABSV:0.2 mb:26 hour fcst: - 2: 11:203600:d=2021032312:O3MR:0.2 mb:26 hour fcst: - 2: 12:229237:d=2021032312:HGT:0.4 mb:26 hour fcst: - 2: 13:250058:d=2021032312:TMP:0.4 mb:26 hour fcst: - 2: 14:265910:d=2021032312:RH:0.4 mb:26 hour fcst: - 2: 15:274913:d=2021032312:SPFH:0.4 mb:26 hour fcst: - 2: 16:295003:d=2021032312:VVEL:0.4 mb:26 hour fcst: - 2: 17:319775:d=2021032312:DZDT:0.4 mb:26 hour fcst: - 2: 18:350374:d=2021032312:UGRD:0.4 mb:26 hour fcst: - 2: 19:369382:d=2021032312:VGRD:0.4 mb:26 hour fcst: - 2: 20:387575:d=2021032312:ABSV:0.4 mb:26 hour fcst: - 2: 21:404325:d=2021032312:O3MR:0.4 mb:26 hour fcst: - 2: 22:429111:d=2021032312:HGT:0.7 mb:26 hour fcst: - 2: 23:447186:d=2021032312:TMP:0.7 mb:26 hour fcst: - 2: 24:462905:d=2021032312:RH:0.7 mb:26 hour fcst: - 2: 25:471812:d=2021032312:SPFH:0.7 mb:26 hour fcst: - 2: 26:491066:d=2021032312:VVEL:0.7 mb:26 hour fcst: - 2: 27:516940:d=2021032312:DZDT:0.7 mb:26 hour fcst: - 2: 28:546201:d=2021032312:UGRD:0.7 mb:26 hour fcst: - 2: 29:564718:d=2021032312:VGRD:0.7 mb:26 hour fcst: - 2: 30:582545:d=2021032312:ABSV:0.7 mb:26 hour fcst: - 2: 31:598954:d=2021032312:O3MR:0.7 mb:26 hour fcst: - 2: + interp_atmos_master.sh[47]export err=0 - 2: + interp_atmos_master.sh[47]err=0 - 2: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 2: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 2: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_3_0p25 - 2: + product_functions.sh[5]local filename=pgb2file_f026_3_0p25 - 2: + product_functions.sh[6]wgrib2 pgb2file_f026_3_0p25 -not_if :RH: -grib pgb2file_f026_3_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_3_0p25.new - 2: 1:0:d=2021032312:O3MR:0.1 mb:26 hour fcst: - 2: 2:621304:d=2021032312:HGT:0.2 mb:26 hour fcst: - 2: 3:1138487:d=2021032312:TMP:0.2 mb:26 hour fcst: - 2: 4:1491458:d=2021032312:RH:0.2 mb:26 hour fcst: - 2: 5:1744977:d=2021032312:SPFH:0.2 mb:26 hour fcst: - 2: 6:2226961:d=2021032312:VVEL:0.2 mb:26 hour fcst: - 2: 7:3044464:d=2021032312:DZDT:0.2 mb:26 hour fcst: - 2: 8:3826191:d=2021032312:UGRD:0.2 mb:26 hour fcst: - 2: 9:4287336:d=2021032312:VGRD:0.2 mb:26 hour fcst: - 2: 10:4725519:d=2021032312:ABSV:0.2 mb:26 hour fcst: - 2: 11:5110070:d=2021032312:O3MR:0.2 mb:26 hour fcst: - 2: 12:5815779:d=2021032312:HGT:0.4 mb:26 hour fcst: - 2: 13:6318707:d=2021032312:TMP:0.4 mb:26 hour fcst: - 2: 14:6665978:d=2021032312:RH:0.4 mb:26 hour fcst: - 2: 15:6897936:d=2021032312:SPFH:0.4 mb:26 hour fcst: - 2: 16:7419908:d=2021032312:VVEL:0.4 mb:26 hour fcst: - 2: 17:8094235:d=2021032312:DZDT:0.4 mb:26 hour fcst: - 2: 18:9016373:d=2021032312:UGRD:0.4 mb:26 hour fcst: - 2: 19:9463048:d=2021032312:VGRD:0.4 mb:26 hour fcst: - 2: 20:9883591:d=2021032312:ABSV:0.4 mb:26 hour fcst: - 2: 21:10242723:d=2021032312:O3MR:0.4 mb:26 hour fcst: - 2: 22:10906011:d=2021032312:HGT:0.7 mb:26 hour fcst: - 2: 23:11316610:d=2021032312:TMP:0.7 mb:26 hour fcst: - 2: 24:11662469:d=2021032312:RH:0.7 mb:26 hour fcst: - 2: 25:11906459:d=2021032312:SPFH:0.7 mb:26 hour fcst: - 2: 26:12427580:d=2021032312:VVEL:0.7 mb:26 hour fcst: - 2: 27:13136036:d=2021032312:DZDT:0.7 mb:26 hour fcst: - 2: 28:13995896:d=2021032312:UGRD:0.7 mb:26 hour fcst: - 2: 29:14429214:d=2021032312:VGRD:0.7 mb:26 hour fcst: - 2: 30:14840117:d=2021032312:ABSV:0.7 mb:26 hour fcst: - 2: 31:15188862:d=2021032312:O3MR:0.7 mb:26 hour fcst: - 2: + product_functions.sh[10]rc=0 - 2: + product_functions.sh[11](( rc == 0 )) - 2: + product_functions.sh[11]mv pgb2file_f026_3_0p25.new pgb2file_f026_3_0p25 - 2: + product_functions.sh[12]return 0 - 2: + interp_atmos_master.sh[56]export err=0 - 2: + interp_atmos_master.sh[56]err=0 - 2: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 2: ++ interp_atmos_master.sh[62]wc -l - 2: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_3_0p25 -match 'LAND|ICEC' - 2: + interp_atmos_master.sh[62]var_count=0 - 2: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 2: + interp_atmos_master.sh[73]exit 0 - 3: + bash[8]'[' -z '' ']' - 3: + bash[9]case "$-" in - 3: + bash[12]__lmod_vx=x - 3: + bash[16]'[' -n x ']' - 3: + bash[16]set +x - 3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 3: Shell debugging restarted - 3: + bash[224]unset __lmod_vx - 3: + interp_atmos_master.sh[7]input_file=tmpfile_f026_4 - 3: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_4 - 3: + interp_atmos_master.sh[9]grid_string=0p25 - 3: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 3: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 3: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 3: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 3: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 3: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 3: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 3: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 3: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 3: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 3: + interp_atmos_master.sh[31]IFS=: - 3: + interp_atmos_master.sh[31]read -ra grids - 3: + interp_atmos_master.sh[33]output_grids= - 3: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 3: + interp_atmos_master.sh[35]gridopt=grid0p25 - 3: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_4_0p25' - 3: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_4 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_4_0p25 - 3: 1:0:d=2021032312:HGT:1 mb:26 hour fcst: - 3: 2:17949:d=2021032312:TMP:1 mb:26 hour fcst: - 3: 3:33754:d=2021032312:RH:1 mb:26 hour fcst: - 3: 4:43532:d=2021032312:SPFH:1 mb:26 hour fcst: - 3: 5:62095:d=2021032312:VVEL:1 mb:26 hour fcst: - 3: 6:88832:d=2021032312:DZDT:1 mb:26 hour fcst: - 3: 7:114789:d=2021032312:UGRD:1 mb:26 hour fcst: - 3: 8:133244:d=2021032312:VGRD:1 mb:26 hour fcst: - 3: 9:150876:d=2021032312:ABSV:1 mb:26 hour fcst: - 3: 10:167165:d=2021032312:O3MR:1 mb:26 hour fcst: - 3: 11:188430:d=2021032312:HGT:2 mb:26 hour fcst: - 3: 12:208635:d=2021032312:TMP:2 mb:26 hour fcst: - 3: 13:225021:d=2021032312:RH:2 mb:26 hour fcst: - 3: 14:233383:d=2021032312:SPFH:2 mb:26 hour fcst: - 3: 15:252411:d=2021032312:VVEL:2 mb:26 hour fcst: - 3: 16:281000:d=2021032312:DZDT:2 mb:26 hour fcst: - 3: 17:307704:d=2021032312:UGRD:2 mb:26 hour fcst: - 3: 18:319192:d=2021032312:VGRD:2 mb:26 hour fcst: - 3: 19:329896:d=2021032312:ABSV:2 mb:26 hour fcst: - 3: 20:346904:d=2021032312:O3MR:2 mb:26 hour fcst: - 3: 21:369415:d=2021032312:HGT:3 mb:26 hour fcst: - 3: 22:389356:d=2021032312:TMP:3 mb:26 hour fcst: - 3: 23:405766:d=2021032312:RH:3 mb:26 hour fcst: - 3: 24:411828:d=2021032312:SPFH:3 mb:26 hour fcst: - 3: 25:433017:d=2021032312:VVEL:3 mb:26 hour fcst: - 3: 26:463051:d=2021032312:DZDT:3 mb:26 hour fcst: - 3: 27:489235:d=2021032312:UGRD:3 mb:26 hour fcst: - 3: 28:500766:d=2021032312:VGRD:3 mb:26 hour fcst: - 3: 29:518728:d=2021032312:ABSV:3 mb:26 hour fcst: - 3: 30:535621:d=2021032312:O3MR:3 mb:26 hour fcst: - 3: 31:558953:d=2021032312:HGT:5 mb:26 hour fcst: - 3: + interp_atmos_master.sh[47]export err=0 - 3: + interp_atmos_master.sh[47]err=0 - 3: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 3: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 3: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_4_0p25 - 3: + product_functions.sh[5]local filename=pgb2file_f026_4_0p25 - 3: + product_functions.sh[6]wgrib2 pgb2file_f026_4_0p25 -not_if :RH: -grib pgb2file_f026_4_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_4_0p25.new - 3: 1:0:d=2021032312:HGT:1 mb:26 hour fcst: - 3: 2:405921:d=2021032312:TMP:1 mb:26 hour fcst: - 3: 3:756793:d=2021032312:RH:1 mb:26 hour fcst: - 3: 4:1014566:d=2021032312:SPFH:1 mb:26 hour fcst: - 3: 5:1509174:d=2021032312:VVEL:1 mb:26 hour fcst: - 3: 6:2253642:d=2021032312:DZDT:1 mb:26 hour fcst: - 3: 7:2976636:d=2021032312:UGRD:1 mb:26 hour fcst: - 3: 8:3403409:d=2021032312:VGRD:1 mb:26 hour fcst: - 3: 9:3808015:d=2021032312:ABSV:1 mb:26 hour fcst: - 3: 10:4150237:d=2021032312:O3MR:1 mb:26 hour fcst: - 3: 11:4683011:d=2021032312:HGT:2 mb:26 hour fcst: - 3: 12:5169366:d=2021032312:TMP:2 mb:26 hour fcst: - 3: 13:5537033:d=2021032312:RH:2 mb:26 hour fcst: - 3: 14:5782932:d=2021032312:SPFH:2 mb:26 hour fcst: - 3: 15:6266679:d=2021032312:VVEL:2 mb:26 hour fcst: - 3: 16:7060659:d=2021032312:DZDT:2 mb:26 hour fcst: - 3: 17:7809550:d=2021032312:UGRD:2 mb:26 hour fcst: - 3: 18:8081326:d=2021032312:VGRD:2 mb:26 hour fcst: - 3: 19:8348668:d=2021032312:ABSV:2 mb:26 hour fcst: - 3: 20:8719934:d=2021032312:O3MR:2 mb:26 hour fcst: - 3: 21:9301307:d=2021032312:HGT:3 mb:26 hour fcst: - 3: 22:9785031:d=2021032312:TMP:3 mb:26 hour fcst: - 3: 23:10148232:d=2021032312:RH:3 mb:26 hour fcst: - 3: 24:10317536:d=2021032312:SPFH:3 mb:26 hour fcst: - 3: 25:10863423:d=2021032312:VVEL:3 mb:26 hour fcst: - 3: 26:11703475:d=2021032312:DZDT:3 mb:26 hour fcst: - 3: 27:12436535:d=2021032312:UGRD:3 mb:26 hour fcst: - 3: 28:12887086:d=2021032312:VGRD:3 mb:26 hour fcst: - 3: 29:13299040:d=2021032312:ABSV:3 mb:26 hour fcst: - 3: 30:13663621:d=2021032312:O3MR:3 mb:26 hour fcst: - 3: 31:14264016:d=2021032312:HGT:5 mb:26 hour fcst: - 3: + product_functions.sh[10]rc=0 - 3: + product_functions.sh[11](( rc == 0 )) - 3: + product_functions.sh[11]mv pgb2file_f026_4_0p25.new pgb2file_f026_4_0p25 - 3: + product_functions.sh[12]return 0 - 3: + interp_atmos_master.sh[56]export err=0 - 3: + interp_atmos_master.sh[56]err=0 - 3: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 3: ++ interp_atmos_master.sh[62]wc -l - 3: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_4_0p25 -match 'LAND|ICEC' - 3: + interp_atmos_master.sh[62]var_count=0 - 3: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 3: + interp_atmos_master.sh[73]exit 0 -23: + bash[8]'[' -z '' ']' -23: + bash[9]case "$-" in -23: + bash[12]__lmod_vx=x -23: + bash[16]'[' -n x ']' -23: + bash[16]set +x -23: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -23: Shell debugging restarted -23: + bash[224]unset __lmod_vx -23: + interp_atmos_master.sh[7]input_file=tmpfile_f026_24 -23: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_24 -23: + interp_atmos_master.sh[9]grid_string=0p25 -23: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -23: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -23: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -23: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -23: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -23: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -23: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -23: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -23: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -23: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -23: + interp_atmos_master.sh[31]IFS=: -23: + interp_atmos_master.sh[31]read -ra grids -23: + interp_atmos_master.sh[33]output_grids= -23: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -23: + interp_atmos_master.sh[35]gridopt=grid0p25 -23: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_24_0p25' -23: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_24 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_24_0p25 -23: 1:0:d=2021032312:CAPE:180-0 mb above ground:26 hour fcst: -23: 2:13920:d=2021032312:CIN:180-0 mb above ground:26 hour fcst: -23: 3:28062:d=2021032312:HPBL:surface:26 hour fcst: -23: 4:59137:d=2021032312:RH:0.33-1 sigma layer:26 hour fcst: -23: 5:77461:d=2021032312:RH:0.44-1 sigma layer:26 hour fcst: -23: 6:95427:d=2021032312:RH:0.72-0.94 sigma layer:26 hour fcst: -23: 7:113889:d=2021032312:RH:0.44-0.72 sigma layer:26 hour fcst: -23: 8:132993:d=2021032312:TMP:0.995 sigma level:26 hour fcst: -23: 9:153871:d=2021032312:POT:0.995 sigma level:26 hour fcst: -23: 10:174716:d=2021032312:RH:0.995 sigma level:26 hour fcst: -23: 11:192465:d=2021032312:UGRD:0.995 sigma level:26 hour fcst: -23: 12:213498:d=2021032312:VGRD:0.995 sigma level:26 hour fcst: -23: 13:235043:d=2021032312:VVEL:0.995 sigma level:26 hour fcst: -23: 14:255963:d=2021032312:CAPE:90-0 mb above ground:26 hour fcst: -23: 15:268752:d=2021032312:CIN:90-0 mb above ground:26 hour fcst: -23: 16:283074:d=2021032312:CAPE:255-0 mb above ground:26 hour fcst: -23: 17:296406:d=2021032312:CIN:255-0 mb above ground:26 hour fcst: -23: 18:307862:d=2021032312:PLPL:255-0 mb above ground:26 hour fcst: -23: 19:331394:d=2021032312:LAND:surface:26 hour fcst: -23: 20:333335:d=2021032312:ICEC:surface:26 hour fcst: -23: 21:336474:d=2021032312:ALBDO:surface:24-26 hour ave fcst: -23: 22:352089:d=2021032312:ICETMP:surface:26 hour fcst: -23: 23:358812:d=2021032312:UGRD:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 24:369468:d=2021032312:VGRD:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 25:380160:d=2021032312:TMP:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 26:394499:d=2021032312:HGT:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 27:411885:d=2021032312:PRES:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 28:430165:d=2021032312:VWSH:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 29:442060:d=2021032312:UGRD:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 30:452902:d=2021032312:VGRD:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 31:463672:d=2021032312:TMP:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 32:478171:d=2021032312:HGT:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 33:495628:d=2021032312:PRES:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 34:513933:d=2021032312:VWSH:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: + interp_atmos_master.sh[47]export err=0 -23: + interp_atmos_master.sh[47]err=0 -23: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -23: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -23: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_24_0p25 -23: + product_functions.sh[5]local filename=pgb2file_f026_24_0p25 -23: + product_functions.sh[6]wgrib2 pgb2file_f026_24_0p25 -not_if :RH: -grib pgb2file_f026_24_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_24_0p25.new -23: 1:0:d=2021032312:CAPE:180-0 mb above ground:26 hour fcst: -23: 2:303765:d=2021032312:CIN:180-0 mb above ground:26 hour fcst: -23: 3:589822:d=2021032312:HPBL:surface:26 hour fcst: -23: 4:1453044:d=2021032312:RH:0.33-1 sigma layer:26 hour fcst: -23: 5:1863549:d=2021032312:RH:0.44-1 sigma layer:26 hour fcst: -23: 6:2263509:d=2021032312:RH:0.72-0.94 sigma layer:26 hour fcst: -23: 7:2674625:d=2021032312:RH:0.44-0.72 sigma layer:26 hour fcst: -23: 8:3113295:d=2021032312:TMP:0.995 sigma level:26 hour fcst: -23: 9:3599371:d=2021032312:POT:0.995 sigma level:26 hour fcst: -23: 10:4089435:d=2021032312:RH:0.995 sigma level:26 hour fcst: -23: 11:4476806:d=2021032312:UGRD:0.995 sigma level:26 hour fcst: -23: 12:4991461:d=2021032312:VGRD:0.995 sigma level:26 hour fcst: -23: 13:5520308:d=2021032312:VVEL:0.995 sigma level:26 hour fcst: -23: 14:6016936:d=2021032312:CAPE:90-0 mb above ground:26 hour fcst: -23: 15:6292604:d=2021032312:CIN:90-0 mb above ground:26 hour fcst: -23: 16:6590237:d=2021032312:CAPE:255-0 mb above ground:26 hour fcst: -23: 17:6889817:d=2021032312:CIN:255-0 mb above ground:26 hour fcst: -23: 18:7122840:d=2021032312:PLPL:255-0 mb above ground:26 hour fcst: -23: 19:7679753:d=2021032312:LAND:surface:26 hour fcst: -23: 20:7699673:d=2021032312:ICEC:surface:26 hour fcst: -23: 21:7754718:d=2021032312:ALBDO:surface:24-26 hour ave fcst: -23: 22:8111614:d=2021032312:ICETMP:surface:26 hour fcst: -23: 23:8334160:d=2021032312:UGRD:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 24:8647665:d=2021032312:VGRD:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 25:8959856:d=2021032312:TMP:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 26:9384489:d=2021032312:HGT:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 27:9934972:d=2021032312:PRES:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 28:10517642:d=2021032312:VWSH:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 29:10867893:d=2021032312:UGRD:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 30:11187462:d=2021032312:VGRD:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 31:11503161:d=2021032312:TMP:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 32:11931725:d=2021032312:HGT:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 33:12484627:d=2021032312:PRES:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 34:13060378:d=2021032312:VWSH:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: + product_functions.sh[10]rc=0 -23: + product_functions.sh[11](( rc == 0 )) -23: + product_functions.sh[11]mv pgb2file_f026_24_0p25.new pgb2file_f026_24_0p25 -23: + product_functions.sh[12]return 0 -23: + interp_atmos_master.sh[56]export err=0 -23: + interp_atmos_master.sh[56]err=0 -23: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -23: ++ interp_atmos_master.sh[62]wc -l -23: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_24_0p25 -match 'LAND|ICEC' -23: + interp_atmos_master.sh[62]var_count=2 -23: + interp_atmos_master.sh[63][[ 2 -eq 2 ]] -23: + interp_atmos_master.sh[64]mod_icec pgb2file_f026_24_0p25 -23: + product_functions.sh[17]local filename=pgb2file_f026_24_0p25 -23: + product_functions.sh[18]wgrib2 pgb2file_f026_24_0p25 -if LAND -rpn sto_1 -fi -if ICEC -rpn 'rcl_1:0:==:*' -fi -set_grib_type same -set_scaling same same -grib_out pgb2file_f026_24_0p25.new -23: 1:0:d=2021032312:CAPE:180-0 mb above ground:26 hour fcst: -23: 2:303765:d=2021032312:CIN:180-0 mb above ground:26 hour fcst: -23: 3:589822:d=2021032312:HPBL:surface:26 hour fcst: -23: 4:1453044:d=2021032312:RH:0.33-1 sigma layer:26 hour fcst: -23: 5:1863549:d=2021032312:RH:0.44-1 sigma layer:26 hour fcst: -23: 6:2263509:d=2021032312:RH:0.72-0.94 sigma layer:26 hour fcst: -23: 7:2674625:d=2021032312:RH:0.44-0.72 sigma layer:26 hour fcst: -23: 8:3113295:d=2021032312:TMP:0.995 sigma level:26 hour fcst: -23: 9:3599371:d=2021032312:POT:0.995 sigma level:26 hour fcst: -23: 10:4089435:d=2021032312:RH:0.995 sigma level:26 hour fcst: -23: 11:4476806:d=2021032312:UGRD:0.995 sigma level:26 hour fcst: -23: 12:4991461:d=2021032312:VGRD:0.995 sigma level:26 hour fcst: -23: 13:5520308:d=2021032312:VVEL:0.995 sigma level:26 hour fcst: -23: 14:6016936:d=2021032312:CAPE:90-0 mb above ground:26 hour fcst: -23: 15:6292604:d=2021032312:CIN:90-0 mb above ground:26 hour fcst: -23: 16:6590237:d=2021032312:CAPE:255-0 mb above ground:26 hour fcst: -23: 17:6889817:d=2021032312:CIN:255-0 mb above ground:26 hour fcst: -23: 18:7122840:d=2021032312:PLPL:255-0 mb above ground:26 hour fcst: -23: 19:7679753:d=2021032312:LAND:surface:26 hour fcst: -23: 20:7699673:d=2021032312:ICEC:surface:26 hour fcst: -23: 21:7754718:d=2021032312:ALBDO:surface:24-26 hour ave fcst: -23: 22:8111614:d=2021032312:ICETMP:surface:26 hour fcst: -23: 23:8334160:d=2021032312:UGRD:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 24:8647665:d=2021032312:VGRD:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 25:8959856:d=2021032312:TMP:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 26:9384489:d=2021032312:HGT:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 27:9934972:d=2021032312:PRES:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 28:10517642:d=2021032312:VWSH:PV=2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 29:10867893:d=2021032312:UGRD:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 30:11187462:d=2021032312:VGRD:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 31:11503161:d=2021032312:TMP:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 32:11931725:d=2021032312:HGT:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 33:12484627:d=2021032312:PRES:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 34:13060378:d=2021032312:VWSH:PV=-2e-06 (Km^2/kg/s) surface:26 hour fcst: -23: + product_functions.sh[24]rc=0 -23: + product_functions.sh[25](( rc == 0 )) -23: + product_functions.sh[25]mv pgb2file_f026_24_0p25.new pgb2file_f026_24_0p25 -23: + product_functions.sh[26]return 0 -23: + interp_atmos_master.sh[65]export err=0 -23: + interp_atmos_master.sh[65]err=0 -23: + interp_atmos_master.sh[66][[ 0 -ne 0 ]] -23: + interp_atmos_master.sh[73]exit 0 - 4: + bash[8]'[' -z '' ']' - 4: + bash[9]case "$-" in - 4: + bash[12]__lmod_vx=x - 4: + bash[16]'[' -n x ']' - 4: + bash[16]set +x - 4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 4: Shell debugging restarted - 4: + bash[224]unset __lmod_vx - 4: + interp_atmos_master.sh[7]input_file=tmpfile_f026_5 - 4: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_5 - 4: + interp_atmos_master.sh[9]grid_string=0p25 - 4: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 4: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 4: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 4: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 4: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 4: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 4: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 4: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 4: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 4: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 4: + interp_atmos_master.sh[31]IFS=: - 4: + interp_atmos_master.sh[31]read -ra grids - 4: + interp_atmos_master.sh[33]output_grids= - 4: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 4: + interp_atmos_master.sh[35]gridopt=grid0p25 - 4: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_5_0p25' - 4: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_5 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_5_0p25 - 4: 1:0:d=2021032312:TMP:5 mb:26 hour fcst: - 4: 2:16634:d=2021032312:RH:5 mb:26 hour fcst: - 4: 3:25811:d=2021032312:SPFH:5 mb:26 hour fcst: - 4: 4:46606:d=2021032312:VVEL:5 mb:26 hour fcst: - 4: 5:75745:d=2021032312:DZDT:5 mb:26 hour fcst: - 4: 6:101733:d=2021032312:UGRD:5 mb:26 hour fcst: - 4: 7:113232:d=2021032312:VGRD:5 mb:26 hour fcst: - 4: 8:132087:d=2021032312:ABSV:5 mb:26 hour fcst: - 4: 9:149699:d=2021032312:O3MR:5 mb:26 hour fcst: - 4: 10:172324:d=2021032312:HGT:7 mb:26 hour fcst: - 4: 11:189982:d=2021032312:TMP:7 mb:26 hour fcst: - 4: 12:206124:d=2021032312:RH:7 mb:26 hour fcst: - 4: 13:212184:d=2021032312:SPFH:7 mb:26 hour fcst: - 4: 14:231577:d=2021032312:VVEL:7 mb:26 hour fcst: - 4: 15:256184:d=2021032312:DZDT:7 mb:26 hour fcst: - 4: 16:281695:d=2021032312:UGRD:7 mb:26 hour fcst: - 4: 17:293089:d=2021032312:VGRD:7 mb:26 hour fcst: - 4: 18:311905:d=2021032312:ABSV:7 mb:26 hour fcst: - 4: 19:329636:d=2021032312:O3MR:7 mb:26 hour fcst: - 4: 20:353227:d=2021032312:HGT:10 mb:26 hour fcst: - 4: 21:373031:d=2021032312:TMP:10 mb:26 hour fcst: - 4: 22:389182:d=2021032312:RH:10 mb:26 hour fcst: - 4: 23:396682:d=2021032312:SPFH:10 mb:26 hour fcst: - 4: 24:414946:d=2021032312:VVEL:10 mb:26 hour fcst: - 4: 25:440465:d=2021032312:DZDT:10 mb:26 hour fcst: - 4: 26:465740:d=2021032312:UGRD:10 mb:26 hour fcst: - 4: 27:477084:d=2021032312:VGRD:10 mb:26 hour fcst: - 4: 28:495604:d=2021032312:ABSV:10 mb:26 hour fcst: - 4: 29:513257:d=2021032312:O3MR:10 mb:26 hour fcst: - 4: 30:531939:d=2021032312:HGT:15 mb:26 hour fcst: - 4: 31:551497:d=2021032312:TMP:15 mb:26 hour fcst: - 4: + interp_atmos_master.sh[47]export err=0 - 4: + interp_atmos_master.sh[47]err=0 - 4: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 4: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 4: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_5_0p25 - 4: + product_functions.sh[5]local filename=pgb2file_f026_5_0p25 - 4: + product_functions.sh[6]wgrib2 pgb2file_f026_5_0p25 -not_if :RH: -grib pgb2file_f026_5_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_5_0p25.new - 4: 1:0:d=2021032312:TMP:5 mb:26 hour fcst: - 4: 2:365422:d=2021032312:RH:5 mb:26 hour fcst: - 4: 3:616409:d=2021032312:SPFH:5 mb:26 hour fcst: - 4: 4:1138395:d=2021032312:VVEL:5 mb:26 hour fcst: - 4: 5:1961941:d=2021032312:DZDT:5 mb:26 hour fcst: - 4: 6:2686220:d=2021032312:UGRD:5 mb:26 hour fcst: - 4: 7:3134521:d=2021032312:VGRD:5 mb:26 hour fcst: - 4: 8:3561400:d=2021032312:ABSV:5 mb:26 hour fcst: - 4: 9:3946819:d=2021032312:O3MR:5 mb:26 hour fcst: - 4: 10:4511982:d=2021032312:HGT:7 mb:26 hour fcst: - 4: 11:4908716:d=2021032312:TMP:7 mb:26 hour fcst: - 4: 12:5259769:d=2021032312:RH:7 mb:26 hour fcst: - 4: 13:5433708:d=2021032312:SPFH:7 mb:26 hour fcst: - 4: 14:5914782:d=2021032312:VVEL:7 mb:26 hour fcst: - 4: 15:6577268:d=2021032312:DZDT:7 mb:26 hour fcst: - 4: 16:7282131:d=2021032312:UGRD:7 mb:26 hour fcst: - 4: 17:7724477:d=2021032312:VGRD:7 mb:26 hour fcst: - 4: 18:8155686:d=2021032312:ABSV:7 mb:26 hour fcst: - 4: 19:8543826:d=2021032312:O3MR:7 mb:26 hour fcst: - 4: 20:9147344:d=2021032312:HGT:10 mb:26 hour fcst: - 4: 21:9617145:d=2021032312:TMP:10 mb:26 hour fcst: - 4: 22:9969512:d=2021032312:RH:10 mb:26 hour fcst: - 4: 23:10193654:d=2021032312:SPFH:10 mb:26 hour fcst: - 4: 24:10638218:d=2021032312:VVEL:10 mb:26 hour fcst: - 4: 25:11332588:d=2021032312:DZDT:10 mb:26 hour fcst: - 4: 26:12018841:d=2021032312:UGRD:10 mb:26 hour fcst: - 4: 27:12462007:d=2021032312:VGRD:10 mb:26 hour fcst: - 4: 28:12888117:d=2021032312:ABSV:10 mb:26 hour fcst: - 4: 29:13278368:d=2021032312:O3MR:10 mb:26 hour fcst: - 4: 30:13703837:d=2021032312:HGT:15 mb:26 hour fcst: - 4: 31:14163866:d=2021032312:TMP:15 mb:26 hour fcst: - 4: + product_functions.sh[10]rc=0 - 4: + product_functions.sh[11](( rc == 0 )) - 4: + product_functions.sh[11]mv pgb2file_f026_5_0p25.new pgb2file_f026_5_0p25 - 4: + product_functions.sh[12]return 0 - 4: + interp_atmos_master.sh[56]export err=0 - 4: + interp_atmos_master.sh[56]err=0 - 4: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 4: ++ interp_atmos_master.sh[62]wc -l - 4: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_5_0p25 -match 'LAND|ICEC' - 4: + interp_atmos_master.sh[62]var_count=0 - 4: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 4: + interp_atmos_master.sh[73]exit 0 -22: + bash[8]'[' -z '' ']' -22: + bash[9]case "$-" in -22: + bash[12]__lmod_vx=x -22: + bash[16]'[' -n x ']' -22: + bash[16]set +x -22: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -22: Shell debugging restarted -22: + bash[224]unset __lmod_vx -22: + interp_atmos_master.sh[7]input_file=tmpfile_f026_23 -22: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_23 -22: + interp_atmos_master.sh[9]grid_string=0p25 -22: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -22: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -22: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -22: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -22: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -22: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -22: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -22: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -22: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -22: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -22: + interp_atmos_master.sh[31]IFS=: -22: + interp_atmos_master.sh[31]read -ra grids -22: + interp_atmos_master.sh[33]output_grids= -22: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -22: + interp_atmos_master.sh[35]gridopt=grid0p25 -22: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_23_0p25' -22: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_23 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_23_0p25 -22: 1:0:d=2021032312:UGRD:40 m above ground:26 hour fcst: -22: 2:21093:d=2021032312:VGRD:40 m above ground:26 hour fcst: -22: 3:42351:d=2021032312:UGRD:50 m above ground:26 hour fcst: -22: 4:63338:d=2021032312:VGRD:50 m above ground:26 hour fcst: -22: 5:84860:d=2021032312:TMP:80 m above ground:26 hour fcst: -22: 6:105656:d=2021032312:SPFH:80 m above ground:26 hour fcst: -22: 7:134401:d=2021032312:PRES:80 m above ground:26 hour fcst: -22: 8:160061:d=2021032312:UGRD:80 m above ground:26 hour fcst: -22: 9:181217:d=2021032312:VGRD:80 m above ground:26 hour fcst: -22: 10:202925:d=2021032312:TMP:100 m above ground:26 hour fcst: -22: 11:223694:d=2021032312:UGRD:100 m above ground:26 hour fcst: -22: 12:245091:d=2021032312:VGRD:100 m above ground:26 hour fcst: -22: 13:266896:d=2021032312:TMP:1829 m above mean sea level:26 hour fcst: -22: 14:288507:d=2021032312:UGRD:1829 m above mean sea level:26 hour fcst: -22: 15:310761:d=2021032312:VGRD:1829 m above mean sea level:26 hour fcst: -22: 16:333622:d=2021032312:TMP:2743 m above mean sea level:26 hour fcst: -22: 17:355259:d=2021032312:UGRD:2743 m above mean sea level:26 hour fcst: -22: 18:377955:d=2021032312:VGRD:2743 m above mean sea level:26 hour fcst: -22: 19:401230:d=2021032312:TMP:3658 m above mean sea level:26 hour fcst: -22: 20:422577:d=2021032312:UGRD:3658 m above mean sea level:26 hour fcst: -22: 21:445639:d=2021032312:VGRD:3658 m above mean sea level:26 hour fcst: -22: 22:469446:d=2021032312:HGT:0C isotherm:26 hour fcst: -22: 23:491056:d=2021032312:RH:0C isotherm:26 hour fcst: -22: 24:510344:d=2021032312:HGT:highest tropospheric freezing level:26 hour fcst: -22: 25:531696:d=2021032312:RH:highest tropospheric freezing level:26 hour fcst: -22: 26:550923:d=2021032312:TMP:30-0 mb above ground:26 hour fcst: -22: 27:571747:d=2021032312:RH:30-0 mb above ground:26 hour fcst: -22: 28:589860:d=2021032312:SPFH:30-0 mb above ground:26 hour fcst: -22: 29:618683:d=2021032312:UGRD:30-0 mb above ground:26 hour fcst: -22: 30:639867:d=2021032312:VGRD:30-0 mb above ground:26 hour fcst: -22: 31:661666:d=2021032312:4LFTX:surface:26 hour fcst: -22: + interp_atmos_master.sh[47]export err=0 -22: + interp_atmos_master.sh[47]err=0 -22: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -22: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -22: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_23_0p25 -22: + product_functions.sh[5]local filename=pgb2file_f026_23_0p25 -22: + product_functions.sh[6]wgrib2 pgb2file_f026_23_0p25 -not_if :RH: -grib pgb2file_f026_23_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_23_0p25.new -22: 1:0:d=2021032312:UGRD:40 m above ground:26 hour fcst: -22: 2:509038:d=2021032312:VGRD:40 m above ground:26 hour fcst: -22: 3:1033756:d=2021032312:UGRD:50 m above ground:26 hour fcst: -22: 4:1545035:d=2021032312:VGRD:50 m above ground:26 hour fcst: -22: 5:2073286:d=2021032312:TMP:80 m above ground:26 hour fcst: -22: 6:2556612:d=2021032312:SPFH:80 m above ground:26 hour fcst: -22: 7:3338652:d=2021032312:PRES:80 m above ground:26 hour fcst: -22: 8:3981973:d=2021032312:UGRD:80 m above ground:26 hour fcst: -22: 9:4498059:d=2021032312:VGRD:80 m above ground:26 hour fcst: -22: 10:5031743:d=2021032312:TMP:100 m above ground:26 hour fcst: -22: 11:5514037:d=2021032312:UGRD:100 m above ground:26 hour fcst: -22: 12:6032170:d=2021032312:VGRD:100 m above ground:26 hour fcst: -22: 13:6569080:d=2021032312:TMP:1829 m above mean sea level:26 hour fcst: -22: 14:7169487:d=2021032312:UGRD:1829 m above mean sea level:26 hour fcst: -22: 15:7797963:d=2021032312:VGRD:1829 m above mean sea level:26 hour fcst: -22: 16:8456270:d=2021032312:TMP:2743 m above mean sea level:26 hour fcst: -22: 17:9040472:d=2021032312:UGRD:2743 m above mean sea level:26 hour fcst: -22: 18:9669890:d=2021032312:VGRD:2743 m above mean sea level:26 hour fcst: -22: 19:10334742:d=2021032312:TMP:3658 m above mean sea level:26 hour fcst: -22: 20:10910641:d=2021032312:UGRD:3658 m above mean sea level:26 hour fcst: -22: 21:11554914:d=2021032312:VGRD:3658 m above mean sea level:26 hour fcst: -22: 22:12228997:d=2021032312:HGT:0C isotherm:26 hour fcst: -22: 23:12820057:d=2021032312:RH:0C isotherm:26 hour fcst: -22: 24:13254265:d=2021032312:HGT:highest tropospheric freezing level:26 hour fcst: -22: 25:13834109:d=2021032312:RH:highest tropospheric freezing level:26 hour fcst: -22: 26:14273395:d=2021032312:TMP:30-0 mb above ground:26 hour fcst: -22: 27:14756784:d=2021032312:RH:30-0 mb above ground:26 hour fcst: -22: 28:15152334:d=2021032312:SPFH:30-0 mb above ground:26 hour fcst: -22: 29:15936659:d=2021032312:UGRD:30-0 mb above ground:26 hour fcst: -22: 30:16453514:d=2021032312:VGRD:30-0 mb above ground:26 hour fcst: -22: 31:16989357:d=2021032312:4LFTX:surface:26 hour fcst: -22: + product_functions.sh[10]rc=0 -22: + product_functions.sh[11](( rc == 0 )) -22: + product_functions.sh[11]mv pgb2file_f026_23_0p25.new pgb2file_f026_23_0p25 -22: + product_functions.sh[12]return 0 -22: + interp_atmos_master.sh[56]export err=0 -22: + interp_atmos_master.sh[56]err=0 -22: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -22: ++ interp_atmos_master.sh[62]wc -l -22: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_23_0p25 -match 'LAND|ICEC' -22: + interp_atmos_master.sh[62]var_count=0 -22: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -22: + interp_atmos_master.sh[73]exit 0 - 1: + bash[8]'[' -z '' ']' - 1: + bash[9]case "$-" in - 1: + bash[12]__lmod_vx=x - 1: + bash[16]'[' -n x ']' - 1: + bash[16]set +x - 1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 1: Shell debugging restarted - 1: + bash[224]unset __lmod_vx - 1: + interp_atmos_master.sh[7]input_file=tmpfile_f026_2 - 1: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_2 - 1: + interp_atmos_master.sh[9]grid_string=0p25 - 1: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 1: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 1: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 1: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 1: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 1: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 1: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 1: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 1: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 1: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 1: + interp_atmos_master.sh[31]IFS=: - 1: + interp_atmos_master.sh[31]read -ra grids - 1: + interp_atmos_master.sh[33]output_grids= - 1: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 1: + interp_atmos_master.sh[35]gridopt=grid0p25 - 1: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_2_0p25' - 1: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_2 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_2_0p25 - 1: 1:0:d=2021032312:ABSV:0.02 mb:26 hour fcst: - 1: 2:16932:d=2021032312:O3MR:0.02 mb:26 hour fcst: - 1: 3:32488:d=2021032312:HGT:0.04 mb:26 hour fcst: - 1: 4:56638:d=2021032312:TMP:0.04 mb:26 hour fcst: - 1: 5:73787:d=2021032312:RH:0.04 mb:26 hour fcst: - 1: 6:79825:d=2021032312:SPFH:0.04 mb:26 hour fcst: - 1: 7:105356:d=2021032312:VVEL:0.04 mb:26 hour fcst: - 1: 8:132853:d=2021032312:DZDT:0.04 mb:26 hour fcst: - 1: 9:157408:d=2021032312:UGRD:0.04 mb:26 hour fcst: - 1: 10:177281:d=2021032312:VGRD:0.04 mb:26 hour fcst: - 1: 11:197017:d=2021032312:ABSV:0.04 mb:26 hour fcst: - 1: 12:215262:d=2021032312:O3MR:0.04 mb:26 hour fcst: - 1: 13:238759:d=2021032312:HGT:0.07 mb:26 hour fcst: - 1: 14:262722:d=2021032312:TMP:0.07 mb:26 hour fcst: - 1: 15:279664:d=2021032312:RH:0.07 mb:26 hour fcst: - 1: 16:289172:d=2021032312:SPFH:0.07 mb:26 hour fcst: - 1: 17:312656:d=2021032312:VVEL:0.07 mb:26 hour fcst: - 1: 18:341415:d=2021032312:DZDT:0.07 mb:26 hour fcst: - 1: 19:366757:d=2021032312:UGRD:0.07 mb:26 hour fcst: - 1: 20:386992:d=2021032312:VGRD:0.07 mb:26 hour fcst: - 1: 21:407205:d=2021032312:ABSV:0.07 mb:26 hour fcst: - 1: 22:425981:d=2021032312:O3MR:0.07 mb:26 hour fcst: - 1: 23:445398:d=2021032312:HGT:0.1 mb:26 hour fcst: - 1: 24:466813:d=2021032312:TMP:0.1 mb:26 hour fcst: - 1: 25:483226:d=2021032312:RH:0.1 mb:26 hour fcst: - 1: 26:491312:d=2021032312:SPFH:0.1 mb:26 hour fcst: - 1: 27:512619:d=2021032312:VVEL:0.1 mb:26 hour fcst: - 1: 28:542167:d=2021032312:DZDT:0.1 mb:26 hour fcst: - 1: 29:567776:d=2021032312:UGRD:0.1 mb:26 hour fcst: - 1: 30:587777:d=2021032312:VGRD:0.1 mb:26 hour fcst: - 1: 31:607800:d=2021032312:ABSV:0.1 mb:26 hour fcst: - 1: + interp_atmos_master.sh[47]export err=0 - 1: + interp_atmos_master.sh[47]err=0 - 1: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 1: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 1: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_2_0p25 - 1: + product_functions.sh[5]local filename=pgb2file_f026_2_0p25 - 1: + product_functions.sh[6]wgrib2 pgb2file_f026_2_0p25 -not_if :RH: -grib pgb2file_f026_2_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_2_0p25.new - 1: 1:0:d=2021032312:ABSV:0.02 mb:26 hour fcst: - 1: 2:365052:d=2021032312:O3MR:0.02 mb:26 hour fcst: - 1: 3:693442:d=2021032312:HGT:0.04 mb:26 hour fcst: - 1: 4:1344753:d=2021032312:TMP:0.04 mb:26 hour fcst: - 1: 5:1736295:d=2021032312:RH:0.04 mb:26 hour fcst: - 1: 6:1878546:d=2021032312:SPFH:0.04 mb:26 hour fcst: - 1: 7:2597409:d=2021032312:VVEL:0.04 mb:26 hour fcst: - 1: 8:3357995:d=2021032312:DZDT:0.04 mb:26 hour fcst: - 1: 9:4026621:d=2021032312:UGRD:0.04 mb:26 hour fcst: - 1: 10:4509039:d=2021032312:VGRD:0.04 mb:26 hour fcst: - 1: 11:4985556:d=2021032312:ABSV:0.04 mb:26 hour fcst: - 1: 12:5395886:d=2021032312:O3MR:0.04 mb:26 hour fcst: - 1: 13:6020534:d=2021032312:HGT:0.07 mb:26 hour fcst: - 1: 14:6662260:d=2021032312:TMP:0.07 mb:26 hour fcst: - 1: 15:7043518:d=2021032312:RH:0.07 mb:26 hour fcst: - 1: 16:7268452:d=2021032312:SPFH:0.07 mb:26 hour fcst: - 1: 17:7864260:d=2021032312:VVEL:0.07 mb:26 hour fcst: - 1: 18:8675695:d=2021032312:DZDT:0.07 mb:26 hour fcst: - 1: 19:9369262:d=2021032312:UGRD:0.07 mb:26 hour fcst: - 1: 20:9872727:d=2021032312:VGRD:0.07 mb:26 hour fcst: - 1: 21:10366520:d=2021032312:ABSV:0.07 mb:26 hour fcst: - 1: 22:10796934:d=2021032312:O3MR:0.07 mb:26 hour fcst: - 1: 23:11262469:d=2021032312:HGT:0.1 mb:26 hour fcst: - 1: 24:11799952:d=2021032312:TMP:0.1 mb:26 hour fcst: - 1: 25:12167504:d=2021032312:RH:0.1 mb:26 hour fcst: - 1: 26:12370512:d=2021032312:SPFH:0.1 mb:26 hour fcst: - 1: 27:12882970:d=2021032312:VVEL:0.1 mb:26 hour fcst: - 1: 28:13709205:d=2021032312:DZDT:0.1 mb:26 hour fcst: - 1: 29:14417656:d=2021032312:UGRD:0.1 mb:26 hour fcst: - 1: 30:14903051:d=2021032312:VGRD:0.1 mb:26 hour fcst: - 1: 31:15385253:d=2021032312:ABSV:0.1 mb:26 hour fcst: - 1: + product_functions.sh[10]rc=0 - 1: + product_functions.sh[11](( rc == 0 )) - 1: + product_functions.sh[11]mv pgb2file_f026_2_0p25.new pgb2file_f026_2_0p25 - 1: + product_functions.sh[12]return 0 - 1: + interp_atmos_master.sh[56]export err=0 - 1: + interp_atmos_master.sh[56]err=0 - 1: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 1: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_2_0p25 -match 'LAND|ICEC' - 1: ++ interp_atmos_master.sh[62]wc -l - 1: + interp_atmos_master.sh[62]var_count=0 - 1: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 1: + interp_atmos_master.sh[73]exit 0 - 7: + bash[8]'[' -z '' ']' - 7: + bash[9]case "$-" in - 7: + bash[12]__lmod_vx=x - 7: + bash[16]'[' -n x ']' - 7: + bash[16]set +x - 7: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 7: Shell debugging restarted - 7: + bash[224]unset __lmod_vx - 7: + interp_atmos_master.sh[7]input_file=tmpfile_f026_8 - 7: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_8 - 7: + interp_atmos_master.sh[9]grid_string=0p25 - 7: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 7: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 7: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 7: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 7: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 7: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 7: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 7: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 7: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 7: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 7: + interp_atmos_master.sh[31]IFS=: - 7: + interp_atmos_master.sh[31]read -ra grids - 7: + interp_atmos_master.sh[33]output_grids= - 7: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 7: + interp_atmos_master.sh[35]gridopt=grid0p25 - 7: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_8_0p25' - 7: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_8 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_8_0p25 - 7: 1:0:d=2021032312:ABSV:70 mb:26 hour fcst: - 7: 2:17698:d=2021032312:O3MR:70 mb:26 hour fcst: - 7: 3:43566:d=2021032312:HGT:100 mb:26 hour fcst: - 7: 4:63172:d=2021032312:TMP:100 mb:26 hour fcst: - 7: 5:80017:d=2021032312:RH:100 mb:26 hour fcst: - 7: 6:90135:d=2021032312:TCDC:100 mb:26 hour fcst: - 7: 7:90314:d=2021032312:SPFH:100 mb:26 hour fcst: - 7: 8:115881:d=2021032312:VVEL:100 mb:26 hour fcst: - 7: 9:140762:d=2021032312:DZDT:100 mb:26 hour fcst: - 7: 10:171017:d=2021032312:UGRD:100 mb:26 hour fcst: - 7: 11:190379:d=2021032312:VGRD:100 mb:26 hour fcst: - 7: 12:209799:d=2021032312:ABSV:100 mb:26 hour fcst: - 7: 13:227977:d=2021032312:CLMR:100 mb:26 hour fcst: - 7: 14:228156:d=2021032312:ICMR:100 mb:26 hour fcst: - 7: 15:229067:d=2021032312:RWMR:100 mb:26 hour fcst: - 7: 16:229246:d=2021032312:SNMR:100 mb:26 hour fcst: - 7: 17:231499:d=2021032312:GRLE:100 mb:26 hour fcst: - 7: 18:231678:d=2021032312:O3MR:100 mb:26 hour fcst: - 7: 19:255701:d=2021032312:HGT:150 mb:26 hour fcst: - 7: 20:276092:d=2021032312:TMP:150 mb:26 hour fcst: - 7: 21:293062:d=2021032312:RH:150 mb:26 hour fcst: - 7: 22:304457:d=2021032312:TCDC:150 mb:26 hour fcst: - 7: 23:308495:d=2021032312:SPFH:150 mb:26 hour fcst: - 7: 24:332748:d=2021032312:VVEL:150 mb:26 hour fcst: - 7: 25:360105:d=2021032312:DZDT:150 mb:26 hour fcst: - 7: 26:385699:d=2021032312:UGRD:150 mb:26 hour fcst: - 7: 27:405701:d=2021032312:VGRD:150 mb:26 hour fcst: - 7: 28:425659:d=2021032312:ABSV:150 mb:26 hour fcst: - 7: 29:444607:d=2021032312:CLMR:150 mb:26 hour fcst: - 7: 30:444786:d=2021032312:ICMR:150 mb:26 hour fcst: - 7: 31:450495:d=2021032312:RWMR:150 mb:26 hour fcst: - 7: + interp_atmos_master.sh[47]export err=0 - 7: + interp_atmos_master.sh[47]err=0 - 7: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 7: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 7: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_8_0p25 - 7: + product_functions.sh[5]local filename=pgb2file_f026_8_0p25 - 7: + product_functions.sh[6]wgrib2 pgb2file_f026_8_0p25 -not_if :RH: -grib pgb2file_f026_8_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_8_0p25.new - 7: 1:0:d=2021032312:ABSV:70 mb:26 hour fcst: - 7: 2:398312:d=2021032312:O3MR:70 mb:26 hour fcst: - 7: 3:1108036:d=2021032312:HGT:100 mb:26 hour fcst: - 7: 4:1576360:d=2021032312:TMP:100 mb:26 hour fcst: - 7: 5:1953046:d=2021032312:RH:100 mb:26 hour fcst: - 7: 6:2184706:d=2021032312:TCDC:100 mb:26 hour fcst: - 7: 7:2184885:d=2021032312:SPFH:100 mb:26 hour fcst: - 7: 8:2842222:d=2021032312:VVEL:100 mb:26 hour fcst: - 7: 9:3512116:d=2021032312:DZDT:100 mb:26 hour fcst: - 7: 10:4409388:d=2021032312:UGRD:100 mb:26 hour fcst: - 7: 11:4876291:d=2021032312:VGRD:100 mb:26 hour fcst: - 7: 12:5339897:d=2021032312:ABSV:100 mb:26 hour fcst: - 7: 13:5746889:d=2021032312:CLMR:100 mb:26 hour fcst: - 7: 14:5747068:d=2021032312:ICMR:100 mb:26 hour fcst: - 7: 15:5759395:d=2021032312:RWMR:100 mb:26 hour fcst: - 7: 16:5759574:d=2021032312:SNMR:100 mb:26 hour fcst: - 7: 17:5807766:d=2021032312:GRLE:100 mb:26 hour fcst: - 7: 18:5807945:d=2021032312:O3MR:100 mb:26 hour fcst: - 7: 19:6422739:d=2021032312:HGT:150 mb:26 hour fcst: - 7: 20:6918060:d=2021032312:TMP:150 mb:26 hour fcst: - 7: 21:7294680:d=2021032312:RH:150 mb:26 hour fcst: - 7: 22:7562695:d=2021032312:TCDC:150 mb:26 hour fcst: - 7: 23:7646182:d=2021032312:SPFH:150 mb:26 hour fcst: - 7: 24:8276291:d=2021032312:VVEL:150 mb:26 hour fcst: - 7: 25:9030588:d=2021032312:DZDT:150 mb:26 hour fcst: - 7: 26:9727941:d=2021032312:UGRD:150 mb:26 hour fcst: - 7: 27:10219315:d=2021032312:VGRD:150 mb:26 hour fcst: - 7: 28:10720229:d=2021032312:ABSV:150 mb:26 hour fcst: - 7: 29:11150431:d=2021032312:CLMR:150 mb:26 hour fcst: - 7: 30:11150610:d=2021032312:ICMR:150 mb:26 hour fcst: - 7: 31:11290830:d=2021032312:RWMR:150 mb:26 hour fcst: - 7: + product_functions.sh[10]rc=0 - 7: + product_functions.sh[11](( rc == 0 )) - 7: + product_functions.sh[11]mv pgb2file_f026_8_0p25.new pgb2file_f026_8_0p25 - 7: + product_functions.sh[12]return 0 - 7: + interp_atmos_master.sh[56]export err=0 - 7: + interp_atmos_master.sh[56]err=0 - 7: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 7: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_8_0p25 -match 'LAND|ICEC' - 7: ++ interp_atmos_master.sh[62]wc -l - 7: + interp_atmos_master.sh[62]var_count=0 - 7: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 7: + interp_atmos_master.sh[73]exit 0 -21: + bash[8]'[' -z '' ']' -21: + bash[9]case "$-" in -21: + bash[12]__lmod_vx=x -21: + bash[16]'[' -n x ']' -21: + bash[16]set +x -21: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -21: Shell debugging restarted -21: + bash[224]unset __lmod_vx -21: + interp_atmos_master.sh[7]input_file=tmpfile_f026_22 -21: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_22 -21: + interp_atmos_master.sh[9]grid_string=0p25 -21: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -21: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -21: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -21: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -21: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -21: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -21: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -21: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -21: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -21: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -21: + interp_atmos_master.sh[31]IFS=: -21: + interp_atmos_master.sh[31]read -ra grids -21: + interp_atmos_master.sh[33]output_grids= -21: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -21: + interp_atmos_master.sh[35]gridopt=grid0p25 -21: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_22_0p25' -21: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_22 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_22_0p25 -21: 1:0:d=2021032312:DLWRF:surface:24-26 hour ave fcst: -21: 2:21808:d=2021032312:USWRF:surface:24-26 hour ave fcst: -21: 3:39976:d=2021032312:ULWRF:surface:24-26 hour ave fcst: -21: 4:58859:d=2021032312:USWRF:top of atmosphere:24-26 hour ave fcst: -21: 5:77865:d=2021032312:ULWRF:top of atmosphere:24-26 hour ave fcst: -21: 6:98957:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Total Aerosol:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 7:126449:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Dust Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 8:144585:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Sea Salt Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 9:172082:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Sulphate Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 10:198005:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Particulate Organic Matter Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 11:223635:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Black Carbon Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 12:245953:d=2021032312:HLCY:3000-0 m above ground:26 hour fcst: -21: 13:261812:d=2021032312:HLCY:1000-0 m above ground:26 hour fcst: -21: 14:284090:d=2021032312:USTM:6000-0 m above ground:26 hour fcst: -21: 15:307132:d=2021032312:VSTM:6000-0 m above ground:26 hour fcst: -21: 16:329936:d=2021032312:PRES:tropopause:26 hour fcst: -21: 17:359668:d=2021032312:ICAHT:tropopause:26 hour fcst: -21: 18:390468:d=2021032312:HGT:tropopause:26 hour fcst: -21: 19:418599:d=2021032312:TMP:tropopause:26 hour fcst: -21: 20:440264:d=2021032312:UGRD:tropopause:26 hour fcst: -21: 21:456214:d=2021032312:VGRD:tropopause:26 hour fcst: -21: 22:472122:d=2021032312:VWSH:tropopause:26 hour fcst: -21: 23:490845:d=2021032312:PRES:max wind:26 hour fcst: -21: 24:525263:d=2021032312:ICAHT:max wind:26 hour fcst: -21: 25:560308:d=2021032312:HGT:max wind:26 hour fcst: -21: 26:595506:d=2021032312:UGRD:max wind:26 hour fcst: -21: 27:611955:d=2021032312:VGRD:max wind:26 hour fcst: -21: 28:629379:d=2021032312:TMP:max wind:26 hour fcst: -21: 29:656035:d=2021032312:UGRD:20 m above ground:26 hour fcst: -21: 30:676555:d=2021032312:VGRD:20 m above ground:26 hour fcst: -21: 31:697493:d=2021032312:UGRD:30 m above ground:26 hour fcst: -21: 32:718440:d=2021032312:VGRD:30 m above ground:26 hour fcst: -21: + interp_atmos_master.sh[47]export err=0 -21: + interp_atmos_master.sh[47]err=0 -21: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -21: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -21: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_22_0p25 -21: + product_functions.sh[5]local filename=pgb2file_f026_22_0p25 -21: + product_functions.sh[6]wgrib2 pgb2file_f026_22_0p25 -not_if :RH: -grib pgb2file_f026_22_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_22_0p25.new -21: 1:0:d=2021032312:DLWRF:surface:24-26 hour ave fcst: -21: 2:526536:d=2021032312:USWRF:surface:24-26 hour ave fcst: -21: 3:980762:d=2021032312:ULWRF:surface:24-26 hour ave fcst: -21: 4:1409656:d=2021032312:USWRF:top of atmosphere:24-26 hour ave fcst: -21: 5:1891200:d=2021032312:ULWRF:top of atmosphere:24-26 hour ave fcst: -21: 6:2402470:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Total Aerosol:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 7:3146316:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Dust Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 8:3592902:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Sea Salt Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 9:4328815:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Sulphate Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 10:5006202:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Particulate Organic Matter Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 11:5664808:d=2021032312:AOTK:entire atmosphere:26 hour fcst:aerosol=Black Carbon Dry:aerosol_size <2e-05:aerosol_wavelength >=5.45e-07,<=5.55e-07 -21: 12:6218199:d=2021032312:HLCY:3000-0 m above ground:26 hour fcst: -21: 13:6556752:d=2021032312:HLCY:1000-0 m above ground:26 hour fcst: -21: 14:7095557:d=2021032312:USTM:6000-0 m above ground:26 hour fcst: -21: 15:7667606:d=2021032312:VSTM:6000-0 m above ground:26 hour fcst: -21: 16:8233419:d=2021032312:PRES:tropopause:26 hour fcst: -21: 17:9033050:d=2021032312:ICAHT:tropopause:26 hour fcst: -21: 18:9883098:d=2021032312:HGT:tropopause:26 hour fcst: -21: 19:10631037:d=2021032312:TMP:tropopause:26 hour fcst: -21: 20:11147709:d=2021032312:UGRD:tropopause:26 hour fcst: -21: 21:11488839:d=2021032312:VGRD:tropopause:26 hour fcst: -21: 22:11829057:d=2021032312:VWSH:tropopause:26 hour fcst: -21: 23:12243092:d=2021032312:PRES:max wind:26 hour fcst: -21: 24:13131966:d=2021032312:ICAHT:max wind:26 hour fcst: -21: 25:14033104:d=2021032312:HGT:max wind:26 hour fcst: -21: 26:14937383:d=2021032312:UGRD:max wind:26 hour fcst: -21: 27:15292486:d=2021032312:VGRD:max wind:26 hour fcst: -21: 28:15668217:d=2021032312:TMP:max wind:26 hour fcst: -21: 29:16331233:d=2021032312:UGRD:20 m above ground:26 hour fcst: -21: 30:16829373:d=2021032312:VGRD:20 m above ground:26 hour fcst: -21: 31:17343233:d=2021032312:UGRD:30 m above ground:26 hour fcst: -21: 32:17849091:d=2021032312:VGRD:30 m above ground:26 hour fcst: -21: + product_functions.sh[10]rc=0 -21: + product_functions.sh[11](( rc == 0 )) -21: + product_functions.sh[11]mv pgb2file_f026_22_0p25.new pgb2file_f026_22_0p25 -21: + product_functions.sh[12]return 0 -21: + interp_atmos_master.sh[56]export err=0 -21: + interp_atmos_master.sh[56]err=0 -21: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -21: ++ interp_atmos_master.sh[62]wc -l -21: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_22_0p25 -match 'LAND|ICEC' -21: + interp_atmos_master.sh[62]var_count=0 -21: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -21: + interp_atmos_master.sh[73]exit 0 -12: + bash[8]'[' -z '' ']' -12: + bash[9]case "$-" in -12: + bash[12]__lmod_vx=x -12: + bash[16]'[' -n x ']' -12: + bash[16]set +x -12: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -12: Shell debugging restarted -12: + bash[224]unset __lmod_vx -12: + interp_atmos_master.sh[7]input_file=tmpfile_f026_13 -12: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_13 -12: + interp_atmos_master.sh[9]grid_string=0p25 -12: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -12: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -12: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -12: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -12: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -12: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -12: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -12: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -12: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -12: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -12: + interp_atmos_master.sh[31]IFS=: -12: + interp_atmos_master.sh[31]read -ra grids -12: + interp_atmos_master.sh[33]output_grids= -12: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -12: + interp_atmos_master.sh[35]gridopt=grid0p25 -12: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_13_0p25' -12: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_13 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_13_0p25 -12: 1:0:d=2021032312:ABSV:550 mb:26 hour fcst: -12: 2:20950:d=2021032312:CLMR:550 mb:26 hour fcst: -12: 3:24233:d=2021032312:ICMR:550 mb:26 hour fcst: -12: 4:40928:d=2021032312:RWMR:550 mb:26 hour fcst: -12: 5:42307:d=2021032312:SNMR:550 mb:26 hour fcst: -12: 6:56815:d=2021032312:GRLE:550 mb:26 hour fcst: -12: 7:58738:d=2021032312:O3MR:550 mb:26 hour fcst: -12: 8:81565:d=2021032312:HGT:600 mb:26 hour fcst: -12: 9:105717:d=2021032312:TMP:600 mb:26 hour fcst: -12: 10:124767:d=2021032312:RH:600 mb:26 hour fcst: -12: 11:144785:d=2021032312:TCDC:600 mb:26 hour fcst: -12: 12:156863:d=2021032312:SPFH:600 mb:26 hour fcst: -12: 13:187127:d=2021032312:VVEL:600 mb:26 hour fcst: -12: 14:212532:d=2021032312:DZDT:600 mb:26 hour fcst: -12: 15:239947:d=2021032312:UGRD:600 mb:26 hour fcst: -12: 16:260832:d=2021032312:VGRD:600 mb:26 hour fcst: -12: 17:282449:d=2021032312:ABSV:600 mb:26 hour fcst: -12: 18:303255:d=2021032312:CLMR:600 mb:26 hour fcst: -12: 19:308091:d=2021032312:ICMR:600 mb:26 hour fcst: -12: 20:321995:d=2021032312:RWMR:600 mb:26 hour fcst: -12: 21:325716:d=2021032312:SNMR:600 mb:26 hour fcst: -12: 22:341504:d=2021032312:GRLE:600 mb:26 hour fcst: -12: 23:343730:d=2021032312:O3MR:600 mb:26 hour fcst: -12: 24:371872:d=2021032312:HGT:650 mb:26 hour fcst: -12: 25:395972:d=2021032312:TMP:650 mb:26 hour fcst: -12: 26:415290:d=2021032312:RH:650 mb:26 hour fcst: -12: 27:435516:d=2021032312:TCDC:650 mb:26 hour fcst: -12: 28:447254:d=2021032312:SPFH:650 mb:26 hour fcst: -12: 29:475962:d=2021032312:VVEL:650 mb:26 hour fcst: -12: 30:501576:d=2021032312:DZDT:650 mb:26 hour fcst: -12: 31:529001:d=2021032312:UGRD:650 mb:26 hour fcst: -12: 32:550098:d=2021032312:VGRD:650 mb:26 hour fcst: -12: + interp_atmos_master.sh[47]export err=0 -12: + interp_atmos_master.sh[47]err=0 -12: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -12: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -12: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_13_0p25 -12: + product_functions.sh[5]local filename=pgb2file_f026_13_0p25 -12: + product_functions.sh[6]wgrib2 pgb2file_f026_13_0p25 -not_if :RH: -grib pgb2file_f026_13_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_13_0p25.new -12: 1:0:d=2021032312:ABSV:550 mb:26 hour fcst: -12: 2:495266:d=2021032312:CLMR:550 mb:26 hour fcst: -12: 3:564842:d=2021032312:ICMR:550 mb:26 hour fcst: -12: 4:941451:d=2021032312:RWMR:550 mb:26 hour fcst: -12: 5:964005:d=2021032312:SNMR:550 mb:26 hour fcst: -12: 6:1279814:d=2021032312:GRLE:550 mb:26 hour fcst: -12: 7:1314072:d=2021032312:O3MR:550 mb:26 hour fcst: -12: 8:1886680:d=2021032312:HGT:600 mb:26 hour fcst: -12: 9:2518940:d=2021032312:TMP:600 mb:26 hour fcst: -12: 10:2945532:d=2021032312:RH:600 mb:26 hour fcst: -12: 11:3409664:d=2021032312:TCDC:600 mb:26 hour fcst: -12: 12:3639029:d=2021032312:SPFH:600 mb:26 hour fcst: -12: 13:4464924:d=2021032312:VVEL:600 mb:26 hour fcst: -12: 14:5137696:d=2021032312:DZDT:600 mb:26 hour fcst: -12: 15:5896710:d=2021032312:UGRD:600 mb:26 hour fcst: -12: 16:6403792:d=2021032312:VGRD:600 mb:26 hour fcst: -12: 17:6940482:d=2021032312:ABSV:600 mb:26 hour fcst: -12: 18:7424050:d=2021032312:CLMR:600 mb:26 hour fcst: -12: 19:7524606:d=2021032312:ICMR:600 mb:26 hour fcst: -12: 20:7829448:d=2021032312:RWMR:600 mb:26 hour fcst: -12: 21:7905919:d=2021032312:SNMR:600 mb:26 hour fcst: -12: 22:8248874:d=2021032312:GRLE:600 mb:26 hour fcst: -12: 23:8290109:d=2021032312:O3MR:600 mb:26 hour fcst: -12: 24:9054083:d=2021032312:HGT:650 mb:26 hour fcst: -12: 25:9684627:d=2021032312:TMP:650 mb:26 hour fcst: -12: 26:10120932:d=2021032312:RH:650 mb:26 hour fcst: -12: 27:10591934:d=2021032312:TCDC:650 mb:26 hour fcst: -12: 28:10817779:d=2021032312:SPFH:650 mb:26 hour fcst: -12: 29:11574608:d=2021032312:VVEL:650 mb:26 hour fcst: -12: 30:12253139:d=2021032312:DZDT:650 mb:26 hour fcst: -12: 31:13009016:d=2021032312:UGRD:650 mb:26 hour fcst: -12: 32:13513627:d=2021032312:VGRD:650 mb:26 hour fcst: -12: + product_functions.sh[10]rc=0 -12: + product_functions.sh[11](( rc == 0 )) -12: + product_functions.sh[11]mv pgb2file_f026_13_0p25.new pgb2file_f026_13_0p25 -12: + product_functions.sh[12]return 0 -12: + interp_atmos_master.sh[56]export err=0 -12: + interp_atmos_master.sh[56]err=0 -12: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -12: ++ interp_atmos_master.sh[62]wc -l -12: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_13_0p25 -match 'LAND|ICEC' -12: + interp_atmos_master.sh[62]var_count=0 -12: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -12: + interp_atmos_master.sh[73]exit 0 -19: + bash[8]'[' -z '' ']' -19: + bash[9]case "$-" in -19: + bash[12]__lmod_vx=x -19: + bash[16]'[' -n x ']' -19: + bash[16]set +x -19: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -19: Shell debugging restarted -19: + bash[224]unset __lmod_vx -19: + interp_atmos_master.sh[7]input_file=tmpfile_f026_20 -19: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_20 -19: + interp_atmos_master.sh[9]grid_string=0p25 -19: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -19: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -19: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -19: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -19: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -19: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -19: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -19: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -19: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -19: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -19: + interp_atmos_master.sh[31]IFS=: -19: + interp_atmos_master.sh[31]read -ra grids -19: + interp_atmos_master.sh[33]output_grids= -19: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -19: + interp_atmos_master.sh[35]gridopt=grid0p25 -19: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_20_0p25' -19: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_20 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_20_0p25 -19: 1:0:d=2021032312:CPRAT:surface:24-26 hour ave fcst: -19: 2:23643:d=2021032312:PRATE:surface:24-26 hour ave fcst: -19: 3:48774:d=2021032312:APCP:surface:24-26 hour acc fcst: -19: 4:56395:d=2021032312:APCP:surface:0-26 hour acc fcst: -19: 5:69742:d=2021032312:ACPCP:surface:24-26 hour acc fcst: -19: 6:75856:d=2021032312:ACPCP:surface:0-26 hour acc fcst: -19: 7:87002:d=2021032312:FROZR:surface:0-26 hour acc fcst: -19: 8:93389:d=2021032312:FRZR:surface:0-26 hour acc fcst: -19: 9:95345:d=2021032312:TSNOWP:surface:0-26 hour acc fcst: -19: 10:103060:d=2021032312:WATR:surface:24-26 hour acc fcst: -19: 11:115804:d=2021032312:FROZR:surface:24-26 hour acc fcst: -19: 12:118499:d=2021032312:FRZR:surface:24-26 hour acc fcst: -19: 13:119223:d=2021032312:TSNOWP:surface:24-26 hour acc fcst: -19: 14:125785:d=2021032312:CSNOW:surface:26 hour fcst: -19: 15:127614:d=2021032312:CICEP:surface:26 hour fcst: -19: 16:127793:d=2021032312:CFRZR:surface:26 hour fcst: -19: 17:128019:d=2021032312:CRAIN:surface:26 hour fcst: -19: 18:131533:d=2021032312:LHTFL:surface:24-26 hour ave fcst: -19: 19:153332:d=2021032312:SHTFL:surface:24-26 hour ave fcst: -19: 20:174127:d=2021032312:GFLUX:surface:24-26 hour ave fcst: -19: 21:186192:d=2021032312:UFLX:surface:24-26 hour ave fcst: -19: 22:202605:d=2021032312:VFLX:surface:24-26 hour ave fcst: -19: 23:219105:d=2021032312:SFCR:surface:26 hour fcst: -19: 24:237520:d=2021032312:FRICV:surface:26 hour fcst: -19: 25:259259:d=2021032312:U-GWD:surface:24-26 hour ave fcst: -19: 26:274174:d=2021032312:V-GWD:surface:24-26 hour ave fcst: -19: 27:289139:d=2021032312:VEG:surface:26 hour fcst: -19: 28:297173:d=2021032312:SOTYP:surface:26 hour fcst: -19: 29:308222:d=2021032312:WILT:surface:26 hour fcst: -19: 30:318132:d=2021032312:FLDCP:surface:26 hour fcst: -19: 31:328338:d=2021032312:SUNSD:surface:26 hour fcst: -19: + interp_atmos_master.sh[47]export err=0 -19: + interp_atmos_master.sh[47]err=0 -19: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -19: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -19: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_20_0p25 -19: + product_functions.sh[5]local filename=pgb2file_f026_20_0p25 -19: + product_functions.sh[6]wgrib2 pgb2file_f026_20_0p25 -not_if :RH: -grib pgb2file_f026_20_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_20_0p25.new -19: 1:0:d=2021032312:CPRAT:surface:24-26 hour ave fcst: -19: 2:621565:d=2021032312:PRATE:surface:24-26 hour ave fcst: -19: 3:1261999:d=2021032312:APCP:surface:24-26 hour acc fcst: -19: 4:1401744:d=2021032312:APCP:surface:0-26 hour acc fcst: -19: 5:1665152:d=2021032312:ACPCP:surface:24-26 hour acc fcst: -19: 6:1773791:d=2021032312:ACPCP:surface:0-26 hour acc fcst: -19: 7:1993230:d=2021032312:FROZR:surface:0-26 hour acc fcst: -19: 8:2119430:d=2021032312:FRZR:surface:0-26 hour acc fcst: -19: 9:2147082:d=2021032312:TSNOWP:surface:0-26 hour acc fcst: -19: 10:2292002:d=2021032312:WATR:surface:24-26 hour acc fcst: -19: 11:2666355:d=2021032312:FROZR:surface:24-26 hour acc fcst: -19: 12:2713137:d=2021032312:FRZR:surface:24-26 hour acc fcst: -19: 13:2720021:d=2021032312:TSNOWP:surface:24-26 hour acc fcst: -19: 14:2840263:d=2021032312:CSNOW:surface:26 hour fcst: -19: 15:2861564:d=2021032312:CICEP:surface:26 hour fcst: -19: 16:2861743:d=2021032312:CFRZR:surface:26 hour fcst: -19: 17:2862086:d=2021032312:CRAIN:surface:26 hour fcst: -19: 18:2927844:d=2021032312:LHTFL:surface:24-26 hour ave fcst: -19: 19:3478451:d=2021032312:SHTFL:surface:24-26 hour ave fcst: -19: 20:3976913:d=2021032312:GFLUX:surface:24-26 hour ave fcst: -19: 21:4347250:d=2021032312:UFLX:surface:24-26 hour ave fcst: -19: 22:4709432:d=2021032312:VFLX:surface:24-26 hour ave fcst: -19: 23:5074710:d=2021032312:SFCR:surface:26 hour fcst: -19: 24:5466115:d=2021032312:FRICV:surface:26 hour fcst: -19: 25:5967828:d=2021032312:U-GWD:surface:24-26 hour ave fcst: -19: 26:6302769:d=2021032312:V-GWD:surface:24-26 hour ave fcst: -19: 27:6636826:d=2021032312:VEG:surface:26 hour fcst: -19: 28:6910009:d=2021032312:SOTYP:surface:26 hour fcst: -19: 29:7147876:d=2021032312:WILT:surface:26 hour fcst: -19: 30:7460357:d=2021032312:FLDCP:surface:26 hour fcst: -19: 31:7782118:d=2021032312:SUNSD:surface:26 hour fcst: -19: + product_functions.sh[10]rc=0 -19: + product_functions.sh[11](( rc == 0 )) -19: + product_functions.sh[11]mv pgb2file_f026_20_0p25.new pgb2file_f026_20_0p25 -19: + product_functions.sh[12]return 0 -19: + interp_atmos_master.sh[56]export err=0 -19: + interp_atmos_master.sh[56]err=0 -19: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -19: ++ interp_atmos_master.sh[62]wc -l -19: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_20_0p25 -match 'LAND|ICEC' -19: + interp_atmos_master.sh[62]var_count=0 -19: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -19: + interp_atmos_master.sh[73]exit 0 - 8: + bash[8]'[' -z '' ']' - 8: + bash[9]case "$-" in - 8: + bash[12]__lmod_vx=x - 8: + bash[16]'[' -n x ']' - 8: + bash[16]set +x - 8: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 8: Shell debugging restarted - 8: + bash[224]unset __lmod_vx - 8: + interp_atmos_master.sh[7]input_file=tmpfile_f026_9 - 8: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_9 - 8: + interp_atmos_master.sh[9]grid_string=0p25 - 8: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 8: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 8: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 8: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 8: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 8: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 8: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 8: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 8: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 8: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 8: + interp_atmos_master.sh[31]IFS=: - 8: + interp_atmos_master.sh[31]read -ra grids - 8: + interp_atmos_master.sh[33]output_grids= - 8: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 8: + interp_atmos_master.sh[35]gridopt=grid0p25 - 8: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_9_0p25' - 8: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_9 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_9_0p25 - 8: 1:0:d=2021032312:SNMR:150 mb:26 hour fcst: - 8: 2:3558:d=2021032312:GRLE:150 mb:26 hour fcst: - 8: 3:4077:d=2021032312:O3MR:150 mb:26 hour fcst: - 8: 4:25951:d=2021032312:HGT:200 mb:26 hour fcst: - 8: 5:47014:d=2021032312:TMP:200 mb:26 hour fcst: - 8: 6:65159:d=2021032312:RH:200 mb:26 hour fcst: - 8: 7:80189:d=2021032312:TCDC:200 mb:26 hour fcst: - 8: 8:86908:d=2021032312:SPFH:200 mb:26 hour fcst: - 8: 9:112663:d=2021032312:VVEL:200 mb:26 hour fcst: - 8: 10:139462:d=2021032312:DZDT:200 mb:26 hour fcst: - 8: 11:165891:d=2021032312:UGRD:200 mb:26 hour fcst: - 8: 12:179279:d=2021032312:VGRD:200 mb:26 hour fcst: - 8: 13:193010:d=2021032312:ABSV:200 mb:26 hour fcst: - 8: 14:213005:d=2021032312:CLMR:200 mb:26 hour fcst: - 8: 15:213184:d=2021032312:ICMR:200 mb:26 hour fcst: - 8: 16:221833:d=2021032312:RWMR:200 mb:26 hour fcst: - 8: 17:222012:d=2021032312:SNMR:200 mb:26 hour fcst: - 8: 18:227464:d=2021032312:GRLE:200 mb:26 hour fcst: - 8: 19:227793:d=2021032312:O3MR:200 mb:26 hour fcst: - 8: 20:248430:d=2021032312:HGT:250 mb:26 hour fcst: - 8: 21:270165:d=2021032312:TMP:250 mb:26 hour fcst: - 8: 22:288999:d=2021032312:RH:250 mb:26 hour fcst: - 8: 23:306821:d=2021032312:TCDC:250 mb:26 hour fcst: - 8: 24:317533:d=2021032312:SPFH:250 mb:26 hour fcst: - 8: 25:345613:d=2021032312:VVEL:250 mb:26 hour fcst: - 8: 26:368347:d=2021032312:DZDT:250 mb:26 hour fcst: - 8: 27:395739:d=2021032312:UGRD:250 mb:26 hour fcst: - 8: 28:409878:d=2021032312:VGRD:250 mb:26 hour fcst: - 8: 29:424734:d=2021032312:ABSV:250 mb:26 hour fcst: - 8: 30:446336:d=2021032312:CLMR:250 mb:26 hour fcst: - 8: 31:446515:d=2021032312:ICMR:250 mb:26 hour fcst: - 8: + interp_atmos_master.sh[47]export err=0 - 8: + interp_atmos_master.sh[47]err=0 - 8: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 8: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 8: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_9_0p25 - 8: + product_functions.sh[5]local filename=pgb2file_f026_9_0p25 - 8: + product_functions.sh[6]wgrib2 pgb2file_f026_9_0p25 -not_if :RH: -grib pgb2file_f026_9_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_9_0p25.new - 8: 1:0:d=2021032312:SNMR:150 mb:26 hour fcst: - 8: 2:58991:d=2021032312:GRLE:150 mb:26 hour fcst: - 8: 3:66133:d=2021032312:O3MR:150 mb:26 hour fcst: - 8: 4:595824:d=2021032312:HGT:200 mb:26 hour fcst: - 8: 5:1122760:d=2021032312:TMP:200 mb:26 hour fcst: - 8: 6:1534196:d=2021032312:RH:200 mb:26 hour fcst: - 8: 7:1882716:d=2021032312:TCDC:200 mb:26 hour fcst: - 8: 8:2022481:d=2021032312:SPFH:200 mb:26 hour fcst: - 8: 9:2697720:d=2021032312:VVEL:200 mb:26 hour fcst: - 8: 10:3421265:d=2021032312:DZDT:200 mb:26 hour fcst: - 8: 11:4156955:d=2021032312:UGRD:200 mb:26 hour fcst: - 8: 12:4463480:d=2021032312:VGRD:200 mb:26 hour fcst: - 8: 13:4771912:d=2021032312:ABSV:200 mb:26 hour fcst: - 8: 14:5243384:d=2021032312:CLMR:200 mb:26 hour fcst: - 8: 15:5243563:d=2021032312:ICMR:200 mb:26 hour fcst: - 8: 16:5426714:d=2021032312:RWMR:200 mb:26 hour fcst: - 8: 17:5426893:d=2021032312:SNMR:200 mb:26 hour fcst: - 8: 18:5530713:d=2021032312:GRLE:200 mb:26 hour fcst: - 8: 19:5533254:d=2021032312:O3MR:200 mb:26 hour fcst: - 8: 20:6021554:d=2021032312:HGT:250 mb:26 hour fcst: - 8: 21:6570383:d=2021032312:TMP:250 mb:26 hour fcst: - 8: 22:6987747:d=2021032312:RH:250 mb:26 hour fcst: - 8: 23:7395947:d=2021032312:TCDC:250 mb:26 hour fcst: - 8: 24:7607385:d=2021032312:SPFH:250 mb:26 hour fcst: - 8: 25:8362092:d=2021032312:VVEL:250 mb:26 hour fcst: - 8: 26:8957350:d=2021032312:DZDT:250 mb:26 hour fcst: - 8: 27:9723019:d=2021032312:UGRD:250 mb:26 hour fcst: - 8: 28:10037109:d=2021032312:VGRD:250 mb:26 hour fcst: - 8: 29:10365898:d=2021032312:ABSV:250 mb:26 hour fcst: - 8: 30:10893147:d=2021032312:CLMR:250 mb:26 hour fcst: - 8: 31:10893326:d=2021032312:ICMR:250 mb:26 hour fcst: - 8: + product_functions.sh[10]rc=0 - 8: + product_functions.sh[11](( rc == 0 )) - 8: + product_functions.sh[11]mv pgb2file_f026_9_0p25.new pgb2file_f026_9_0p25 - 8: + product_functions.sh[12]return 0 - 8: + interp_atmos_master.sh[56]export err=0 - 8: + interp_atmos_master.sh[56]err=0 - 8: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 8: ++ interp_atmos_master.sh[62]wc -l - 8: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_9_0p25 -match 'LAND|ICEC' - 8: + interp_atmos_master.sh[62]var_count=0 - 8: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 8: + interp_atmos_master.sh[73]exit 0 -16: + bash[8]'[' -z '' ']' -16: + bash[9]case "$-" in -16: + bash[12]__lmod_vx=x -16: + bash[16]'[' -n x ']' -16: + bash[16]set +x -16: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -16: Shell debugging restarted -16: + bash[224]unset __lmod_vx -16: + interp_atmos_master.sh[7]input_file=tmpfile_f026_17 -16: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_17 -16: + interp_atmos_master.sh[9]grid_string=0p25 -16: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -16: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -16: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -16: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -16: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -16: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -16: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -16: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -16: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -16: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -16: + interp_atmos_master.sh[31]IFS=: -16: + interp_atmos_master.sh[31]read -ra grids -16: + interp_atmos_master.sh[33]output_grids= -16: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -16: + interp_atmos_master.sh[35]gridopt=grid0p25 -16: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_17_0p25' -16: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_17 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_17_0p25 -16: 1:0:d=2021032312:ABSV:925 mb:26 hour fcst: -16: 2:21268:d=2021032312:CLMR:925 mb:26 hour fcst: -16: 3:29782:d=2021032312:ICMR:925 mb:26 hour fcst: -16: 4:37451:d=2021032312:RWMR:925 mb:26 hour fcst: -16: 5:49697:d=2021032312:SNMR:925 mb:26 hour fcst: -16: 6:59292:d=2021032312:GRLE:925 mb:26 hour fcst: -16: 7:62068:d=2021032312:O3MR:925 mb:26 hour fcst: -16: 8:88020:d=2021032312:HGT:950 mb:26 hour fcst: -16: 9:113372:d=2021032312:TMP:950 mb:26 hour fcst: -16: 10:134416:d=2021032312:RH:950 mb:26 hour fcst: -16: 11:153937:d=2021032312:TCDC:950 mb:26 hour fcst: -16: 12:164015:d=2021032312:SPFH:950 mb:26 hour fcst: -16: 13:191269:d=2021032312:VVEL:950 mb:26 hour fcst: -16: 14:215112:d=2021032312:DZDT:950 mb:26 hour fcst: -16: 15:240000:d=2021032312:UGRD:950 mb:26 hour fcst: -16: 16:261508:d=2021032312:VGRD:950 mb:26 hour fcst: -16: 17:283448:d=2021032312:ABSV:950 mb:26 hour fcst: -16: 18:304539:d=2021032312:CLMR:950 mb:26 hour fcst: -16: 19:311008:d=2021032312:ICMR:950 mb:26 hour fcst: -16: 20:318130:d=2021032312:RWMR:950 mb:26 hour fcst: -16: 21:330423:d=2021032312:SNMR:950 mb:26 hour fcst: -16: 22:339233:d=2021032312:GRLE:950 mb:26 hour fcst: -16: 23:341781:d=2021032312:O3MR:950 mb:26 hour fcst: -16: 24:367716:d=2021032312:HINDEX:surface:26 hour fcst: -16: 25:372816:d=2021032312:HGT:975 mb:26 hour fcst: -16: 26:398359:d=2021032312:TMP:975 mb:26 hour fcst: -16: 27:419242:d=2021032312:RH:975 mb:26 hour fcst: -16: 28:438314:d=2021032312:TCDC:975 mb:26 hour fcst: -16: 29:446507:d=2021032312:SPFH:975 mb:26 hour fcst: -16: 30:473386:d=2021032312:VVEL:975 mb:26 hour fcst: -16: 31:496124:d=2021032312:DZDT:975 mb:26 hour fcst: -16: + interp_atmos_master.sh[47]export err=0 -16: + interp_atmos_master.sh[47]err=0 -16: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -16: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -16: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_17_0p25 -16: + product_functions.sh[5]local filename=pgb2file_f026_17_0p25 -16: + product_functions.sh[6]wgrib2 pgb2file_f026_17_0p25 -not_if :RH: -grib pgb2file_f026_17_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_17_0p25.new -16: 1:0:d=2021032312:ABSV:925 mb:26 hour fcst: -16: 2:505569:d=2021032312:CLMR:925 mb:26 hour fcst: -16: 3:697910:d=2021032312:ICMR:925 mb:26 hour fcst: -16: 4:874313:d=2021032312:RWMR:925 mb:26 hour fcst: -16: 5:1122562:d=2021032312:SNMR:925 mb:26 hour fcst: -16: 6:1344491:d=2021032312:GRLE:925 mb:26 hour fcst: -16: 7:1397509:d=2021032312:O3MR:925 mb:26 hour fcst: -16: 8:2100716:d=2021032312:HGT:950 mb:26 hour fcst: -16: 9:2763188:d=2021032312:TMP:950 mb:26 hour fcst: -16: 10:3257994:d=2021032312:RH:950 mb:26 hour fcst: -16: 11:3700903:d=2021032312:TCDC:950 mb:26 hour fcst: -16: 12:3901673:d=2021032312:SPFH:950 mb:26 hour fcst: -16: 13:4641077:d=2021032312:VVEL:950 mb:26 hour fcst: -16: 14:5252195:d=2021032312:DZDT:950 mb:26 hour fcst: -16: 15:5900786:d=2021032312:UGRD:950 mb:26 hour fcst: -16: 16:6421381:d=2021032312:VGRD:950 mb:26 hour fcst: -16: 17:6960569:d=2021032312:ABSV:950 mb:26 hour fcst: -16: 18:7464546:d=2021032312:CLMR:950 mb:26 hour fcst: -16: 19:7606352:d=2021032312:ICMR:950 mb:26 hour fcst: -16: 20:7773125:d=2021032312:RWMR:950 mb:26 hour fcst: -16: 21:8025475:d=2021032312:SNMR:950 mb:26 hour fcst: -16: 22:8226254:d=2021032312:GRLE:950 mb:26 hour fcst: -16: 23:8274413:d=2021032312:O3MR:950 mb:26 hour fcst: -16: 24:8975882:d=2021032312:HINDEX:surface:26 hour fcst: -16: 25:9155803:d=2021032312:HGT:975 mb:26 hour fcst: -16: 26:9836163:d=2021032312:TMP:975 mb:26 hour fcst: -16: 27:10325649:d=2021032312:RH:975 mb:26 hour fcst: -16: 28:10754035:d=2021032312:TCDC:975 mb:26 hour fcst: -16: 29:10911391:d=2021032312:SPFH:975 mb:26 hour fcst: -16: 30:11643967:d=2021032312:VVEL:975 mb:26 hour fcst: -16: 31:12216184:d=2021032312:DZDT:975 mb:26 hour fcst: -16: + product_functions.sh[10]rc=0 -16: + product_functions.sh[11](( rc == 0 )) -16: + product_functions.sh[11]mv pgb2file_f026_17_0p25.new pgb2file_f026_17_0p25 -16: + product_functions.sh[12]return 0 -16: + interp_atmos_master.sh[56]export err=0 -16: + interp_atmos_master.sh[56]err=0 -16: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -16: ++ interp_atmos_master.sh[62]wc -l -16: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_17_0p25 -match 'LAND|ICEC' -16: + interp_atmos_master.sh[62]var_count=0 -16: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -16: + interp_atmos_master.sh[73]exit 0 -18: + bash[8]'[' -z '' ']' -18: + bash[9]case "$-" in -18: + bash[12]__lmod_vx=x -18: + bash[16]'[' -n x ']' -18: + bash[16]set +x -18: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -18: Shell debugging restarted -18: + bash[224]unset __lmod_vx -18: + interp_atmos_master.sh[7]input_file=tmpfile_f026_19 -18: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_19 -18: + interp_atmos_master.sh[9]grid_string=0p25 -18: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -18: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -18: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -18: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -18: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -18: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -18: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -18: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -18: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -18: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -18: + interp_atmos_master.sh[31]IFS=: -18: + interp_atmos_master.sh[31]read -ra grids -18: + interp_atmos_master.sh[33]output_grids= -18: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -18: + interp_atmos_master.sh[35]gridopt=grid0p25 -18: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_19_0p25' -18: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_19 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_19_0p25 -18: 1:0:d=2021032312:TMP:surface:26 hour fcst: -18: 2:16636:d=2021032312:TSOIL:0-0.1 m below ground:26 hour fcst: -18: 3:28604:d=2021032312:SOILW:0-0.1 m below ground:26 hour fcst: -18: 4:37521:d=2021032312:SOILL:0-0.1 m below ground:26 hour fcst: -18: 5:46321:d=2021032312:TSOIL:0.1-0.4 m below ground:26 hour fcst: -18: 6:61814:d=2021032312:SOILW:0.1-0.4 m below ground:26 hour fcst: -18: 7:70828:d=2021032312:SOILL:0.1-0.4 m below ground:26 hour fcst: -18: 8:79720:d=2021032312:TSOIL:0.4-1 m below ground:26 hour fcst: -18: 9:94169:d=2021032312:SOILW:0.4-1 m below ground:26 hour fcst: -18: 10:103302:d=2021032312:SOILL:0.4-1 m below ground:26 hour fcst: -18: 11:112225:d=2021032312:TSOIL:1-2 m below ground:26 hour fcst: -18: 12:126515:d=2021032312:SOILW:1-2 m below ground:26 hour fcst: -18: 13:135572:d=2021032312:SOILL:1-2 m below ground:26 hour fcst: -18: 14:144649:d=2021032312:CNWAT:surface:26 hour fcst: -18: 15:150171:d=2021032312:WEASD:surface:26 hour fcst: -18: 16:162996:d=2021032312:SNOD:surface:26 hour fcst: -18: 17:177375:d=2021032312:PEVPR:surface:26 hour fcst: -18: 18:193167:d=2021032312:ICETK:surface:26 hour fcst: -18: 19:196575:d=2021032312:TMP:2 m above ground:26 hour fcst: -18: 20:217395:d=2021032312:SPFH:2 m above ground:26 hour fcst: -18: 21:246209:d=2021032312:DPT:2 m above ground:26 hour fcst: -18: 22:267471:d=2021032312:RH:2 m above ground:26 hour fcst: -18: 23:285985:d=2021032312:APTMP:2 m above ground:26 hour fcst: -18: 24:300384:d=2021032312:TMAX:2 m above ground:24-26 hour max fcst: -18: 25:321120:d=2021032312:TMIN:2 m above ground:24-26 hour min fcst: -18: 26:341912:d=2021032312:UGRD:10 m above ground:26 hour fcst: -18: 27:362457:d=2021032312:VGRD:10 m above ground:26 hour fcst: -18: 28:383375:d=2021032312:ICEG:10 m above mean sea level:26 hour fcst: -18: 29:383991:d=2021032312:CPOFP:surface:26 hour fcst: -18: 30:404337:d=2021032312:CPRAT:surface:26 hour fcst: -18: 31:425985:d=2021032312:PRATE:surface:26 hour fcst: -18: + interp_atmos_master.sh[47]export err=0 -18: + interp_atmos_master.sh[47]err=0 -18: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -18: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -18: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_19_0p25 -18: + product_functions.sh[5]local filename=pgb2file_f026_19_0p25 -18: + product_functions.sh[6]wgrib2 pgb2file_f026_19_0p25 -not_if :RH: -grib pgb2file_f026_19_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_19_0p25.new -18: 1:0:d=2021032312:TMP:surface:26 hour fcst: -18: 2:342215:d=2021032312:TSOIL:0-0.1 m below ground:26 hour fcst: -18: 3:676575:d=2021032312:SOILW:0-0.1 m below ground:26 hour fcst: -18: 4:963442:d=2021032312:SOILL:0-0.1 m below ground:26 hour fcst: -18: 5:1252049:d=2021032312:TSOIL:0.1-0.4 m below ground:26 hour fcst: -18: 6:1711429:d=2021032312:SOILW:0.1-0.4 m below ground:26 hour fcst: -18: 7:2002921:d=2021032312:SOILL:0.1-0.4 m below ground:26 hour fcst: -18: 8:2296959:d=2021032312:TSOIL:0.4-1 m below ground:26 hour fcst: -18: 9:2734738:d=2021032312:SOILW:0.4-1 m below ground:26 hour fcst: -18: 10:3031164:d=2021032312:SOILL:0.4-1 m below ground:26 hour fcst: -18: 11:3326778:d=2021032312:TSOIL:1-2 m below ground:26 hour fcst: -18: 12:3756749:d=2021032312:SOILW:1-2 m below ground:26 hour fcst: -18: 13:4047625:d=2021032312:SOILL:1-2 m below ground:26 hour fcst: -18: 14:4343747:d=2021032312:CNWAT:surface:26 hour fcst: -18: 15:4529459:d=2021032312:WEASD:surface:26 hour fcst: -18: 16:4898167:d=2021032312:SNOD:surface:26 hour fcst: -18: 17:5291857:d=2021032312:PEVPR:surface:26 hour fcst: -18: 18:5775165:d=2021032312:ICETK:surface:26 hour fcst: -18: 19:5828525:d=2021032312:TMP:2 m above ground:26 hour fcst: -18: 20:6311649:d=2021032312:SPFH:2 m above ground:26 hour fcst: -18: 21:7096030:d=2021032312:DPT:2 m above ground:26 hour fcst: -18: 22:7600937:d=2021032312:RH:2 m above ground:26 hour fcst: -18: 23:8011155:d=2021032312:APTMP:2 m above ground:26 hour fcst: -18: 24:8323038:d=2021032312:TMAX:2 m above ground:24-26 hour max fcst: -18: 25:8804463:d=2021032312:TMIN:2 m above ground:24-26 hour min fcst: -18: 26:9286726:d=2021032312:UGRD:10 m above ground:26 hour fcst: -18: 27:9786107:d=2021032312:VGRD:10 m above ground:26 hour fcst: -18: 28:10297616:d=2021032312:ICEG:10 m above mean sea level:26 hour fcst: -18: 29:10308009:d=2021032312:CPOFP:surface:26 hour fcst: -18: 30:10732428:d=2021032312:CPRAT:surface:26 hour fcst: -18: 31:11269501:d=2021032312:PRATE:surface:26 hour fcst: -18: + product_functions.sh[10]rc=0 -18: + product_functions.sh[11](( rc == 0 )) -18: + product_functions.sh[11]mv pgb2file_f026_19_0p25.new pgb2file_f026_19_0p25 -18: + product_functions.sh[12]return 0 -18: + interp_atmos_master.sh[56]export err=0 -18: + interp_atmos_master.sh[56]err=0 -18: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -18: ++ interp_atmos_master.sh[62]wc -l -18: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_19_0p25 -match 'LAND|ICEC' -18: + interp_atmos_master.sh[62]var_count=0 -18: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -18: + interp_atmos_master.sh[73]exit 0 -13: + bash[8]'[' -z '' ']' -13: + bash[9]case "$-" in -13: + bash[12]__lmod_vx=x -13: + bash[16]'[' -n x ']' -13: + bash[16]set +x -13: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -13: Shell debugging restarted -13: + bash[224]unset __lmod_vx -13: + interp_atmos_master.sh[7]input_file=tmpfile_f026_14 -13: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_14 -13: + interp_atmos_master.sh[9]grid_string=0p25 -13: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -13: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -13: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -13: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -13: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -13: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -13: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -13: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -13: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -13: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -13: + interp_atmos_master.sh[31]IFS=: -13: + interp_atmos_master.sh[31]read -ra grids -13: + interp_atmos_master.sh[33]output_grids= -13: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -13: + interp_atmos_master.sh[35]gridopt=grid0p25 -13: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_14_0p25' -13: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_14 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_14_0p25 -13: 1:0:d=2021032312:ABSV:650 mb:26 hour fcst: -13: 2:20879:d=2021032312:CLMR:650 mb:26 hour fcst: -13: 3:25908:d=2021032312:ICMR:650 mb:26 hour fcst: -13: 4:38311:d=2021032312:RWMR:650 mb:26 hour fcst: -13: 5:43029:d=2021032312:SNMR:650 mb:26 hour fcst: -13: 6:57907:d=2021032312:GRLE:650 mb:26 hour fcst: -13: 7:60220:d=2021032312:O3MR:650 mb:26 hour fcst: -13: 8:87743:d=2021032312:HGT:700 mb:26 hour fcst: -13: 9:111928:d=2021032312:TMP:700 mb:26 hour fcst: -13: 10:131700:d=2021032312:RH:700 mb:26 hour fcst: -13: 11:151960:d=2021032312:TCDC:700 mb:26 hour fcst: -13: 12:163552:d=2021032312:SPFH:700 mb:26 hour fcst: -13: 13:192121:d=2021032312:VVEL:700 mb:26 hour fcst: -13: 14:217841:d=2021032312:DZDT:700 mb:26 hour fcst: -13: 15:245202:d=2021032312:UGRD:700 mb:26 hour fcst: -13: 16:266087:d=2021032312:VGRD:700 mb:26 hour fcst: -13: 17:287686:d=2021032312:ABSV:700 mb:26 hour fcst: -13: 18:308722:d=2021032312:CLMR:700 mb:26 hour fcst: -13: 19:315588:d=2021032312:ICMR:700 mb:26 hour fcst: -13: 20:326866:d=2021032312:RWMR:700 mb:26 hour fcst: -13: 21:332579:d=2021032312:SNMR:700 mb:26 hour fcst: -13: 22:346619:d=2021032312:GRLE:700 mb:26 hour fcst: -13: 23:348496:d=2021032312:O3MR:700 mb:26 hour fcst: -13: 24:375792:d=2021032312:HGT:750 mb:26 hour fcst: -13: 25:400197:d=2021032312:TMP:750 mb:26 hour fcst: -13: 26:420135:d=2021032312:RH:750 mb:26 hour fcst: -13: 27:440580:d=2021032312:TCDC:750 mb:26 hour fcst: -13: 28:452803:d=2021032312:SPFH:750 mb:26 hour fcst: -13: 29:481776:d=2021032312:VVEL:750 mb:26 hour fcst: -13: 30:507560:d=2021032312:DZDT:750 mb:26 hour fcst: -13: 31:534803:d=2021032312:UGRD:750 mb:26 hour fcst: -13: 32:555663:d=2021032312:VGRD:750 mb:26 hour fcst: -13: + interp_atmos_master.sh[47]export err=0 -13: + interp_atmos_master.sh[47]err=0 -13: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -13: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -13: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_14_0p25 -13: + product_functions.sh[5]local filename=pgb2file_f026_14_0p25 -13: + product_functions.sh[6]wgrib2 pgb2file_f026_14_0p25 -not_if :RH: -grib pgb2file_f026_14_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_14_0p25.new -13: 1:0:d=2021032312:ABSV:650 mb:26 hour fcst: -13: 2:498473:d=2021032312:CLMR:650 mb:26 hour fcst: -13: 3:601148:d=2021032312:ICMR:650 mb:26 hour fcst: -13: 4:877114:d=2021032312:RWMR:650 mb:26 hour fcst: -13: 5:973303:d=2021032312:SNMR:650 mb:26 hour fcst: -13: 6:1294416:d=2021032312:GRLE:650 mb:26 hour fcst: -13: 7:1335111:d=2021032312:O3MR:650 mb:26 hour fcst: -13: 8:2089908:d=2021032312:HGT:700 mb:26 hour fcst: -13: 9:2719088:d=2021032312:TMP:700 mb:26 hour fcst: -13: 10:3169200:d=2021032312:RH:700 mb:26 hour fcst: -13: 11:3642535:d=2021032312:TCDC:700 mb:26 hour fcst: -13: 12:3873730:d=2021032312:SPFH:700 mb:26 hour fcst: -13: 13:4646914:d=2021032312:VVEL:700 mb:26 hour fcst: -13: 14:5330696:d=2021032312:DZDT:700 mb:26 hour fcst: -13: 15:6085043:d=2021032312:UGRD:700 mb:26 hour fcst: -13: 16:6588083:d=2021032312:VGRD:700 mb:26 hour fcst: -13: 17:7119878:d=2021032312:ABSV:700 mb:26 hour fcst: -13: 18:7623834:d=2021032312:CLMR:700 mb:26 hour fcst: -13: 19:7768293:d=2021032312:ICMR:700 mb:26 hour fcst: -13: 20:8026553:d=2021032312:RWMR:700 mb:26 hour fcst: -13: 21:8137385:d=2021032312:SNMR:700 mb:26 hour fcst: -13: 22:8447535:d=2021032312:GRLE:700 mb:26 hour fcst: -13: 23:8478427:d=2021032312:O3MR:700 mb:26 hour fcst: -13: 24:9226050:d=2021032312:HGT:750 mb:26 hour fcst: -13: 25:9856824:d=2021032312:TMP:750 mb:26 hour fcst: -13: 26:10321394:d=2021032312:RH:750 mb:26 hour fcst: -13: 27:10799743:d=2021032312:TCDC:750 mb:26 hour fcst: -13: 28:11044845:d=2021032312:SPFH:750 mb:26 hour fcst: -13: 29:11834402:d=2021032312:VVEL:750 mb:26 hour fcst: -13: 30:12520975:d=2021032312:DZDT:750 mb:26 hour fcst: -13: 31:13271851:d=2021032312:UGRD:750 mb:26 hour fcst: -13: 32:13776865:d=2021032312:VGRD:750 mb:26 hour fcst: -13: + product_functions.sh[10]rc=0 -13: + product_functions.sh[11](( rc == 0 )) -13: + product_functions.sh[11]mv pgb2file_f026_14_0p25.new pgb2file_f026_14_0p25 -13: + product_functions.sh[12]return 0 -13: + interp_atmos_master.sh[56]export err=0 -13: + interp_atmos_master.sh[56]err=0 -13: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -13: ++ interp_atmos_master.sh[62]wc -l -13: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_14_0p25 -match 'LAND|ICEC' -13: + interp_atmos_master.sh[62]var_count=0 -13: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -13: + interp_atmos_master.sh[73]exit 0 -10: + bash[8]'[' -z '' ']' -10: + bash[9]case "$-" in -10: + bash[12]__lmod_vx=x -10: + bash[16]'[' -n x ']' -10: + bash[16]set +x -10: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -10: Shell debugging restarted -10: + bash[224]unset __lmod_vx -10: + interp_atmos_master.sh[7]input_file=tmpfile_f026_11 -10: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_11 -10: + interp_atmos_master.sh[9]grid_string=0p25 -10: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -10: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -10: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -10: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -10: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -10: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -10: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -10: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -10: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -10: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -10: + interp_atmos_master.sh[31]IFS=: -10: + interp_atmos_master.sh[31]read -ra grids -10: + interp_atmos_master.sh[33]output_grids= -10: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -10: + interp_atmos_master.sh[35]gridopt=grid0p25 -10: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_11_0p25' -10: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_11 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_11_0p25 -10: 1:0:d=2021032312:ICMR:350 mb:26 hour fcst: -10: 2:16999:d=2021032312:RWMR:350 mb:26 hour fcst: -10: 3:17178:d=2021032312:SNMR:350 mb:26 hour fcst: -10: 4:29132:d=2021032312:GRLE:350 mb:26 hour fcst: -10: 5:29623:d=2021032312:O3MR:350 mb:26 hour fcst: -10: 6:55110:d=2021032312:HGT:400 mb:26 hour fcst: -10: 7:77044:d=2021032312:TMP:400 mb:26 hour fcst: -10: 8:95423:d=2021032312:RH:400 mb:26 hour fcst: -10: 9:114625:d=2021032312:TCDC:400 mb:26 hour fcst: -10: 10:127920:d=2021032312:SPFH:400 mb:26 hour fcst: -10: 11:155680:d=2021032312:VVEL:400 mb:26 hour fcst: -10: 12:180493:d=2021032312:DZDT:400 mb:26 hour fcst: -10: 13:208593:d=2021032312:UGRD:400 mb:26 hour fcst: -10: 14:230636:d=2021032312:VGRD:400 mb:26 hour fcst: -10: 15:253295:d=2021032312:ABSV:400 mb:26 hour fcst: -10: 16:275267:d=2021032312:CLMR:400 mb:26 hour fcst: -10: 17:275669:d=2021032312:ICMR:400 mb:26 hour fcst: -10: 18:292746:d=2021032312:RWMR:400 mb:26 hour fcst: -10: 19:293124:d=2021032312:SNMR:400 mb:26 hour fcst: -10: 20:307070:d=2021032312:GRLE:400 mb:26 hour fcst: -10: 21:308110:d=2021032312:O3MR:400 mb:26 hour fcst: -10: 22:332655:d=2021032312:HGT:450 mb:26 hour fcst: -10: 23:356574:d=2021032312:TMP:450 mb:26 hour fcst: -10: 24:374972:d=2021032312:RH:450 mb:26 hour fcst: -10: 25:394338:d=2021032312:TCDC:450 mb:26 hour fcst: -10: 26:406934:d=2021032312:SPFH:450 mb:26 hour fcst: -10: 27:436218:d=2021032312:VVEL:450 mb:26 hour fcst: -10: 28:461368:d=2021032312:DZDT:450 mb:26 hour fcst: -10: 29:489386:d=2021032312:UGRD:450 mb:26 hour fcst: -10: 30:511016:d=2021032312:VGRD:450 mb:26 hour fcst: -10: 31:533295:d=2021032312:ABSV:450 mb:26 hour fcst: -10: + interp_atmos_master.sh[47]export err=0 -10: + interp_atmos_master.sh[47]err=0 -10: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -10: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -10: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_11_0p25 -10: + product_functions.sh[5]local filename=pgb2file_f026_11_0p25 -10: + product_functions.sh[6]wgrib2 pgb2file_f026_11_0p25 -not_if :RH: -grib pgb2file_f026_11_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_11_0p25.new -10: 1:0:d=2021032312:ICMR:350 mb:26 hour fcst: -10: 2:375055:d=2021032312:RWMR:350 mb:26 hour fcst: -10: 3:375234:d=2021032312:SNMR:350 mb:26 hour fcst: -10: 4:616971:d=2021032312:GRLE:350 mb:26 hour fcst: -10: 5:623748:d=2021032312:O3MR:350 mb:26 hour fcst: -10: 6:1292683:d=2021032312:HGT:400 mb:26 hour fcst: -10: 7:1828834:d=2021032312:TMP:400 mb:26 hour fcst: -10: 8:2238310:d=2021032312:RH:400 mb:26 hour fcst: -10: 9:2674286:d=2021032312:TCDC:400 mb:26 hour fcst: -10: 10:2915349:d=2021032312:SPFH:400 mb:26 hour fcst: -10: 11:3656285:d=2021032312:VVEL:400 mb:26 hour fcst: -10: 12:4318146:d=2021032312:DZDT:400 mb:26 hour fcst: -10: 13:5110332:d=2021032312:UGRD:400 mb:26 hour fcst: -10: 14:5654924:d=2021032312:VGRD:400 mb:26 hour fcst: -10: 15:6230326:d=2021032312:ABSV:400 mb:26 hour fcst: -10: 16:6766687:d=2021032312:CLMR:400 mb:26 hour fcst: -10: 17:6771111:d=2021032312:ICMR:400 mb:26 hour fcst: -10: 18:7147248:d=2021032312:RWMR:400 mb:26 hour fcst: -10: 19:7150526:d=2021032312:SNMR:400 mb:26 hour fcst: -10: 20:7446644:d=2021032312:GRLE:400 mb:26 hour fcst: -10: 21:7462717:d=2021032312:O3MR:400 mb:26 hour fcst: -10: 22:8094718:d=2021032312:HGT:450 mb:26 hour fcst: -10: 23:8716213:d=2021032312:TMP:450 mb:26 hour fcst: -10: 24:9130773:d=2021032312:RH:450 mb:26 hour fcst: -10: 25:9572732:d=2021032312:TCDC:450 mb:26 hour fcst: -10: 26:9809674:d=2021032312:SPFH:450 mb:26 hour fcst: -10: 27:10608040:d=2021032312:VVEL:450 mb:26 hour fcst: -10: 28:11273359:d=2021032312:DZDT:450 mb:26 hour fcst: -10: 29:12059557:d=2021032312:UGRD:450 mb:26 hour fcst: -10: 30:12593267:d=2021032312:VGRD:450 mb:26 hour fcst: -10: 31:13157420:d=2021032312:ABSV:450 mb:26 hour fcst: -10: + product_functions.sh[10]rc=0 -10: + product_functions.sh[11](( rc == 0 )) -10: + product_functions.sh[11]mv pgb2file_f026_11_0p25.new pgb2file_f026_11_0p25 -10: + product_functions.sh[12]return 0 -10: + interp_atmos_master.sh[56]export err=0 -10: + interp_atmos_master.sh[56]err=0 -10: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -10: ++ interp_atmos_master.sh[62]wc -l -10: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_11_0p25 -match 'LAND|ICEC' -10: + interp_atmos_master.sh[62]var_count=0 -10: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -10: + interp_atmos_master.sh[73]exit 0 -20: + bash[8]'[' -z '' ']' -20: + bash[9]case "$-" in -20: + bash[12]__lmod_vx=x -20: + bash[16]'[' -n x ']' -20: + bash[16]set +x -20: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -20: Shell debugging restarted -20: + bash[224]unset __lmod_vx -20: + interp_atmos_master.sh[7]input_file=tmpfile_f026_21 -20: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_21 -20: + interp_atmos_master.sh[9]grid_string=0p25 -20: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -20: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -20: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -20: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -20: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -20: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -20: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -20: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -20: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -20: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -20: + interp_atmos_master.sh[31]IFS=: -20: + interp_atmos_master.sh[31]read -ra grids -20: + interp_atmos_master.sh[33]output_grids= -20: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -20: + interp_atmos_master.sh[35]gridopt=grid0p25 -20: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_21_0p25' -20: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_21 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_21_0p25 -20: 1:0:d=2021032312:LFTX:surface:26 hour fcst: -20: 2:13756:d=2021032312:CAPE:surface:26 hour fcst: -20: 3:28837:d=2021032312:CIN:surface:26 hour fcst: -20: 4:41781:d=2021032312:PWAT:entire atmosphere (considered as a single layer):26 hour fcst: -20: 5:68695:d=2021032312:CWAT:entire atmosphere (considered as a single layer):26 hour fcst: -20: 6:95169:d=2021032312:RH:entire atmosphere (considered as a single layer):26 hour fcst: -20: 7:110291:d=2021032312:TOZNE:entire atmosphere (considered as a single layer):26 hour fcst: -20: 8:124606:d=2021032312:LCDC:low cloud layer:26 hour fcst: -20: 9:142744:d=2021032312:LCDC:low cloud layer:24-26 hour ave fcst: -20: 10:161440:d=2021032312:MCDC:middle cloud layer:26 hour fcst: -20: 11:177060:d=2021032312:MCDC:middle cloud layer:24-26 hour ave fcst: -20: 12:192991:d=2021032312:HCDC:high cloud layer:26 hour fcst: -20: 13:207841:d=2021032312:HCDC:high cloud layer:24-26 hour ave fcst: -20: 14:224265:d=2021032312:TCDC:entire atmosphere (considered as a single layer):26 hour fcst: -20: 15:241316:d=2021032312:TCDC:entire atmosphere (considered as a single layer):24-26 hour ave fcst: -20: 16:259378:d=2021032312:HGT:cloud ceiling:26 hour fcst: -20: 17:291358:d=2021032312:PRES:convective cloud bottom level:26 hour fcst: -20: 18:304996:d=2021032312:PRES:low cloud bottom level:24-26 hour ave fcst: -20: 19:326605:d=2021032312:PRES:middle cloud bottom level:24-26 hour ave fcst: -20: 20:343408:d=2021032312:PRES:high cloud bottom level:24-26 hour ave fcst: -20: 21:360844:d=2021032312:PRES:convective cloud top level:26 hour fcst: -20: 22:376170:d=2021032312:PRES:low cloud top level:24-26 hour ave fcst: -20: 23:397688:d=2021032312:PRES:middle cloud top level:24-26 hour ave fcst: -20: 24:414333:d=2021032312:PRES:high cloud top level:24-26 hour ave fcst: -20: 25:432264:d=2021032312:TMP:low cloud top level:24-26 hour ave fcst: -20: 26:448490:d=2021032312:TMP:middle cloud top level:24-26 hour ave fcst: -20: 27:461792:d=2021032312:TMP:high cloud top level:24-26 hour ave fcst: -20: 28:476733:d=2021032312:TCDC:convective cloud layer:26 hour fcst: -20: 29:497177:d=2021032312:TCDC:boundary layer cloud layer:24-26 hour ave fcst: -20: 30:513314:d=2021032312:CWORK:entire atmosphere (considered as a single layer):24-26 hour ave fcst: -20: 31:526185:d=2021032312:DSWRF:surface:24-26 hour ave fcst: -20: + interp_atmos_master.sh[47]export err=0 -20: + interp_atmos_master.sh[47]err=0 -20: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -20: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -20: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_21_0p25 -20: + product_functions.sh[5]local filename=pgb2file_f026_21_0p25 -20: + product_functions.sh[6]wgrib2 pgb2file_f026_21_0p25 -not_if :RH: -grib pgb2file_f026_21_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_21_0p25.new -20: 1:0:d=2021032312:LFTX:surface:26 hour fcst: -20: 2:298337:d=2021032312:CAPE:surface:26 hour fcst: -20: 3:629231:d=2021032312:CIN:surface:26 hour fcst: -20: 4:888968:d=2021032312:PWAT:entire atmosphere (considered as a single layer):26 hour fcst: -20: 5:1619830:d=2021032312:CWAT:entire atmosphere (considered as a single layer):26 hour fcst: -20: 6:2286108:d=2021032312:RH:entire atmosphere (considered as a single layer):26 hour fcst: -20: 7:2622871:d=2021032312:TOZNE:entire atmosphere (considered as a single layer):26 hour fcst: -20: 8:2931165:d=2021032312:LCDC:low cloud layer:26 hour fcst: -20: 9:3273944:d=2021032312:LCDC:low cloud layer:24-26 hour ave fcst: -20: 10:3635844:d=2021032312:MCDC:middle cloud layer:26 hour fcst: -20: 11:3905039:d=2021032312:MCDC:middle cloud layer:24-26 hour ave fcst: -20: 12:4196999:d=2021032312:HCDC:high cloud layer:26 hour fcst: -20: 13:4444858:d=2021032312:HCDC:high cloud layer:24-26 hour ave fcst: -20: 14:4729739:d=2021032312:TCDC:entire atmosphere (considered as a single layer):26 hour fcst: -20: 15:5041171:d=2021032312:TCDC:entire atmosphere (considered as a single layer):24-26 hour ave fcst: -20: 16:5383221:d=2021032312:HGT:cloud ceiling:26 hour fcst: -20: 17:6058226:d=2021032312:PRES:convective cloud bottom level:26 hour fcst: -20: 18:6551029:d=2021032312:PRES:low cloud bottom level:24-26 hour ave fcst: -20: 19:7241771:d=2021032312:PRES:middle cloud bottom level:24-26 hour ave fcst: -20: 20:7771046:d=2021032312:PRES:high cloud bottom level:24-26 hour ave fcst: -20: 21:8326661:d=2021032312:PRES:convective cloud top level:26 hour fcst: -20: 22:8871059:d=2021032312:PRES:low cloud top level:24-26 hour ave fcst: -20: 23:9544630:d=2021032312:PRES:middle cloud top level:24-26 hour ave fcst: -20: 24:10041527:d=2021032312:PRES:high cloud top level:24-26 hour ave fcst: -20: 25:10624454:d=2021032312:TMP:low cloud top level:24-26 hour ave fcst: -20: 26:11129616:d=2021032312:TMP:middle cloud top level:24-26 hour ave fcst: -20: 27:11537426:d=2021032312:TMP:high cloud top level:24-26 hour ave fcst: -20: 28:12024125:d=2021032312:TCDC:convective cloud layer:26 hour fcst: -20: 29:12500891:d=2021032312:TCDC:boundary layer cloud layer:24-26 hour ave fcst: -20: 30:12803397:d=2021032312:CWORK:entire atmosphere (considered as a single layer):24-26 hour ave fcst: -20: 31:13104585:d=2021032312:DSWRF:surface:24-26 hour ave fcst: -20: + product_functions.sh[10]rc=0 -20: + product_functions.sh[11](( rc == 0 )) -20: + product_functions.sh[11]mv pgb2file_f026_21_0p25.new pgb2file_f026_21_0p25 -20: + product_functions.sh[12]return 0 -20: + interp_atmos_master.sh[56]export err=0 -20: + interp_atmos_master.sh[56]err=0 -20: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -20: ++ interp_atmos_master.sh[62]wc -l -20: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_21_0p25 -match 'LAND|ICEC' -20: + interp_atmos_master.sh[62]var_count=0 -20: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -20: + interp_atmos_master.sh[73]exit 0 -14: + bash[8]'[' -z '' ']' -14: + bash[9]case "$-" in -14: + bash[12]__lmod_vx=x -14: + bash[16]'[' -n x ']' -14: + bash[16]set +x -14: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -14: Shell debugging restarted -14: + bash[224]unset __lmod_vx -14: + interp_atmos_master.sh[7]input_file=tmpfile_f026_15 -14: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_15 -14: + interp_atmos_master.sh[9]grid_string=0p25 -14: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -14: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -14: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -14: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -14: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -14: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -14: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -14: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -14: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -14: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -14: + interp_atmos_master.sh[31]IFS=: -14: + interp_atmos_master.sh[31]read -ra grids -14: + interp_atmos_master.sh[33]output_grids= -14: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -14: + interp_atmos_master.sh[35]gridopt=grid0p25 -14: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_15_0p25' -14: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_15 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_15_0p25 -14: 1:0:d=2021032312:ABSV:750 mb:26 hour fcst: -14: 2:21145:d=2021032312:CLMR:750 mb:26 hour fcst: -14: 3:30478:d=2021032312:ICMR:750 mb:26 hour fcst: -14: 4:41921:d=2021032312:RWMR:750 mb:26 hour fcst: -14: 5:47794:d=2021032312:SNMR:750 mb:26 hour fcst: -14: 6:61401:d=2021032312:GRLE:750 mb:26 hour fcst: -14: 7:63489:d=2021032312:O3MR:750 mb:26 hour fcst: -14: 8:90730:d=2021032312:HGT:800 mb:26 hour fcst: -14: 9:115065:d=2021032312:TMP:800 mb:26 hour fcst: -14: 10:135478:d=2021032312:RH:800 mb:26 hour fcst: -14: 11:155916:d=2021032312:TCDC:800 mb:26 hour fcst: -14: 12:169420:d=2021032312:SPFH:800 mb:26 hour fcst: -14: 13:198951:d=2021032312:VVEL:800 mb:26 hour fcst: -14: 14:224661:d=2021032312:DZDT:800 mb:26 hour fcst: -14: 15:251681:d=2021032312:UGRD:800 mb:26 hour fcst: -14: 16:272625:d=2021032312:VGRD:800 mb:26 hour fcst: -14: 17:294377:d=2021032312:ABSV:800 mb:26 hour fcst: -14: 18:315589:d=2021032312:CLMR:800 mb:26 hour fcst: -14: 19:327139:d=2021032312:ICMR:800 mb:26 hour fcst: -14: 20:338170:d=2021032312:RWMR:800 mb:26 hour fcst: -14: 21:346418:d=2021032312:SNMR:800 mb:26 hour fcst: -14: 22:359275:d=2021032312:GRLE:800 mb:26 hour fcst: -14: 23:361641:d=2021032312:O3MR:800 mb:26 hour fcst: -14: 24:388694:d=2021032312:HGT:850 mb:26 hour fcst: -14: 25:413474:d=2021032312:TMP:850 mb:26 hour fcst: -14: 26:434351:d=2021032312:RH:850 mb:26 hour fcst: -14: 27:454609:d=2021032312:TCDC:850 mb:26 hour fcst: -14: 28:469333:d=2021032312:SPFH:850 mb:26 hour fcst: -14: 29:498888:d=2021032312:VVEL:850 mb:26 hour fcst: -14: 30:524406:d=2021032312:DZDT:850 mb:26 hour fcst: -14: 31:550976:d=2021032312:UGRD:850 mb:26 hour fcst: -14: 32:572219:d=2021032312:VGRD:850 mb:26 hour fcst: -14: + interp_atmos_master.sh[47]export err=0 -14: + interp_atmos_master.sh[47]err=0 -14: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -14: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -14: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_15_0p25 -14: + product_functions.sh[5]local filename=pgb2file_f026_15_0p25 -14: + product_functions.sh[6]wgrib2 pgb2file_f026_15_0p25 -not_if :RH: -grib pgb2file_f026_15_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_15_0p25.new -14: 1:0:d=2021032312:ABSV:750 mb:26 hour fcst: -14: 2:502789:d=2021032312:CLMR:750 mb:26 hour fcst: -14: 3:703794:d=2021032312:ICMR:750 mb:26 hour fcst: -14: 4:973236:d=2021032312:RWMR:750 mb:26 hour fcst: -14: 5:1083911:d=2021032312:SNMR:750 mb:26 hour fcst: -14: 6:1379273:d=2021032312:GRLE:750 mb:26 hour fcst: -14: 7:1415536:d=2021032312:O3MR:750 mb:26 hour fcst: -14: 8:2146900:d=2021032312:HGT:800 mb:26 hour fcst: -14: 9:2781732:d=2021032312:TMP:800 mb:26 hour fcst: -14: 10:3261255:d=2021032312:RH:800 mb:26 hour fcst: -14: 11:3742660:d=2021032312:TCDC:800 mb:26 hour fcst: -14: 12:4015914:d=2021032312:SPFH:800 mb:26 hour fcst: -14: 13:4814309:d=2021032312:VVEL:800 mb:26 hour fcst: -14: 14:5499978:d=2021032312:DZDT:800 mb:26 hour fcst: -14: 15:6243896:d=2021032312:UGRD:800 mb:26 hour fcst: -14: 16:6753169:d=2021032312:VGRD:800 mb:26 hour fcst: -14: 17:7290688:d=2021032312:ABSV:800 mb:26 hour fcst: -14: 18:7797038:d=2021032312:CLMR:800 mb:26 hour fcst: -14: 19:8055373:d=2021032312:ICMR:800 mb:26 hour fcst: -14: 20:8316623:d=2021032312:RWMR:800 mb:26 hour fcst: -14: 21:8474387:d=2021032312:SNMR:800 mb:26 hour fcst: -14: 22:8758553:d=2021032312:GRLE:800 mb:26 hour fcst: -14: 23:8801614:d=2021032312:O3MR:800 mb:26 hour fcst: -14: 24:9524697:d=2021032312:HGT:850 mb:26 hour fcst: -14: 25:10165280:d=2021032312:TMP:850 mb:26 hour fcst: -14: 26:10657264:d=2021032312:RH:850 mb:26 hour fcst: -14: 27:11125616:d=2021032312:TCDC:850 mb:26 hour fcst: -14: 28:11423201:d=2021032312:SPFH:850 mb:26 hour fcst: -14: 29:12226506:d=2021032312:VVEL:850 mb:26 hour fcst: -14: 30:12905587:d=2021032312:DZDT:850 mb:26 hour fcst: -14: 31:13639144:d=2021032312:UGRD:850 mb:26 hour fcst: -14: 32:14150271:d=2021032312:VGRD:850 mb:26 hour fcst: -14: + product_functions.sh[10]rc=0 -14: + product_functions.sh[11](( rc == 0 )) -14: + product_functions.sh[11]mv pgb2file_f026_15_0p25.new pgb2file_f026_15_0p25 -14: + product_functions.sh[12]return 0 -14: + interp_atmos_master.sh[56]export err=0 -14: + interp_atmos_master.sh[56]err=0 -14: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -14: ++ interp_atmos_master.sh[62]wc -l -14: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_15_0p25 -match 'LAND|ICEC' -14: + interp_atmos_master.sh[62]var_count=0 -14: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -14: + interp_atmos_master.sh[73]exit 0 - 5: + bash[8]'[' -z '' ']' - 5: + bash[9]case "$-" in - 5: + bash[12]__lmod_vx=x - 5: + bash[16]'[' -n x ']' - 5: + bash[16]set +x - 5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 5: Shell debugging restarted - 5: + bash[224]unset __lmod_vx - 5: + interp_atmos_master.sh[7]input_file=tmpfile_f026_6 - 5: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_6 - 5: + interp_atmos_master.sh[9]grid_string=0p25 - 5: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 5: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 5: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 5: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 5: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 5: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 5: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 5: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 5: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 5: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 5: + interp_atmos_master.sh[31]IFS=: - 5: + interp_atmos_master.sh[31]read -ra grids - 5: + interp_atmos_master.sh[33]output_grids= - 5: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 5: + interp_atmos_master.sh[35]gridopt=grid0p25 - 5: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_6_0p25' - 5: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_6 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_6_0p25 - 5: 1:0:d=2021032312:RH:15 mb:26 hour fcst: - 5: 2:5269:d=2021032312:SPFH:15 mb:26 hour fcst: - 5: 3:20698:d=2021032312:VVEL:15 mb:26 hour fcst: - 5: 4:47221:d=2021032312:DZDT:15 mb:26 hour fcst: - 5: 5:77542:d=2021032312:UGRD:15 mb:26 hour fcst: - 5: 6:88743:d=2021032312:VGRD:15 mb:26 hour fcst: - 5: 7:107207:d=2021032312:ABSV:15 mb:26 hour fcst: - 5: 8:124842:d=2021032312:O3MR:15 mb:26 hour fcst: - 5: 9:143732:d=2021032312:HGT:20 mb:26 hour fcst: - 5: 10:163088:d=2021032312:TMP:20 mb:26 hour fcst: - 5: 11:179000:d=2021032312:RH:20 mb:26 hour fcst: - 5: 12:185160:d=2021032312:SPFH:20 mb:26 hour fcst: - 5: 13:199393:d=2021032312:VVEL:20 mb:26 hour fcst: - 5: 14:226834:d=2021032312:DZDT:20 mb:26 hour fcst: - 5: 15:257164:d=2021032312:UGRD:20 mb:26 hour fcst: - 5: 16:275822:d=2021032312:VGRD:20 mb:26 hour fcst: - 5: 17:294337:d=2021032312:ABSV:20 mb:26 hour fcst: - 5: 18:311850:d=2021032312:O3MR:20 mb:26 hour fcst: - 5: 19:336588:d=2021032312:HGT:30 mb:26 hour fcst: - 5: 20:355650:d=2021032312:TMP:30 mb:26 hour fcst: - 5: 21:371551:d=2021032312:RH:30 mb:26 hour fcst: - 5: 22:379145:d=2021032312:SPFH:30 mb:26 hour fcst: - 5: 23:394452:d=2021032312:VVEL:30 mb:26 hour fcst: - 5: 24:422776:d=2021032312:DZDT:30 mb:26 hour fcst: - 5: 25:452666:d=2021032312:UGRD:30 mb:26 hour fcst: - 5: 26:471239:d=2021032312:VGRD:30 mb:26 hour fcst: - 5: 27:489745:d=2021032312:ABSV:30 mb:26 hour fcst: - 5: 28:507422:d=2021032312:O3MR:30 mb:26 hour fcst: - 5: 29:534134:d=2021032312:HGT:40 mb:26 hour fcst: - 5: 30:553010:d=2021032312:TMP:40 mb:26 hour fcst: - 5: 31:569303:d=2021032312:RH:40 mb:26 hour fcst: - 5: + interp_atmos_master.sh[47]export err=0 - 5: + interp_atmos_master.sh[47]err=0 - 5: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 5: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 5: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_6_0p25 - 5: + product_functions.sh[5]local filename=pgb2file_f026_6_0p25 - 5: + product_functions.sh[6]wgrib2 pgb2file_f026_6_0p25 -not_if :RH: -grib pgb2file_f026_6_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_6_0p25.new - 5: 1:0:d=2021032312:RH:15 mb:26 hour fcst: - 5: 2:116002:d=2021032312:SPFH:15 mb:26 hour fcst: - 5: 3:490188:d=2021032312:VVEL:15 mb:26 hour fcst: - 5: 4:1227743:d=2021032312:DZDT:15 mb:26 hour fcst: - 5: 5:2143253:d=2021032312:UGRD:15 mb:26 hour fcst: - 5: 6:2580895:d=2021032312:VGRD:15 mb:26 hour fcst: - 5: 7:3008380:d=2021032312:ABSV:15 mb:26 hour fcst: - 5: 8:3397566:d=2021032312:O3MR:15 mb:26 hour fcst: - 5: 9:3841796:d=2021032312:HGT:20 mb:26 hour fcst: - 5: 10:4295335:d=2021032312:TMP:20 mb:26 hour fcst: - 5: 11:4646340:d=2021032312:RH:20 mb:26 hour fcst: - 5: 12:4810336:d=2021032312:SPFH:20 mb:26 hour fcst: - 5: 13:5144759:d=2021032312:VVEL:20 mb:26 hour fcst: - 5: 14:5906784:d=2021032312:DZDT:20 mb:26 hour fcst: - 5: 15:6815394:d=2021032312:UGRD:20 mb:26 hour fcst: - 5: 16:7249463:d=2021032312:VGRD:20 mb:26 hour fcst: - 5: 17:7677830:d=2021032312:ABSV:20 mb:26 hour fcst: - 5: 18:8067952:d=2021032312:O3MR:20 mb:26 hour fcst: - 5: 19:8730740:d=2021032312:HGT:30 mb:26 hour fcst: - 5: 20:9181343:d=2021032312:TMP:30 mb:26 hour fcst: - 5: 21:9533478:d=2021032312:RH:30 mb:26 hour fcst: - 5: 22:9745402:d=2021032312:SPFH:30 mb:26 hour fcst: - 5: 23:10111124:d=2021032312:VVEL:30 mb:26 hour fcst: - 5: 24:10914308:d=2021032312:DZDT:30 mb:26 hour fcst: - 5: 25:11806731:d=2021032312:UGRD:30 mb:26 hour fcst: - 5: 26:12241339:d=2021032312:VGRD:30 mb:26 hour fcst: - 5: 27:12671437:d=2021032312:ABSV:30 mb:26 hour fcst: - 5: 28:13066141:d=2021032312:O3MR:30 mb:26 hour fcst: - 5: 29:13796206:d=2021032312:HGT:40 mb:26 hour fcst: - 5: 30:14235398:d=2021032312:TMP:40 mb:26 hour fcst: - 5: 31:14598362:d=2021032312:RH:40 mb:26 hour fcst: - 5: + product_functions.sh[10]rc=0 - 5: + product_functions.sh[11](( rc == 0 )) - 5: + product_functions.sh[11]mv pgb2file_f026_6_0p25.new pgb2file_f026_6_0p25 - 5: + product_functions.sh[12]return 0 - 5: + interp_atmos_master.sh[56]export err=0 - 5: + interp_atmos_master.sh[56]err=0 - 5: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 5: ++ interp_atmos_master.sh[62]wc -l - 5: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_6_0p25 -match 'LAND|ICEC' - 5: + interp_atmos_master.sh[62]var_count=0 - 5: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 5: + interp_atmos_master.sh[73]exit 0 -11: + bash[8]'[' -z '' ']' -11: + bash[9]case "$-" in -11: + bash[12]__lmod_vx=x -11: + bash[16]'[' -n x ']' -11: + bash[16]set +x -11: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -11: Shell debugging restarted -11: + bash[224]unset __lmod_vx -11: + interp_atmos_master.sh[7]input_file=tmpfile_f026_12 -11: + interp_atmos_master.sh[8]output_file_prefix=pgb2file_f026_12 -11: + interp_atmos_master.sh[9]grid_string=0p25 -11: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -11: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -11: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -11: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -11: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -11: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -11: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -11: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -11: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -11: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -11: + interp_atmos_master.sh[31]IFS=: -11: + interp_atmos_master.sh[31]read -ra grids -11: + interp_atmos_master.sh[33]output_grids= -11: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -11: + interp_atmos_master.sh[35]gridopt=grid0p25 -11: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_12_0p25' -11: + interp_atmos_master.sh[40]wgrib2 tmpfile_f026_12 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2file_f026_12_0p25 -11: 1:0:d=2021032312:CLMR:450 mb:26 hour fcst: -11: 2:860:d=2021032312:ICMR:450 mb:26 hour fcst: -11: 3:17600:d=2021032312:RWMR:450 mb:26 hour fcst: -11: 4:18120:d=2021032312:SNMR:450 mb:26 hour fcst: -11: 5:31726:d=2021032312:GRLE:450 mb:26 hour fcst: -11: 6:33134:d=2021032312:O3MR:450 mb:26 hour fcst: -11: 7:57239:d=2021032312:HGT:500 mb:26 hour fcst: -11: 8:80967:d=2021032312:TMP:500 mb:26 hour fcst: -11: 9:99661:d=2021032312:RH:500 mb:26 hour fcst: -11: 10:119196:d=2021032312:TCDC:500 mb:26 hour fcst: -11: 11:131497:d=2021032312:SPFH:500 mb:26 hour fcst: -11: 12:159752:d=2021032312:VVEL:500 mb:26 hour fcst: -11: 13:184993:d=2021032312:DZDT:500 mb:26 hour fcst: -11: 14:212649:d=2021032312:UGRD:500 mb:26 hour fcst: -11: 15:234121:d=2021032312:VGRD:500 mb:26 hour fcst: -11: 16:256121:d=2021032312:ABSV:500 mb:26 hour fcst: -11: 17:277239:d=2021032312:CLMR:500 mb:26 hour fcst: -11: 18:278867:d=2021032312:ICMR:500 mb:26 hour fcst: -11: 19:294978:d=2021032312:RWMR:500 mb:26 hour fcst: -11: 20:295940:d=2021032312:SNMR:500 mb:26 hour fcst: -11: 21:310221:d=2021032312:GRLE:500 mb:26 hour fcst: -11: 22:312106:d=2021032312:O3MR:500 mb:26 hour fcst: -11: 23:335557:d=2021032312:HGT:550 mb:26 hour fcst: -11: 24:359092:d=2021032312:TMP:550 mb:26 hour fcst: -11: 25:377939:d=2021032312:RH:550 mb:26 hour fcst: -11: 26:397700:d=2021032312:TCDC:550 mb:26 hour fcst: -11: 27:409894:d=2021032312:SPFH:550 mb:26 hour fcst: -11: 28:439403:d=2021032312:VVEL:550 mb:26 hour fcst: -11: 29:464734:d=2021032312:DZDT:550 mb:26 hour fcst: -11: 30:492281:d=2021032312:UGRD:550 mb:26 hour fcst: -11: 31:513293:d=2021032312:VGRD:550 mb:26 hour fcst: -11: + interp_atmos_master.sh[47]export err=0 -11: + interp_atmos_master.sh[47]err=0 -11: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -11: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -11: + interp_atmos_master.sh[55]trim_rh pgb2file_f026_12_0p25 -11: + product_functions.sh[5]local filename=pgb2file_f026_12_0p25 -11: + product_functions.sh[6]wgrib2 pgb2file_f026_12_0p25 -not_if :RH: -grib pgb2file_f026_12_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2file_f026_12_0p25.new -11: 1:0:d=2021032312:CLMR:450 mb:26 hour fcst: -11: 2:16646:d=2021032312:ICMR:450 mb:26 hour fcst: -11: 3:385823:d=2021032312:RWMR:450 mb:26 hour fcst: -11: 4:391988:d=2021032312:SNMR:450 mb:26 hour fcst: -11: 5:678678:d=2021032312:GRLE:450 mb:26 hour fcst: -11: 6:696629:d=2021032312:O3MR:450 mb:26 hour fcst: -11: 7:1311446:d=2021032312:HGT:500 mb:26 hour fcst: -11: 8:1923197:d=2021032312:TMP:500 mb:26 hour fcst: -11: 9:2352131:d=2021032312:RH:500 mb:26 hour fcst: -11: 10:2798639:d=2021032312:TCDC:500 mb:26 hour fcst: -11: 11:3029490:d=2021032312:SPFH:500 mb:26 hour fcst: -11: 12:3786904:d=2021032312:VVEL:500 mb:26 hour fcst: -11: 13:4456343:d=2021032312:DZDT:500 mb:26 hour fcst: -11: 14:5236568:d=2021032312:UGRD:500 mb:26 hour fcst: -11: 15:5759236:d=2021032312:VGRD:500 mb:26 hour fcst: -11: 16:6310038:d=2021032312:ABSV:500 mb:26 hour fcst: -11: 17:6814510:d=2021032312:CLMR:500 mb:26 hour fcst: -11: 18:6847223:d=2021032312:ICMR:500 mb:26 hour fcst: -11: 19:7203180:d=2021032312:RWMR:500 mb:26 hour fcst: -11: 20:7218500:d=2021032312:SNMR:500 mb:26 hour fcst: -11: 21:7526755:d=2021032312:GRLE:500 mb:26 hour fcst: -11: 22:7553608:d=2021032312:O3MR:500 mb:26 hour fcst: -11: 23:8152437:d=2021032312:HGT:550 mb:26 hour fcst: -11: 24:8749891:d=2021032312:TMP:550 mb:26 hour fcst: -11: 25:9172145:d=2021032312:RH:550 mb:26 hour fcst: -11: 26:9623373:d=2021032312:TCDC:550 mb:26 hour fcst: -11: 27:9856282:d=2021032312:SPFH:550 mb:26 hour fcst: -11: 28:10638342:d=2021032312:VVEL:550 mb:26 hour fcst: -11: 29:11310758:d=2021032312:DZDT:550 mb:26 hour fcst: -11: 30:12075938:d=2021032312:UGRD:550 mb:26 hour fcst: -11: 31:12588809:d=2021032312:VGRD:550 mb:26 hour fcst: -11: + product_functions.sh[10]rc=0 -11: + product_functions.sh[11](( rc == 0 )) -11: + product_functions.sh[11]mv pgb2file_f026_12_0p25.new pgb2file_f026_12_0p25 -11: + product_functions.sh[12]return 0 -11: + interp_atmos_master.sh[56]export err=0 -11: + interp_atmos_master.sh[56]err=0 -11: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -11: ++ interp_atmos_master.sh[62]wc -l -11: ++ interp_atmos_master.sh[62]wgrib2 pgb2file_f026_12_0p25 -match 'LAND|ICEC' -11: + interp_atmos_master.sh[62]var_count=0 -11: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -11: + interp_atmos_master.sh[73]exit 0 -+ run_mpmd.sh[113]exit 0 -+ run_mpmd.sh[1]postamble run_mpmd.sh 1753756075 0 -+ preamble.sh[62]set +x -End run_mpmd.sh at 02:28:16 with error code 0 (time elapsed: 00:00:21) -+ exglobal_atmos_products.sh[142]true -+ exglobal_atmos_products.sh[143]export err=0 -+ exglobal_atmos_products.sh[143]err=0 -+ exglobal_atmos_products.sh[144][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[149]mv mpmd.out mpmd_1.out -+ exglobal_atmos_products.sh[153]echo 'Concatenating processor-specific grib2 files into a single product file' -Concatenating processor-specific grib2 files into a single product file -+ exglobal_atmos_products.sh[154](( iproc = 1 )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_1_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_1_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_1 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_2_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_2_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_2 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_3_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_3_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_3 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_4_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_4_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_4 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_5_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_5_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_5 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_6_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_6_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_6 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_7_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_7_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_7 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_8_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_8_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_8 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_9_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_9_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_9 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_10_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_10_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_10 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_11_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_11_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_11 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_12_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_12_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_12 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_13_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_13_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_13 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_14_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_14_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_14 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_15_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_15_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_15 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_16_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_16_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_16 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_17_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_17_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_17 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_18_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_18_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_18 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_19_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_19_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_19 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_20_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_20_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_20 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_21_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_21_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_21 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_22_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_22_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_22 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_23_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_23_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_23 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2file_f026_24_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2file_f026_24_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfile_f026_24 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[164]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[165]prod_dir=COMOUT_ATMOS_GRIB_0p25 -+ exglobal_atmos_products.sh[166]cpfs pgb2file_f026_0p25 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f026 -+ cpfs[3]'[' 2 -ne 2 ']' -+ cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f026 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f026 = ./ ']' -+ cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f026 ']' -+ cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f026 -+ cpfs[16]cp pgb2file_f026_0p25 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f026.cptmp -+ cpfs[18]'[' 0 -ne 0 ']' -+ cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f026.cptmp -+ cpfs[23]'[' 0 -ne 0 ']' -+ cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f026.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f026 -+ cpfs[28]'[' 0 -ne 0 ']' -+ exglobal_atmos_products.sh[167]wgrib2 -s pgb2file_f026_0p25 -+ exglobal_atmos_products.sh[170]echo 'Finished processing nset = 1' -Finished processing nset = 1 -+ exglobal_atmos_products.sh[154](( nset++ )) -+ exglobal_atmos_products.sh[154](( nset <= downset )) -+ exglobal_atmos_products.sh[74]echo 'Begin processing nset = 2' -Begin processing nset = 2 -+ exglobal_atmos_products.sh[77]nproc=24 -+ exglobal_atmos_products.sh[80][[ 2 == 1 ]] -+ exglobal_atmos_products.sh[82][[ 2 == 2 ]] -+ exglobal_atmos_products.sh[83]grp=b -+ exglobal_atmos_products.sh[87]tmpfile=tmpfileb_f026 -++ exglobal_atmos_products.sh[90]wgrib2 tmpfileb_f026 -++ exglobal_atmos_products.sh[90]wc -l -+ exglobal_atmos_products.sh[90]ncount=349 -+ exglobal_atmos_products.sh[91][[ 24 -gt 349 ]] -+ exglobal_atmos_products.sh[95]inv=14 -+ exglobal_atmos_products.sh[96]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601/poescript -+ exglobal_atmos_products.sh[98]last=0 -+ exglobal_atmos_products.sh[134](( iproc = 1 )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=1 -+ exglobal_atmos_products.sh[101]last=14 -+ exglobal_atmos_products.sh[102][[ 14 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 14 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 1 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 1:14 -grib tmpfileb_f026_1 -1:0:d=2021032312:HGT:1 mb:26 hour fcst: -2:17949:d=2021032312:TMP:1 mb:26 hour fcst: -3:33754:d=2021032312:RH:1 mb:26 hour fcst: -4:43532:d=2021032312:UGRD:1 mb:26 hour fcst: -5:61987:d=2021032312:VGRD:1 mb:26 hour fcst: -6:79619:d=2021032312:ABSV:1 mb:26 hour fcst: -7:95908:d=2021032312:O3MR:1 mb:26 hour fcst: -8:117173:d=2021032312:HGT:2 mb:26 hour fcst: -9:137378:d=2021032312:TMP:2 mb:26 hour fcst: -10:153764:d=2021032312:RH:2 mb:26 hour fcst: -11:162126:d=2021032312:UGRD:2 mb:26 hour fcst: -12:173614:d=2021032312:VGRD:2 mb:26 hour fcst: -13:184318:d=2021032312:ABSV:2 mb:26 hour fcst: -14:201326:d=2021032312:O3MR:2 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_1 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_1 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_1 pgb2bfile_f026_1 0p25' -+ exglobal_atmos_products.sh[133][[ 14 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=15 -+ exglobal_atmos_products.sh[101]last=28 -+ exglobal_atmos_products.sh[102][[ 28 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 28 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 2 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 15:28 -grib tmpfileb_f026_2 -15:223837:d=2021032312:HGT:3 mb:26 hour fcst: -16:243778:d=2021032312:TMP:3 mb:26 hour fcst: -17:260188:d=2021032312:RH:3 mb:26 hour fcst: -18:266250:d=2021032312:UGRD:3 mb:26 hour fcst: -19:277781:d=2021032312:VGRD:3 mb:26 hour fcst: -20:295743:d=2021032312:ABSV:3 mb:26 hour fcst: -21:312636:d=2021032312:O3MR:3 mb:26 hour fcst: -22:335968:d=2021032312:HGT:5 mb:26 hour fcst: -23:353707:d=2021032312:TMP:5 mb:26 hour fcst: -24:370341:d=2021032312:RH:5 mb:26 hour fcst: -25:379518:d=2021032312:UGRD:5 mb:26 hour fcst: -26:391017:d=2021032312:VGRD:5 mb:26 hour fcst: -27:409872:d=2021032312:ABSV:5 mb:26 hour fcst: -28:427484:d=2021032312:O3MR:5 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_2 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_2 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_2 pgb2bfile_f026_2 0p25' -+ exglobal_atmos_products.sh[133][[ 28 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=29 -+ exglobal_atmos_products.sh[101]last=42 -+ exglobal_atmos_products.sh[102][[ 42 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 42 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 3 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 29:42 -grib tmpfileb_f026_3 -29:450109:d=2021032312:HGT:7 mb:26 hour fcst: -30:467767:d=2021032312:TMP:7 mb:26 hour fcst: -31:483909:d=2021032312:RH:7 mb:26 hour fcst: -32:489969:d=2021032312:UGRD:7 mb:26 hour fcst: -33:501363:d=2021032312:VGRD:7 mb:26 hour fcst: -34:520179:d=2021032312:ABSV:7 mb:26 hour fcst: -35:537910:d=2021032312:O3MR:7 mb:26 hour fcst: -36:561501:d=2021032312:TCDC:70 mb:26 hour fcst: -37:561680:d=2021032312:CLMR:70 mb:26 hour fcst: -38:561859:d=2021032312:ICMR:70 mb:26 hour fcst: -39:562038:d=2021032312:RWMR:70 mb:26 hour fcst: -40:562217:d=2021032312:SNMR:70 mb:26 hour fcst: -41:562396:d=2021032312:GRLE:70 mb:26 hour fcst: -42:562575:d=2021032312:HGT:125 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_3 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_3 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_3 pgb2bfile_f026_3 0p25' -+ exglobal_atmos_products.sh[133][[ 42 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=43 -+ exglobal_atmos_products.sh[101]last=56 -+ exglobal_atmos_products.sh[102][[ 56 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 56 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 4 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 43:56 -grib tmpfileb_f026_4 -43:582612:d=2021032312:TMP:125 mb:26 hour fcst: -44:599549:d=2021032312:RH:125 mb:26 hour fcst: -45:610029:d=2021032312:TCDC:125 mb:26 hour fcst: -46:612623:d=2021032312:VVEL:125 mb:26 hour fcst: -47:638713:d=2021032312:DZDT:125 mb:26 hour fcst: -48:663790:d=2021032312:UGRD:125 mb:26 hour fcst: -49:683446:d=2021032312:VGRD:125 mb:26 hour fcst: -50:703089:d=2021032312:ABSV:125 mb:26 hour fcst: -51:721526:d=2021032312:CLMR:125 mb:26 hour fcst: -52:721705:d=2021032312:ICMR:125 mb:26 hour fcst: -53:726433:d=2021032312:RWMR:125 mb:26 hour fcst: -54:726612:d=2021032312:SNMR:125 mb:26 hour fcst: -55:730787:d=2021032312:GRLE:125 mb:26 hour fcst: -56:731197:d=2021032312:HGT:175 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_4 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_4 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_4 pgb2bfile_f026_4 0p25' -+ exglobal_atmos_products.sh[133][[ 56 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=57 -+ exglobal_atmos_products.sh[101]last=70 -+ exglobal_atmos_products.sh[102][[ 70 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 70 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 5 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 57:70 -grib tmpfileb_f026_5 -57:751862:d=2021032312:TMP:175 mb:26 hour fcst: -58:769408:d=2021032312:RH:175 mb:26 hour fcst: -59:782581:d=2021032312:TCDC:175 mb:26 hour fcst: -60:787458:d=2021032312:VVEL:175 mb:26 hour fcst: -61:815832:d=2021032312:DZDT:175 mb:26 hour fcst: -62:841888:d=2021032312:UGRD:175 mb:26 hour fcst: -63:854887:d=2021032312:VGRD:175 mb:26 hour fcst: -64:875595:d=2021032312:ABSV:175 mb:26 hour fcst: -65:895002:d=2021032312:CLMR:175 mb:26 hour fcst: -66:895181:d=2021032312:ICMR:175 mb:26 hour fcst: -67:901866:d=2021032312:RWMR:175 mb:26 hour fcst: -68:902045:d=2021032312:SNMR:175 mb:26 hour fcst: -69:906419:d=2021032312:GRLE:175 mb:26 hour fcst: -70:906837:d=2021032312:HGT:225 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_5 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_5 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_5 pgb2bfile_f026_5 0p25' -+ exglobal_atmos_products.sh[133][[ 70 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=71 -+ exglobal_atmos_products.sh[101]last=84 -+ exglobal_atmos_products.sh[102][[ 84 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 84 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 6 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 71:84 -grib tmpfileb_f026_6 -71:928309:d=2021032312:TMP:225 mb:26 hour fcst: -72:946922:d=2021032312:RH:225 mb:26 hour fcst: -73:963598:d=2021032312:TCDC:225 mb:26 hour fcst: -74:972428:d=2021032312:VVEL:225 mb:26 hour fcst: -75:999965:d=2021032312:DZDT:225 mb:26 hour fcst: -76:1026711:d=2021032312:UGRD:225 mb:26 hour fcst: -77:1040510:d=2021032312:VGRD:225 mb:26 hour fcst: -78:1054859:d=2021032312:ABSV:225 mb:26 hour fcst: -79:1075679:d=2021032312:CLMR:225 mb:26 hour fcst: -80:1075858:d=2021032312:ICMR:225 mb:26 hour fcst: -81:1087358:d=2021032312:RWMR:225 mb:26 hour fcst: -82:1087537:d=2021032312:SNMR:225 mb:26 hour fcst: -83:1094390:d=2021032312:GRLE:225 mb:26 hour fcst: -84:1094665:d=2021032312:HGT:275 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_6 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_6 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_6 pgb2bfile_f026_6 0p25' -+ exglobal_atmos_products.sh[133][[ 84 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=85 -+ exglobal_atmos_products.sh[101]last=98 -+ exglobal_atmos_products.sh[102][[ 98 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 98 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 7 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 85:98 -grib tmpfileb_f026_7 -85:1116842:d=2021032312:TMP:275 mb:26 hour fcst: -86:1135652:d=2021032312:RH:275 mb:26 hour fcst: -87:1154092:d=2021032312:TCDC:275 mb:26 hour fcst: -88:1166038:d=2021032312:VVEL:275 mb:26 hour fcst: -89:1189349:d=2021032312:DZDT:275 mb:26 hour fcst: -90:1217203:d=2021032312:UGRD:275 mb:26 hour fcst: -91:1231550:d=2021032312:VGRD:275 mb:26 hour fcst: -92:1246939:d=2021032312:ABSV:275 mb:26 hour fcst: -93:1269019:d=2021032312:CLMR:275 mb:26 hour fcst: -94:1269198:d=2021032312:ICMR:275 mb:26 hour fcst: -95:1283800:d=2021032312:RWMR:275 mb:26 hour fcst: -96:1283979:d=2021032312:SNMR:275 mb:26 hour fcst: -97:1294142:d=2021032312:GRLE:275 mb:26 hour fcst: -98:1294435:d=2021032312:HGT:325 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_7 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_7 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_7 pgb2bfile_f026_7 0p25' -+ exglobal_atmos_products.sh[133][[ 98 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=99 -+ exglobal_atmos_products.sh[101]last=112 -+ exglobal_atmos_products.sh[102][[ 112 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 112 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 8 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 99:112 -grib tmpfileb_f026_8 -99:1316676:d=2021032312:TMP:325 mb:26 hour fcst: -100:1335018:d=2021032312:RH:325 mb:26 hour fcst: -101:1353955:d=2021032312:TCDC:325 mb:26 hour fcst: -102:1367098:d=2021032312:VVEL:325 mb:26 hour fcst: -103:1391541:d=2021032312:DZDT:325 mb:26 hour fcst: -104:1419752:d=2021032312:UGRD:325 mb:26 hour fcst: -105:1434403:d=2021032312:VGRD:325 mb:26 hour fcst: -106:1449667:d=2021032312:ABSV:325 mb:26 hour fcst: -107:1471932:d=2021032312:CLMR:325 mb:26 hour fcst: -108:1472111:d=2021032312:ICMR:325 mb:26 hour fcst: -109:1488749:d=2021032312:RWMR:325 mb:26 hour fcst: -110:1488928:d=2021032312:SNMR:325 mb:26 hour fcst: -111:1500889:d=2021032312:GRLE:325 mb:26 hour fcst: -112:1501304:d=2021032312:HGT:375 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_8 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_8 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_8 pgb2bfile_f026_8 0p25' -+ exglobal_atmos_products.sh[133][[ 112 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=113 -+ exglobal_atmos_products.sh[101]last=126 -+ exglobal_atmos_products.sh[102][[ 126 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 126 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 9 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 113:126 -grib tmpfileb_f026_9 -113:1523361:d=2021032312:TMP:375 mb:26 hour fcst: -114:1541686:d=2021032312:RH:375 mb:26 hour fcst: -115:1560850:d=2021032312:TCDC:375 mb:26 hour fcst: -116:1574060:d=2021032312:VVEL:375 mb:26 hour fcst: -117:1598892:d=2021032312:DZDT:375 mb:26 hour fcst: -118:1626992:d=2021032312:UGRD:375 mb:26 hour fcst: -119:1641474:d=2021032312:VGRD:375 mb:26 hour fcst: -120:1656572:d=2021032312:ABSV:375 mb:26 hour fcst: -121:1678709:d=2021032312:CLMR:375 mb:26 hour fcst: -122:1678965:d=2021032312:ICMR:375 mb:26 hour fcst: -123:1696054:d=2021032312:RWMR:375 mb:26 hour fcst: -124:1696311:d=2021032312:SNMR:375 mb:26 hour fcst: -125:1709438:d=2021032312:GRLE:375 mb:26 hour fcst: -126:1710052:d=2021032312:HGT:425 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_9 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_9 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_9 pgb2bfile_f026_9 0p25' -+ exglobal_atmos_products.sh[133][[ 126 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=127 -+ exglobal_atmos_products.sh[101]last=140 -+ exglobal_atmos_products.sh[102][[ 140 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 140 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 10 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 127:140 -grib tmpfileb_f026_10 -127:1734096:d=2021032312:TMP:425 mb:26 hour fcst: -128:1752540:d=2021032312:RH:425 mb:26 hour fcst: -129:1771812:d=2021032312:TCDC:425 mb:26 hour fcst: -130:1784682:d=2021032312:VVEL:425 mb:26 hour fcst: -131:1809757:d=2021032312:DZDT:425 mb:26 hour fcst: -132:1837942:d=2021032312:UGRD:425 mb:26 hour fcst: -133:1859796:d=2021032312:VGRD:425 mb:26 hour fcst: -134:1882313:d=2021032312:ABSV:425 mb:26 hour fcst: -135:1903922:d=2021032312:CLMR:425 mb:26 hour fcst: -136:1904576:d=2021032312:ICMR:425 mb:26 hour fcst: -137:1921351:d=2021032312:RWMR:425 mb:26 hour fcst: -138:1921828:d=2021032312:SNMR:425 mb:26 hour fcst: -139:1936371:d=2021032312:GRLE:425 mb:26 hour fcst: -140:1937548:d=2021032312:HGT:475 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_10 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_10 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_10 pgb2bfile_f026_10 0p25' -+ exglobal_atmos_products.sh[133][[ 140 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=141 -+ exglobal_atmos_products.sh[101]last=154 -+ exglobal_atmos_products.sh[102][[ 154 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 154 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 11 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 141:154 -grib tmpfileb_f026_11 -141:1961394:d=2021032312:TMP:475 mb:26 hour fcst: -142:1979870:d=2021032312:RH:475 mb:26 hour fcst: -143:1999528:d=2021032312:TCDC:475 mb:26 hour fcst: -144:2012041:d=2021032312:VVEL:475 mb:26 hour fcst: -145:2037048:d=2021032312:DZDT:475 mb:26 hour fcst: -146:2065003:d=2021032312:UGRD:475 mb:26 hour fcst: -147:2086581:d=2021032312:VGRD:475 mb:26 hour fcst: -148:2108738:d=2021032312:ABSV:475 mb:26 hour fcst: -149:2130019:d=2021032312:CLMR:475 mb:26 hour fcst: -150:2131201:d=2021032312:ICMR:475 mb:26 hour fcst: -151:2147556:d=2021032312:RWMR:475 mb:26 hour fcst: -152:2148239:d=2021032312:SNMR:475 mb:26 hour fcst: -153:2162201:d=2021032312:GRLE:475 mb:26 hour fcst: -154:2163716:d=2021032312:HGT:525 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_11 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_11 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_11 pgb2bfile_f026_11 0p25' -+ exglobal_atmos_products.sh[133][[ 154 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=155 -+ exglobal_atmos_products.sh[101]last=168 -+ exglobal_atmos_products.sh[102][[ 168 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 168 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 12 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 155:168 -grib tmpfileb_f026_12 -155:2187353:d=2021032312:TMP:525 mb:26 hour fcst: -156:2206146:d=2021032312:RH:525 mb:26 hour fcst: -157:2225822:d=2021032312:TCDC:525 mb:26 hour fcst: -158:2238062:d=2021032312:VVEL:525 mb:26 hour fcst: -159:2263165:d=2021032312:DZDT:525 mb:26 hour fcst: -160:2290761:d=2021032312:UGRD:525 mb:26 hour fcst: -161:2311941:d=2021032312:VGRD:525 mb:26 hour fcst: -162:2333884:d=2021032312:ABSV:525 mb:26 hour fcst: -163:2354874:d=2021032312:CLMR:525 mb:26 hour fcst: -164:2357190:d=2021032312:ICMR:525 mb:26 hour fcst: -165:2374421:d=2021032312:RWMR:525 mb:26 hour fcst: -166:2375550:d=2021032312:SNMR:525 mb:26 hour fcst: -167:2389965:d=2021032312:GRLE:525 mb:26 hour fcst: -168:2391613:d=2021032312:HGT:575 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_12 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_12 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_12 pgb2bfile_f026_12 0p25' -+ exglobal_atmos_products.sh[133][[ 168 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=169 -+ exglobal_atmos_products.sh[101]last=182 -+ exglobal_atmos_products.sh[102][[ 182 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 182 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 13 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 169:182 -grib tmpfileb_f026_13 -169:2415110:d=2021032312:TMP:575 mb:26 hour fcst: -170:2434055:d=2021032312:RH:575 mb:26 hour fcst: -171:2453926:d=2021032312:TCDC:575 mb:26 hour fcst: -172:2466128:d=2021032312:VVEL:575 mb:26 hour fcst: -173:2491344:d=2021032312:DZDT:575 mb:26 hour fcst: -174:2518792:d=2021032312:UGRD:575 mb:26 hour fcst: -175:2539738:d=2021032312:VGRD:575 mb:26 hour fcst: -176:2561211:d=2021032312:ABSV:575 mb:26 hour fcst: -177:2582054:d=2021032312:CLMR:575 mb:26 hour fcst: -178:2586313:d=2021032312:ICMR:575 mb:26 hour fcst: -179:2602306:d=2021032312:RWMR:575 mb:26 hour fcst: -180:2604661:d=2021032312:SNMR:575 mb:26 hour fcst: -181:2619219:d=2021032312:GRLE:575 mb:26 hour fcst: -182:2621247:d=2021032312:HGT:625 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_13 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_13 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_13 pgb2bfile_f026_13 0p25' -+ exglobal_atmos_products.sh[133][[ 182 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=183 -+ exglobal_atmos_products.sh[101]last=196 -+ exglobal_atmos_products.sh[102][[ 196 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 196 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 14 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 183:196 -grib tmpfileb_f026_14 -183:2645543:d=2021032312:TMP:625 mb:26 hour fcst: -184:2664688:d=2021032312:RH:625 mb:26 hour fcst: -185:2684775:d=2021032312:TCDC:625 mb:26 hour fcst: -186:2696592:d=2021032312:VVEL:625 mb:26 hour fcst: -187:2722073:d=2021032312:DZDT:625 mb:26 hour fcst: -188:2749483:d=2021032312:UGRD:625 mb:26 hour fcst: -189:2770484:d=2021032312:VGRD:625 mb:26 hour fcst: -190:2792136:d=2021032312:ABSV:625 mb:26 hour fcst: -191:2812918:d=2021032312:CLMR:625 mb:26 hour fcst: -192:2817976:d=2021032312:ICMR:625 mb:26 hour fcst: -193:2831011:d=2021032312:RWMR:625 mb:26 hour fcst: -194:2835250:d=2021032312:SNMR:625 mb:26 hour fcst: -195:2850518:d=2021032312:GRLE:625 mb:26 hour fcst: -196:2852946:d=2021032312:HGT:675 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_14 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_14 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_14 pgb2bfile_f026_14 0p25' -+ exglobal_atmos_products.sh[133][[ 196 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=197 -+ exglobal_atmos_products.sh[101]last=210 -+ exglobal_atmos_products.sh[102][[ 210 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 210 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 15 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 197:210 -grib tmpfileb_f026_15 -197:2877300:d=2021032312:TMP:675 mb:26 hour fcst: -198:2896800:d=2021032312:RH:675 mb:26 hour fcst: -199:2917071:d=2021032312:TCDC:675 mb:26 hour fcst: -200:2928668:d=2021032312:VVEL:675 mb:26 hour fcst: -201:2954326:d=2021032312:DZDT:675 mb:26 hour fcst: -202:2981727:d=2021032312:UGRD:675 mb:26 hour fcst: -203:3002781:d=2021032312:VGRD:675 mb:26 hour fcst: -204:3024441:d=2021032312:ABSV:675 mb:26 hour fcst: -205:3045533:d=2021032312:CLMR:675 mb:26 hour fcst: -206:3051122:d=2021032312:ICMR:675 mb:26 hour fcst: -207:3062872:d=2021032312:RWMR:675 mb:26 hour fcst: -208:3067948:d=2021032312:SNMR:675 mb:26 hour fcst: -209:3082449:d=2021032312:GRLE:675 mb:26 hour fcst: -210:3084335:d=2021032312:HGT:725 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_15 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_15 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_15 pgb2bfile_f026_15 0p25' -+ exglobal_atmos_products.sh[133][[ 210 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=211 -+ exglobal_atmos_products.sh[101]last=224 -+ exglobal_atmos_products.sh[102][[ 224 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 224 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 16 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 211:224 -grib tmpfileb_f026_16 -211:3108484:d=2021032312:TMP:725 mb:26 hour fcst: -212:3128233:d=2021032312:RH:725 mb:26 hour fcst: -213:3148547:d=2021032312:TCDC:725 mb:26 hour fcst: -214:3160524:d=2021032312:VVEL:725 mb:26 hour fcst: -215:3186300:d=2021032312:DZDT:725 mb:26 hour fcst: -216:3213612:d=2021032312:UGRD:725 mb:26 hour fcst: -217:3234426:d=2021032312:VGRD:725 mb:26 hour fcst: -218:3256044:d=2021032312:ABSV:725 mb:26 hour fcst: -219:3277121:d=2021032312:CLMR:725 mb:26 hour fcst: -220:3285119:d=2021032312:ICMR:725 mb:26 hour fcst: -221:3296821:d=2021032312:RWMR:725 mb:26 hour fcst: -222:3302555:d=2021032312:SNMR:725 mb:26 hour fcst: -223:3316398:d=2021032312:GRLE:725 mb:26 hour fcst: -224:3318331:d=2021032312:HGT:775 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_16 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_16 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_16 pgb2bfile_f026_16 0p25' -+ exglobal_atmos_products.sh[133][[ 224 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=225 -+ exglobal_atmos_products.sh[101]last=238 -+ exglobal_atmos_products.sh[102][[ 238 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 238 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 17 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 225:238 -grib tmpfileb_f026_17 -225:3342790:d=2021032312:TMP:775 mb:26 hour fcst: -226:3362886:d=2021032312:RH:775 mb:26 hour fcst: -227:3383319:d=2021032312:TCDC:775 mb:26 hour fcst: -228:3395896:d=2021032312:VVEL:775 mb:26 hour fcst: -229:3421652:d=2021032312:DZDT:775 mb:26 hour fcst: -230:3448767:d=2021032312:UGRD:775 mb:26 hour fcst: -231:3469631:d=2021032312:VGRD:775 mb:26 hour fcst: -232:3491358:d=2021032312:ABSV:775 mb:26 hour fcst: -233:3512531:d=2021032312:CLMR:775 mb:26 hour fcst: -234:3522801:d=2021032312:ICMR:775 mb:26 hour fcst: -235:3534067:d=2021032312:RWMR:775 mb:26 hour fcst: -236:3540835:d=2021032312:SNMR:775 mb:26 hour fcst: -237:3554071:d=2021032312:GRLE:775 mb:26 hour fcst: -238:3556605:d=2021032312:HGT:825 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_17 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_17 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_17 pgb2bfile_f026_17 0p25' -+ exglobal_atmos_products.sh[133][[ 238 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=239 -+ exglobal_atmos_products.sh[101]last=252 -+ exglobal_atmos_products.sh[102][[ 252 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 252 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 18 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 239:252 -grib tmpfileb_f026_18 -239:3581200:d=2021032312:TMP:825 mb:26 hour fcst: -240:3601843:d=2021032312:RH:825 mb:26 hour fcst: -241:3622229:d=2021032312:TCDC:825 mb:26 hour fcst: -242:3636315:d=2021032312:VVEL:825 mb:26 hour fcst: -243:3661962:d=2021032312:DZDT:825 mb:26 hour fcst: -244:3688873:d=2021032312:UGRD:825 mb:26 hour fcst: -245:3710102:d=2021032312:VGRD:825 mb:26 hour fcst: -246:3731957:d=2021032312:ABSV:825 mb:26 hour fcst: -247:3753212:d=2021032312:CLMR:825 mb:26 hour fcst: -248:3765878:d=2021032312:ICMR:825 mb:26 hour fcst: -249:3776210:d=2021032312:RWMR:825 mb:26 hour fcst: -250:3785691:d=2021032312:SNMR:825 mb:26 hour fcst: -251:3797957:d=2021032312:GRLE:825 mb:26 hour fcst: -252:3800388:d=2021032312:HGT:875 mb:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_18 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_18 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_18 pgb2bfile_f026_18 0p25' -+ exglobal_atmos_products.sh[133][[ 252 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=253 -+ exglobal_atmos_products.sh[101]last=266 -+ exglobal_atmos_products.sh[102][[ 266 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 266 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 19 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 253:266 -grib tmpfileb_f026_19 -253:3825117:d=2021032312:TMP:875 mb:26 hour fcst: -254:3846047:d=2021032312:RH:875 mb:26 hour fcst: -255:3866040:d=2021032312:TCDC:875 mb:26 hour fcst: -256:3880277:d=2021032312:VVEL:875 mb:26 hour fcst: -257:3905621:d=2021032312:DZDT:875 mb:26 hour fcst: -258:3932004:d=2021032312:UGRD:875 mb:26 hour fcst: -259:3953307:d=2021032312:VGRD:875 mb:26 hour fcst: -260:3975158:d=2021032312:ABSV:875 mb:26 hour fcst: -261:3996406:d=2021032312:CLMR:875 mb:26 hour fcst: -262:4008633:d=2021032312:ICMR:875 mb:26 hour fcst: -263:4017794:d=2021032312:RWMR:875 mb:26 hour fcst: -264:4029118:d=2021032312:SNMR:875 mb:26 hour fcst: -265:4040117:d=2021032312:GRLE:875 mb:26 hour fcst: -266:4042514:d=2021032312:SOILL:0-0.1 m below ground:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_19 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_19 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_19 pgb2bfile_f026_19 0p25' -+ exglobal_atmos_products.sh[133][[ 266 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=267 -+ exglobal_atmos_products.sh[101]last=280 -+ exglobal_atmos_products.sh[102][[ 280 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 280 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 20 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 267:280 -grib tmpfileb_f026_20 -267:4051314:d=2021032312:SOILL:0.1-0.4 m below ground:26 hour fcst: -268:4060206:d=2021032312:SOILL:0.4-1 m below ground:26 hour fcst: -269:4069129:d=2021032312:SOILL:1-2 m below ground:26 hour fcst: -270:4078206:d=2021032312:CNWAT:surface:26 hour fcst: -271:4083728:d=2021032312:ICETK:surface:26 hour fcst: -272:4087136:d=2021032312:DUVB:surface:24-26 hour ave fcst: -273:4104566:d=2021032312:CDUVB:surface:24-26 hour ave fcst: -274:4118286:d=2021032312:TMP:305 m above mean sea level:26 hour fcst: -275:4136491:d=2021032312:UGRD:305 m above mean sea level:26 hour fcst: -276:4155654:d=2021032312:VGRD:305 m above mean sea level:26 hour fcst: -277:4175297:d=2021032312:TMP:457 m above mean sea level:26 hour fcst: -278:4194705:d=2021032312:UGRD:457 m above mean sea level:26 hour fcst: -279:4214947:d=2021032312:VGRD:457 m above mean sea level:26 hour fcst: -280:4235510:d=2021032312:TMP:610 m above mean sea level:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_20 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_20 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_20 pgb2bfile_f026_20 0p25' -+ exglobal_atmos_products.sh[133][[ 280 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=281 -+ exglobal_atmos_products.sh[101]last=294 -+ exglobal_atmos_products.sh[102][[ 294 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 294 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 21 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 281:294 -grib tmpfileb_f026_21 -281:4255849:d=2021032312:UGRD:610 m above mean sea level:26 hour fcst: -282:4276670:d=2021032312:VGRD:610 m above mean sea level:26 hour fcst: -283:4297866:d=2021032312:TMP:914 m above mean sea level:26 hour fcst: -284:4318766:d=2021032312:UGRD:914 m above mean sea level:26 hour fcst: -285:4340123:d=2021032312:VGRD:914 m above mean sea level:26 hour fcst: -286:4361929:d=2021032312:TMP:4572 m above mean sea level:26 hour fcst: -287:4383126:d=2021032312:UGRD:4572 m above mean sea level:26 hour fcst: -288:4406329:d=2021032312:VGRD:4572 m above mean sea level:26 hour fcst: -289:4430326:d=2021032312:TMP:60-30 mb above ground:26 hour fcst: -290:4451197:d=2021032312:RH:60-30 mb above ground:26 hour fcst: -291:4469834:d=2021032312:SPFH:60-30 mb above ground:26 hour fcst: -292:4498612:d=2021032312:UGRD:60-30 mb above ground:26 hour fcst: -293:4520221:d=2021032312:VGRD:60-30 mb above ground:26 hour fcst: -294:4542376:d=2021032312:TMP:90-60 mb above ground:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_21 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_21 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_21 pgb2bfile_f026_21 0p25' -+ exglobal_atmos_products.sh[133][[ 294 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=295 -+ exglobal_atmos_products.sh[101]last=308 -+ exglobal_atmos_products.sh[102][[ 308 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 308 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 22 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 295:308 -grib tmpfileb_f026_22 -295:4563245:d=2021032312:RH:90-60 mb above ground:26 hour fcst: -296:4582310:d=2021032312:SPFH:90-60 mb above ground:26 hour fcst: -297:4611215:d=2021032312:UGRD:90-60 mb above ground:26 hour fcst: -298:4632614:d=2021032312:VGRD:90-60 mb above ground:26 hour fcst: -299:4654622:d=2021032312:TMP:120-90 mb above ground:26 hour fcst: -300:4675548:d=2021032312:RH:120-90 mb above ground:26 hour fcst: -301:4694886:d=2021032312:SPFH:120-90 mb above ground:26 hour fcst: -302:4723844:d=2021032312:UGRD:120-90 mb above ground:26 hour fcst: -303:4745079:d=2021032312:VGRD:120-90 mb above ground:26 hour fcst: -304:4766846:d=2021032312:TMP:150-120 mb above ground:26 hour fcst: -305:4787753:d=2021032312:RH:150-120 mb above ground:26 hour fcst: -306:4807384:d=2021032312:SPFH:150-120 mb above ground:26 hour fcst: -307:4836430:d=2021032312:UGRD:150-120 mb above ground:26 hour fcst: -308:4857469:d=2021032312:VGRD:150-120 mb above ground:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_22 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_22 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_22 pgb2bfile_f026_22 0p25' -+ exglobal_atmos_products.sh[133][[ 308 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=309 -+ exglobal_atmos_products.sh[101]last=322 -+ exglobal_atmos_products.sh[102][[ 322 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 322 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 23 -eq 24 ]] -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 309:322 -grib tmpfileb_f026_23 -309:4879153:d=2021032312:TMP:180-150 mb above ground:26 hour fcst: -310:4900298:d=2021032312:RH:180-150 mb above ground:26 hour fcst: -311:4920304:d=2021032312:SPFH:180-150 mb above ground:26 hour fcst: -312:4949715:d=2021032312:UGRD:180-150 mb above ground:26 hour fcst: -313:4970769:d=2021032312:VGRD:180-150 mb above ground:26 hour fcst: -314:4992443:d=2021032312:UGRD:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -315:5003098:d=2021032312:VGRD:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -316:5013642:d=2021032312:TMP:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -317:5028310:d=2021032312:HGT:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -318:5045052:d=2021032312:PRES:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -319:5061938:d=2021032312:VWSH:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -320:5072935:d=2021032312:UGRD:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -321:5083998:d=2021032312:VGRD:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -322:5098449:d=2021032312:TMP:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_23 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_23 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_23 pgb2bfile_f026_23 0p25' -+ exglobal_atmos_products.sh[133][[ 322 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( iproc++ )) -+ exglobal_atmos_products.sh[134](( iproc <= nproc )) -+ exglobal_atmos_products.sh[100]first=323 -+ exglobal_atmos_products.sh[101]last=336 -+ exglobal_atmos_products.sh[102][[ 336 -gt 349 ]] -+ exglobal_atmos_products.sh[109]set +e -+ exglobal_atmos_products.sh[111]wgrib2 -d 336 tmpfileb_f026 -+ exglobal_atmos_products.sh[111]grep -E -i 'ugrd|ustm|uflx|u-gwd|land|maxuw' -+ exglobal_atmos_products.sh[112]rc=1 -+ exglobal_atmos_products.sh[113]set_strict -+ environment[0][[ YES == \Y\E\S ]] -+ environment[1]set -eu -+ exglobal_atmos_products.sh[114][[ 1 == 0 ]] -+ exglobal_atmos_products.sh[117][[ 24 -eq 24 ]] -+ exglobal_atmos_products.sh[118]last=349 -+ exglobal_atmos_products.sh[122]wgrib2 tmpfileb_f026 -for 323:349 -grib tmpfileb_f026_24 -323:5113825:d=2021032312:HGT:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -324:5131510:d=2021032312:PRES:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -325:5149233:d=2021032312:VWSH:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -326:5160581:d=2021032312:UGRD:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -327:5171867:d=2021032312:VGRD:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -328:5183149:d=2021032312:TMP:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -329:5198966:d=2021032312:HGT:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -330:5217615:d=2021032312:PRES:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -331:5236221:d=2021032312:VWSH:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -332:5248484:d=2021032312:UGRD:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -333:5259796:d=2021032312:VGRD:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -334:5271118:d=2021032312:TMP:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -335:5286926:d=2021032312:HGT:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -336:5305438:d=2021032312:PRES:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -337:5323942:d=2021032312:VWSH:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -338:5336293:d=2021032312:UGRD:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -339:5347272:d=2021032312:VGRD:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -340:5358265:d=2021032312:TMP:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -341:5373214:d=2021032312:HGT:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -342:5391140:d=2021032312:PRES:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -343:5408839:d=2021032312:VWSH:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -344:5420945:d=2021032312:UGRD:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -345:5432002:d=2021032312:VGRD:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -346:5443035:d=2021032312:TMP:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -347:5458112:d=2021032312:HGT:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -348:5475950:d=2021032312:PRES:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -349:5493592:d=2021032312:VWSH:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -+ exglobal_atmos_products.sh[122]true -+ exglobal_atmos_products.sh[123]export err=0 -+ exglobal_atmos_products.sh[123]err=0 -+ exglobal_atmos_products.sh[124][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[127]input_file=tmpfileb_f026_24 -+ exglobal_atmos_products.sh[128]output_file_prefix=pgb2bfile_f026_24 -+ exglobal_atmos_products.sh[129]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_24 pgb2bfile_f026_24 0p25' -+ exglobal_atmos_products.sh[133][[ 349 -eq 349 ]] -+ exglobal_atmos_products.sh[134](( pproc = iproc+1 )) -+ exglobal_atmos_products.sh[134](( pproc < nproc )) -+ exglobal_atmos_products.sh[137]break -+ exglobal_atmos_products.sh[142]/work2/noaa/global/mterry/global-workflow_forked/ush/run_mpmd.sh /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601/poescript -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ run_mpmd.sh[31]source /work2/noaa/global/mterry/global-workflow_forked/ush/preamble.sh -++ preamble.sh[20]set +x -Begin run_mpmd.sh at Tue Jul 29 02:28:31 UTC 2025 -+ run_mpmd.sh[33]cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601/poescript -+ run_mpmd.sh[36][[ YES != \Y\E\S ]] -+ run_mpmd.sh[46]export OMP_NUM_THREADS=1 -+ run_mpmd.sh[46]OMP_NUM_THREADS=1 -++ run_mpmd.sh[49]wc -l -+ run_mpmd.sh[49]nprocs=24 -+ run_mpmd.sh[52]mpmd_cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601/mpmd_cmdfile -+ run_mpmd.sh[53][[ -s /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601/mpmd_cmdfile ]] -+ run_mpmd.sh[55]cat - INFO: Executing MPMD job, STDOUT redirected for each process separately - INFO: On failure, logs for each job will be available in /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601/mpmd.proc_num.out - INFO: The proc_num corresponds to the line in '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601/mpmd_cmdfile' -+ run_mpmd.sh[61][[ srun -l --export=ALL --hint=nomultithread =~ ^srun.* ]] -+ run_mpmd.sh[65]nm=0 -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '0 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_1 pgb2bfile_f026_1 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '1 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_2 pgb2bfile_f026_2 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '2 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_3 pgb2bfile_f026_3 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '3 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_4 pgb2bfile_f026_4 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '4 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_5 pgb2bfile_f026_5 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '5 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_6 pgb2bfile_f026_6 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '6 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_7 pgb2bfile_f026_7 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '7 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_8 pgb2bfile_f026_8 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '8 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_9 pgb2bfile_f026_9 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '9 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_10 pgb2bfile_f026_10 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '10 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_11 pgb2bfile_f026_11 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '11 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_12 pgb2bfile_f026_12 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '12 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_13 pgb2bfile_f026_13 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '13 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_14 pgb2bfile_f026_14 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '14 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_15 pgb2bfile_f026_15 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '15 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_16 pgb2bfile_f026_16 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '16 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_17 pgb2bfile_f026_17 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '17 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_18 pgb2bfile_f026_18 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '18 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_19 pgb2bfile_f026_19 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '19 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_20 pgb2bfile_f026_20 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '20 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_21 pgb2bfile_f026_21 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '21 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_22 pgb2bfile_f026_22 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '22 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_23 pgb2bfile_f026_23 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '23 /work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_master.sh tmpfileb_f026_24 pgb2bfile_f026_24 0p25' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[72]set +e -+ run_mpmd.sh[74]srun -l --export=ALL --hint=nomultithread --multi-prog --output=mpmd.%j.%t.out -n 24 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601/mpmd_cmdfile -+ run_mpmd.sh[75]err=0 -+ run_mpmd.sh[76]set_strict -+ preamble.sh[35][[ YES == \Y\E\S ]] -+ preamble.sh[37]set -eu -+ run_mpmd.sh[103][[ 0 -eq 0 ]] -+ run_mpmd.sh[104]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601/mpmd_cmdfile -++ run_mpmd.sh[105]find . -name 'mpmd.*.out' -+ run_mpmd.sh[105]out_files='./mpmd.5951061.17.out -./mpmd.5951061.0.out -./mpmd.5951061.6.out -./mpmd.5951061.15.out -./mpmd.5951061.9.out -./mpmd.5951061.2.out -./mpmd.5951061.3.out -./mpmd.5951061.23.out -./mpmd.5951061.4.out -./mpmd.5951061.22.out -./mpmd.5951061.1.out -./mpmd.5951061.7.out -./mpmd.5951061.21.out -./mpmd.5951061.12.out -./mpmd.5951061.19.out -./mpmd.5951061.8.out -./mpmd.5951061.16.out -./mpmd.5951061.18.out -./mpmd.5951061.13.out -./mpmd.5951061.10.out -./mpmd.5951061.20.out -./mpmd.5951061.14.out -./mpmd.5951061.5.out -./mpmd.5951061.11.out' -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.17.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.17.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.0.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.0.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.6.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.6.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.15.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.15.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.9.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.9.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.2.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.2.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.3.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.3.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.23.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.23.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.4.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.4.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.22.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.22.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.1.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.1.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.7.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.7.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.21.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.21.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.12.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.12.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.19.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.19.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.8.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.8.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.16.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.16.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.18.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.18.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.13.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.13.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.10.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.10.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.20.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.20.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.14.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.14.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.5.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.5.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951061.11.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951061.11.out -+ run_mpmd.sh[110]cat mpmd.out -17: + bash[8]'[' -z '' ']' -17: + bash[9]case "$-" in -17: + bash[12]__lmod_vx=x -17: + bash[16]'[' -n x ']' -17: + bash[16]set +x -17: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -17: Shell debugging restarted -17: + bash[224]unset __lmod_vx -17: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_18 -17: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_18 -17: + interp_atmos_master.sh[9]grid_string=0p25 -17: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -17: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -17: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -17: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -17: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -17: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -17: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -17: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -17: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -17: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -17: + interp_atmos_master.sh[31]IFS=: -17: + interp_atmos_master.sh[31]read -ra grids -17: + interp_atmos_master.sh[33]output_grids= -17: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -17: + interp_atmos_master.sh[35]gridopt=grid0p25 -17: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_18_0p25' -17: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_18 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_18_0p25 -17: 1:0:d=2021032312:TMP:825 mb:26 hour fcst: -17: 2:20643:d=2021032312:RH:825 mb:26 hour fcst: -17: 3:41029:d=2021032312:TCDC:825 mb:26 hour fcst: -17: 4:55115:d=2021032312:VVEL:825 mb:26 hour fcst: -17: 5:80762:d=2021032312:DZDT:825 mb:26 hour fcst: -17: 6:107673:d=2021032312:UGRD:825 mb:26 hour fcst: -17: 7:128902:d=2021032312:VGRD:825 mb:26 hour fcst: -17: 8:150757:d=2021032312:ABSV:825 mb:26 hour fcst: -17: 9:172012:d=2021032312:CLMR:825 mb:26 hour fcst: -17: 10:184678:d=2021032312:ICMR:825 mb:26 hour fcst: -17: 11:195010:d=2021032312:RWMR:825 mb:26 hour fcst: -17: 12:204491:d=2021032312:SNMR:825 mb:26 hour fcst: -17: 13:216757:d=2021032312:GRLE:825 mb:26 hour fcst: -17: 14:219188:d=2021032312:HGT:875 mb:26 hour fcst: -17: + interp_atmos_master.sh[47]export err=0 -17: + interp_atmos_master.sh[47]err=0 -17: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -17: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -17: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_18_0p25 -17: + product_functions.sh[5]local filename=pgb2bfile_f026_18_0p25 -17: + product_functions.sh[6]wgrib2 pgb2bfile_f026_18_0p25 -not_if :RH: -grib pgb2bfile_f026_18_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_18_0p25.new -17: 1:0:d=2021032312:TMP:825 mb:26 hour fcst: -17: 2:488694:d=2021032312:RH:825 mb:26 hour fcst: -17: 3:968646:d=2021032312:TCDC:825 mb:26 hour fcst: -17: 4:1254829:d=2021032312:VVEL:825 mb:26 hour fcst: -17: 5:1937942:d=2021032312:DZDT:825 mb:26 hour fcst: -17: 6:2677612:d=2021032312:UGRD:825 mb:26 hour fcst: -17: 7:3188278:d=2021032312:VGRD:825 mb:26 hour fcst: -17: 8:3728067:d=2021032312:ABSV:825 mb:26 hour fcst: -17: 9:4236466:d=2021032312:CLMR:825 mb:26 hour fcst: -17: 10:4508591:d=2021032312:ICMR:825 mb:26 hour fcst: -17: 11:4753156:d=2021032312:RWMR:825 mb:26 hour fcst: -17: 12:4932902:d=2021032312:SNMR:825 mb:26 hour fcst: -17: 13:5208332:d=2021032312:GRLE:825 mb:26 hour fcst: -17: 14:5252261:d=2021032312:HGT:875 mb:26 hour fcst: -17: + product_functions.sh[10]rc=0 -17: + product_functions.sh[11](( rc == 0 )) -17: + product_functions.sh[11]mv pgb2bfile_f026_18_0p25.new pgb2bfile_f026_18_0p25 -17: + product_functions.sh[12]return 0 -17: + interp_atmos_master.sh[56]export err=0 -17: + interp_atmos_master.sh[56]err=0 -17: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -17: ++ interp_atmos_master.sh[62]wc -l -17: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_18_0p25 -match 'LAND|ICEC' -17: + interp_atmos_master.sh[62]var_count=0 -17: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -17: + interp_atmos_master.sh[73]exit 0 - 0: + bash[8]'[' -z '' ']' - 0: + bash[9]case "$-" in - 0: + bash[12]__lmod_vx=x - 0: + bash[16]'[' -n x ']' - 0: + bash[16]set +x - 0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 0: Shell debugging restarted - 0: + bash[224]unset __lmod_vx - 0: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_1 - 0: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_1 - 0: + interp_atmos_master.sh[9]grid_string=0p25 - 0: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 0: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 0: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 0: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 0: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 0: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 0: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 0: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 0: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 0: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 0: + interp_atmos_master.sh[31]IFS=: - 0: + interp_atmos_master.sh[31]read -ra grids - 0: + interp_atmos_master.sh[33]output_grids= - 0: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 0: + interp_atmos_master.sh[35]gridopt=grid0p25 - 0: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_1_0p25' - 0: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_1 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_1_0p25 - 0: 1:0:d=2021032312:HGT:1 mb:26 hour fcst: - 0: 2:17949:d=2021032312:TMP:1 mb:26 hour fcst: - 0: 3:33754:d=2021032312:RH:1 mb:26 hour fcst: - 0: 4:43532:d=2021032312:UGRD:1 mb:26 hour fcst: - 0: 5:61987:d=2021032312:VGRD:1 mb:26 hour fcst: - 0: 6:79619:d=2021032312:ABSV:1 mb:26 hour fcst: - 0: 7:95908:d=2021032312:O3MR:1 mb:26 hour fcst: - 0: 8:117173:d=2021032312:HGT:2 mb:26 hour fcst: - 0: 9:137378:d=2021032312:TMP:2 mb:26 hour fcst: - 0: 10:153764:d=2021032312:RH:2 mb:26 hour fcst: - 0: 11:162126:d=2021032312:UGRD:2 mb:26 hour fcst: - 0: 12:173614:d=2021032312:VGRD:2 mb:26 hour fcst: - 0: 13:184318:d=2021032312:ABSV:2 mb:26 hour fcst: - 0: 14:201326:d=2021032312:O3MR:2 mb:26 hour fcst: - 0: + interp_atmos_master.sh[47]export err=0 - 0: + interp_atmos_master.sh[47]err=0 - 0: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 0: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 0: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_1_0p25 - 0: + product_functions.sh[5]local filename=pgb2bfile_f026_1_0p25 - 0: + product_functions.sh[6]wgrib2 pgb2bfile_f026_1_0p25 -not_if :RH: -grib pgb2bfile_f026_1_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_1_0p25.new - 0: 1:0:d=2021032312:HGT:1 mb:26 hour fcst: - 0: 2:405921:d=2021032312:TMP:1 mb:26 hour fcst: - 0: 3:756793:d=2021032312:RH:1 mb:26 hour fcst: - 0: 4:1014566:d=2021032312:UGRD:1 mb:26 hour fcst: - 0: 5:1441339:d=2021032312:VGRD:1 mb:26 hour fcst: - 0: 6:1845945:d=2021032312:ABSV:1 mb:26 hour fcst: - 0: 7:2188167:d=2021032312:O3MR:1 mb:26 hour fcst: - 0: 8:2720941:d=2021032312:HGT:2 mb:26 hour fcst: - 0: 9:3207296:d=2021032312:TMP:2 mb:26 hour fcst: - 0: 10:3574963:d=2021032312:RH:2 mb:26 hour fcst: - 0: 11:3820862:d=2021032312:UGRD:2 mb:26 hour fcst: - 0: 12:4092638:d=2021032312:VGRD:2 mb:26 hour fcst: - 0: 13:4359980:d=2021032312:ABSV:2 mb:26 hour fcst: - 0: 14:4731246:d=2021032312:O3MR:2 mb:26 hour fcst: - 0: + product_functions.sh[10]rc=0 - 0: + product_functions.sh[11](( rc == 0 )) - 0: + product_functions.sh[11]mv pgb2bfile_f026_1_0p25.new pgb2bfile_f026_1_0p25 - 0: + product_functions.sh[12]return 0 - 0: + interp_atmos_master.sh[56]export err=0 - 0: + interp_atmos_master.sh[56]err=0 - 0: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 0: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_1_0p25 -match 'LAND|ICEC' - 0: ++ interp_atmos_master.sh[62]wc -l - 0: + interp_atmos_master.sh[62]var_count=0 - 0: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 0: + interp_atmos_master.sh[73]exit 0 - 6: + bash[8]'[' -z '' ']' - 6: + bash[9]case "$-" in - 6: + bash[12]__lmod_vx=x - 6: + bash[16]'[' -n x ']' - 6: + bash[16]set +x - 6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 6: Shell debugging restarted - 6: + bash[224]unset __lmod_vx - 6: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_7 - 6: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_7 - 6: + interp_atmos_master.sh[9]grid_string=0p25 - 6: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 6: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 6: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 6: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 6: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 6: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 6: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 6: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 6: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 6: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 6: + interp_atmos_master.sh[31]IFS=: - 6: + interp_atmos_master.sh[31]read -ra grids - 6: + interp_atmos_master.sh[33]output_grids= - 6: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 6: + interp_atmos_master.sh[35]gridopt=grid0p25 - 6: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_7_0p25' - 6: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_7 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_7_0p25 - 6: 1:0:d=2021032312:TMP:275 mb:26 hour fcst: - 6: 2:18810:d=2021032312:RH:275 mb:26 hour fcst: - 6: 3:37250:d=2021032312:TCDC:275 mb:26 hour fcst: - 6: 4:49196:d=2021032312:VVEL:275 mb:26 hour fcst: - 6: 5:72507:d=2021032312:DZDT:275 mb:26 hour fcst: - 6: 6:100361:d=2021032312:UGRD:275 mb:26 hour fcst: - 6: 7:114708:d=2021032312:VGRD:275 mb:26 hour fcst: - 6: 8:130097:d=2021032312:ABSV:275 mb:26 hour fcst: - 6: 9:152177:d=2021032312:CLMR:275 mb:26 hour fcst: - 6: 10:152356:d=2021032312:ICMR:275 mb:26 hour fcst: - 6: 11:166958:d=2021032312:RWMR:275 mb:26 hour fcst: - 6: 12:167137:d=2021032312:SNMR:275 mb:26 hour fcst: - 6: 13:177300:d=2021032312:GRLE:275 mb:26 hour fcst: - 6: 14:177593:d=2021032312:HGT:325 mb:26 hour fcst: - 6: + interp_atmos_master.sh[47]export err=0 - 6: + interp_atmos_master.sh[47]err=0 - 6: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 6: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 6: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_7_0p25 - 6: + product_functions.sh[5]local filename=pgb2bfile_f026_7_0p25 - 6: + product_functions.sh[6]wgrib2 pgb2bfile_f026_7_0p25 -not_if :RH: -grib pgb2bfile_f026_7_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_7_0p25.new - 6: 1:0:d=2021032312:TMP:275 mb:26 hour fcst: - 6: 2:416157:d=2021032312:RH:275 mb:26 hour fcst: - 6: 3:834924:d=2021032312:TCDC:275 mb:26 hour fcst: - 6: 4:1064982:d=2021032312:VVEL:275 mb:26 hour fcst: - 6: 5:1680799:d=2021032312:DZDT:275 mb:26 hour fcst: - 6: 6:2458235:d=2021032312:UGRD:275 mb:26 hour fcst: - 6: 7:2774196:d=2021032312:VGRD:275 mb:26 hour fcst: - 6: 8:3106746:d=2021032312:ABSV:275 mb:26 hour fcst: - 6: 9:3649838:d=2021032312:CLMR:275 mb:26 hour fcst: - 6: 10:3650017:d=2021032312:ICMR:275 mb:26 hour fcst: - 6: 11:3968506:d=2021032312:RWMR:275 mb:26 hour fcst: - 6: 12:3968685:d=2021032312:SNMR:275 mb:26 hour fcst: - 6: 13:4169544:d=2021032312:GRLE:275 mb:26 hour fcst: - 6: 14:4171268:d=2021032312:HGT:325 mb:26 hour fcst: - 6: + product_functions.sh[10]rc=0 - 6: + product_functions.sh[11](( rc == 0 )) - 6: + product_functions.sh[11]mv pgb2bfile_f026_7_0p25.new pgb2bfile_f026_7_0p25 - 6: + product_functions.sh[12]return 0 - 6: + interp_atmos_master.sh[56]export err=0 - 6: + interp_atmos_master.sh[56]err=0 - 6: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 6: ++ interp_atmos_master.sh[62]wc -l - 6: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_7_0p25 -match 'LAND|ICEC' - 6: + interp_atmos_master.sh[62]var_count=0 - 6: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 6: + interp_atmos_master.sh[73]exit 0 -15: + bash[8]'[' -z '' ']' -15: + bash[9]case "$-" in -15: + bash[12]__lmod_vx=x -15: + bash[16]'[' -n x ']' -15: + bash[16]set +x -15: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -15: Shell debugging restarted -15: + bash[224]unset __lmod_vx -15: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_16 -15: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_16 -15: + interp_atmos_master.sh[9]grid_string=0p25 -15: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -15: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -15: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -15: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -15: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -15: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -15: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -15: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -15: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -15: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -15: + interp_atmos_master.sh[31]IFS=: -15: + interp_atmos_master.sh[31]read -ra grids -15: + interp_atmos_master.sh[33]output_grids= -15: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -15: + interp_atmos_master.sh[35]gridopt=grid0p25 -15: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_16_0p25' -15: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_16 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_16_0p25 -15: 1:0:d=2021032312:TMP:725 mb:26 hour fcst: -15: 2:19749:d=2021032312:RH:725 mb:26 hour fcst: -15: 3:40063:d=2021032312:TCDC:725 mb:26 hour fcst: -15: 4:52040:d=2021032312:VVEL:725 mb:26 hour fcst: -15: 5:77816:d=2021032312:DZDT:725 mb:26 hour fcst: -15: 6:105128:d=2021032312:UGRD:725 mb:26 hour fcst: -15: 7:125942:d=2021032312:VGRD:725 mb:26 hour fcst: -15: 8:147560:d=2021032312:ABSV:725 mb:26 hour fcst: -15: 9:168637:d=2021032312:CLMR:725 mb:26 hour fcst: -15: 10:176635:d=2021032312:ICMR:725 mb:26 hour fcst: -15: 11:188337:d=2021032312:RWMR:725 mb:26 hour fcst: -15: 12:194071:d=2021032312:SNMR:725 mb:26 hour fcst: -15: 13:207914:d=2021032312:GRLE:725 mb:26 hour fcst: -15: 14:209847:d=2021032312:HGT:775 mb:26 hour fcst: -15: + interp_atmos_master.sh[47]export err=0 -15: + interp_atmos_master.sh[47]err=0 -15: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -15: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -15: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_16_0p25 -15: + product_functions.sh[5]local filename=pgb2bfile_f026_16_0p25 -15: + product_functions.sh[6]wgrib2 pgb2bfile_f026_16_0p25 -not_if :RH: -grib pgb2bfile_f026_16_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_16_0p25.new -15: 1:0:d=2021032312:TMP:725 mb:26 hour fcst: -15: 2:456695:d=2021032312:RH:725 mb:26 hour fcst: -15: 3:931081:d=2021032312:TCDC:725 mb:26 hour fcst: -15: 4:1167628:d=2021032312:VVEL:725 mb:26 hour fcst: -15: 5:1853413:d=2021032312:DZDT:725 mb:26 hour fcst: -15: 6:2606967:d=2021032312:UGRD:725 mb:26 hour fcst: -15: 7:3110341:d=2021032312:VGRD:725 mb:26 hour fcst: -15: 8:3643230:d=2021032312:ABSV:725 mb:26 hour fcst: -15: 9:4142741:d=2021032312:CLMR:725 mb:26 hour fcst: -15: 10:4316175:d=2021032312:ICMR:725 mb:26 hour fcst: -15: 11:4590777:d=2021032312:RWMR:725 mb:26 hour fcst: -15: 12:4702650:d=2021032312:SNMR:725 mb:26 hour fcst: -15: 13:5007077:d=2021032312:GRLE:725 mb:26 hour fcst: -15: 14:5039552:d=2021032312:HGT:775 mb:26 hour fcst: -15: + product_functions.sh[10]rc=0 -15: + product_functions.sh[11](( rc == 0 )) -15: + product_functions.sh[11]mv pgb2bfile_f026_16_0p25.new pgb2bfile_f026_16_0p25 -15: + product_functions.sh[12]return 0 -15: + interp_atmos_master.sh[56]export err=0 -15: + interp_atmos_master.sh[56]err=0 -15: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -15: ++ interp_atmos_master.sh[62]wc -l -15: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_16_0p25 -match 'LAND|ICEC' -15: + interp_atmos_master.sh[62]var_count=0 -15: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -15: + interp_atmos_master.sh[73]exit 0 - 9: + bash[8]'[' -z '' ']' - 9: + bash[9]case "$-" in - 9: + bash[12]__lmod_vx=x - 9: + bash[16]'[' -n x ']' - 9: + bash[16]set +x - 9: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 9: Shell debugging restarted - 9: + bash[224]unset __lmod_vx - 9: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_10 - 9: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_10 - 9: + interp_atmos_master.sh[9]grid_string=0p25 - 9: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 9: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 9: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 9: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 9: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 9: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 9: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 9: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 9: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 9: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 9: + interp_atmos_master.sh[31]IFS=: - 9: + interp_atmos_master.sh[31]read -ra grids - 9: + interp_atmos_master.sh[33]output_grids= - 9: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 9: + interp_atmos_master.sh[35]gridopt=grid0p25 - 9: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_10_0p25' - 9: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_10 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_10_0p25 - 9: 1:0:d=2021032312:TMP:425 mb:26 hour fcst: - 9: 2:18444:d=2021032312:RH:425 mb:26 hour fcst: - 9: 3:37716:d=2021032312:TCDC:425 mb:26 hour fcst: - 9: 4:50586:d=2021032312:VVEL:425 mb:26 hour fcst: - 9: 5:75661:d=2021032312:DZDT:425 mb:26 hour fcst: - 9: 6:103846:d=2021032312:UGRD:425 mb:26 hour fcst: - 9: 7:125700:d=2021032312:VGRD:425 mb:26 hour fcst: - 9: 8:148217:d=2021032312:ABSV:425 mb:26 hour fcst: - 9: 9:169826:d=2021032312:CLMR:425 mb:26 hour fcst: - 9: 10:170480:d=2021032312:ICMR:425 mb:26 hour fcst: - 9: 11:187255:d=2021032312:RWMR:425 mb:26 hour fcst: - 9: 12:187732:d=2021032312:SNMR:425 mb:26 hour fcst: - 9: 13:202275:d=2021032312:GRLE:425 mb:26 hour fcst: - 9: 14:203452:d=2021032312:HGT:475 mb:26 hour fcst: - 9: + interp_atmos_master.sh[47]export err=0 - 9: + interp_atmos_master.sh[47]err=0 - 9: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 9: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 9: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_10_0p25 - 9: + product_functions.sh[5]local filename=pgb2bfile_f026_10_0p25 - 9: + product_functions.sh[6]wgrib2 pgb2bfile_f026_10_0p25 -not_if :RH: -grib pgb2bfile_f026_10_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_10_0p25.new - 9: 1:0:d=2021032312:TMP:425 mb:26 hour fcst: - 9: 2:414394:d=2021032312:RH:425 mb:26 hour fcst: - 9: 3:855143:d=2021032312:TCDC:425 mb:26 hour fcst: - 9: 4:1091110:d=2021032312:VVEL:425 mb:26 hour fcst: - 9: 5:1755418:d=2021032312:DZDT:425 mb:26 hour fcst: - 9: 6:2544457:d=2021032312:UGRD:425 mb:26 hour fcst: - 9: 7:3085708:d=2021032312:VGRD:425 mb:26 hour fcst: - 9: 8:3655510:d=2021032312:ABSV:425 mb:26 hour fcst: - 9: 9:4186118:d=2021032312:CLMR:425 mb:26 hour fcst: - 9: 10:4197081:d=2021032312:ICMR:425 mb:26 hour fcst: - 9: 11:4569532:d=2021032312:RWMR:425 mb:26 hour fcst: - 9: 12:4574587:d=2021032312:SNMR:425 mb:26 hour fcst: - 9: 13:4887038:d=2021032312:GRLE:425 mb:26 hour fcst: - 9: 14:4902789:d=2021032312:HGT:475 mb:26 hour fcst: - 9: + product_functions.sh[10]rc=0 - 9: + product_functions.sh[11](( rc == 0 )) - 9: + product_functions.sh[11]mv pgb2bfile_f026_10_0p25.new pgb2bfile_f026_10_0p25 - 9: + product_functions.sh[12]return 0 - 9: + interp_atmos_master.sh[56]export err=0 - 9: + interp_atmos_master.sh[56]err=0 - 9: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 9: ++ interp_atmos_master.sh[62]wc -l - 9: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_10_0p25 -match 'LAND|ICEC' - 9: + interp_atmos_master.sh[62]var_count=0 - 9: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 9: + interp_atmos_master.sh[73]exit 0 - 2: + bash[8]'[' -z '' ']' - 2: + bash[9]case "$-" in - 2: + bash[12]__lmod_vx=x - 2: + bash[16]'[' -n x ']' - 2: + bash[16]set +x - 2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 2: Shell debugging restarted - 2: + bash[224]unset __lmod_vx - 2: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_3 - 2: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_3 - 2: + interp_atmos_master.sh[9]grid_string=0p25 - 2: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 2: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 2: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 2: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 2: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 2: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 2: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 2: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 2: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 2: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 2: + interp_atmos_master.sh[31]IFS=: - 2: + interp_atmos_master.sh[31]read -ra grids - 2: + interp_atmos_master.sh[33]output_grids= - 2: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 2: + interp_atmos_master.sh[35]gridopt=grid0p25 - 2: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_3_0p25' - 2: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_3 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_3_0p25 - 2: 1:0:d=2021032312:HGT:7 mb:26 hour fcst: - 2: 2:17658:d=2021032312:TMP:7 mb:26 hour fcst: - 2: 3:33800:d=2021032312:RH:7 mb:26 hour fcst: - 2: 4:39860:d=2021032312:UGRD:7 mb:26 hour fcst: - 2: 5:51254:d=2021032312:VGRD:7 mb:26 hour fcst: - 2: 6:70070:d=2021032312:ABSV:7 mb:26 hour fcst: - 2: 7:87801:d=2021032312:O3MR:7 mb:26 hour fcst: - 2: 8:111392:d=2021032312:TCDC:70 mb:26 hour fcst: - 2: 9:111571:d=2021032312:CLMR:70 mb:26 hour fcst: - 2: 10:111750:d=2021032312:ICMR:70 mb:26 hour fcst: - 2: 11:111929:d=2021032312:RWMR:70 mb:26 hour fcst: - 2: 12:112108:d=2021032312:SNMR:70 mb:26 hour fcst: - 2: 13:112287:d=2021032312:GRLE:70 mb:26 hour fcst: - 2: 14:112466:d=2021032312:HGT:125 mb:26 hour fcst: - 2: + interp_atmos_master.sh[47]export err=0 - 2: + interp_atmos_master.sh[47]err=0 - 2: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 2: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 2: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_3_0p25 - 2: + product_functions.sh[5]local filename=pgb2bfile_f026_3_0p25 - 2: + product_functions.sh[6]wgrib2 pgb2bfile_f026_3_0p25 -not_if :RH: -grib pgb2bfile_f026_3_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_3_0p25.new - 2: 1:0:d=2021032312:HGT:7 mb:26 hour fcst: - 2: 2:396734:d=2021032312:TMP:7 mb:26 hour fcst: - 2: 3:747787:d=2021032312:RH:7 mb:26 hour fcst: - 2: 4:921726:d=2021032312:UGRD:7 mb:26 hour fcst: - 2: 5:1364072:d=2021032312:VGRD:7 mb:26 hour fcst: - 2: 6:1795281:d=2021032312:ABSV:7 mb:26 hour fcst: - 2: 7:2183421:d=2021032312:O3MR:7 mb:26 hour fcst: - 2: 8:2786939:d=2021032312:TCDC:70 mb:26 hour fcst: - 2: 9:2787118:d=2021032312:CLMR:70 mb:26 hour fcst: - 2: 10:2787297:d=2021032312:ICMR:70 mb:26 hour fcst: - 2: 11:2787476:d=2021032312:RWMR:70 mb:26 hour fcst: - 2: 12:2787655:d=2021032312:SNMR:70 mb:26 hour fcst: - 2: 13:2787834:d=2021032312:GRLE:70 mb:26 hour fcst: - 2: 14:2788013:d=2021032312:HGT:125 mb:26 hour fcst: - 2: + product_functions.sh[10]rc=0 - 2: + product_functions.sh[11](( rc == 0 )) - 2: + product_functions.sh[11]mv pgb2bfile_f026_3_0p25.new pgb2bfile_f026_3_0p25 - 2: + product_functions.sh[12]return 0 - 2: + interp_atmos_master.sh[56]export err=0 - 2: + interp_atmos_master.sh[56]err=0 - 2: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 2: ++ interp_atmos_master.sh[62]wc -l - 2: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_3_0p25 -match 'LAND|ICEC' - 2: + interp_atmos_master.sh[62]var_count=0 - 2: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 2: + interp_atmos_master.sh[73]exit 0 - 3: + bash[8]'[' -z '' ']' - 3: + bash[9]case "$-" in - 3: + bash[12]__lmod_vx=x - 3: + bash[16]'[' -n x ']' - 3: + bash[16]set +x - 3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 3: Shell debugging restarted - 3: + bash[224]unset __lmod_vx - 3: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_4 - 3: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_4 - 3: + interp_atmos_master.sh[9]grid_string=0p25 - 3: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 3: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 3: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 3: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 3: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 3: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 3: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 3: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 3: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 3: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 3: + interp_atmos_master.sh[31]IFS=: - 3: + interp_atmos_master.sh[31]read -ra grids - 3: + interp_atmos_master.sh[33]output_grids= - 3: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 3: + interp_atmos_master.sh[35]gridopt=grid0p25 - 3: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_4_0p25' - 3: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_4 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_4_0p25 - 3: 1:0:d=2021032312:TMP:125 mb:26 hour fcst: - 3: 2:16937:d=2021032312:RH:125 mb:26 hour fcst: - 3: 3:27417:d=2021032312:TCDC:125 mb:26 hour fcst: - 3: 4:30011:d=2021032312:VVEL:125 mb:26 hour fcst: - 3: 5:56101:d=2021032312:DZDT:125 mb:26 hour fcst: - 3: 6:81178:d=2021032312:UGRD:125 mb:26 hour fcst: - 3: 7:100834:d=2021032312:VGRD:125 mb:26 hour fcst: - 3: 8:120477:d=2021032312:ABSV:125 mb:26 hour fcst: - 3: 9:138914:d=2021032312:CLMR:125 mb:26 hour fcst: - 3: 10:139093:d=2021032312:ICMR:125 mb:26 hour fcst: - 3: 11:143821:d=2021032312:RWMR:125 mb:26 hour fcst: - 3: 12:144000:d=2021032312:SNMR:125 mb:26 hour fcst: - 3: 13:148175:d=2021032312:GRLE:125 mb:26 hour fcst: - 3: 14:148585:d=2021032312:HGT:175 mb:26 hour fcst: - 3: + interp_atmos_master.sh[47]export err=0 - 3: + interp_atmos_master.sh[47]err=0 - 3: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 3: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 3: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_4_0p25 - 3: + product_functions.sh[5]local filename=pgb2bfile_f026_4_0p25 - 3: + product_functions.sh[6]wgrib2 pgb2bfile_f026_4_0p25 -not_if :RH: -grib pgb2bfile_f026_4_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_4_0p25.new - 3: 1:0:d=2021032312:TMP:125 mb:26 hour fcst: - 3: 2:377291:d=2021032312:RH:125 mb:26 hour fcst: - 3: 3:624032:d=2021032312:TCDC:125 mb:26 hour fcst: - 3: 4:677349:d=2021032312:VVEL:125 mb:26 hour fcst: - 3: 5:1379047:d=2021032312:DZDT:125 mb:26 hour fcst: - 3: 6:2061862:d=2021032312:UGRD:125 mb:26 hour fcst: - 3: 7:2545057:d=2021032312:VGRD:125 mb:26 hour fcst: - 3: 8:3024420:d=2021032312:ABSV:125 mb:26 hour fcst: - 3: 9:3443625:d=2021032312:CLMR:125 mb:26 hour fcst: - 3: 10:3443804:d=2021032312:ICMR:125 mb:26 hour fcst: - 3: 11:3552557:d=2021032312:RWMR:125 mb:26 hour fcst: - 3: 12:3552736:d=2021032312:SNMR:125 mb:26 hour fcst: - 3: 13:3634686:d=2021032312:GRLE:125 mb:26 hour fcst: - 3: 14:3639060:d=2021032312:HGT:175 mb:26 hour fcst: - 3: + product_functions.sh[10]rc=0 - 3: + product_functions.sh[11](( rc == 0 )) - 3: + product_functions.sh[11]mv pgb2bfile_f026_4_0p25.new pgb2bfile_f026_4_0p25 - 3: + product_functions.sh[12]return 0 - 3: + interp_atmos_master.sh[56]export err=0 - 3: + interp_atmos_master.sh[56]err=0 - 3: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 3: ++ interp_atmos_master.sh[62]wc -l - 3: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_4_0p25 -match 'LAND|ICEC' - 3: + interp_atmos_master.sh[62]var_count=0 - 3: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 3: + interp_atmos_master.sh[73]exit 0 -23: + bash[8]'[' -z '' ']' -23: + bash[9]case "$-" in -23: + bash[12]__lmod_vx=x -23: + bash[16]'[' -n x ']' -23: + bash[16]set +x -23: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -23: Shell debugging restarted -23: + bash[224]unset __lmod_vx -23: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_24 -23: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_24 -23: + interp_atmos_master.sh[9]grid_string=0p25 -23: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -23: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -23: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -23: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -23: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -23: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -23: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -23: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -23: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -23: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -23: + interp_atmos_master.sh[31]IFS=: -23: + interp_atmos_master.sh[31]read -ra grids -23: + interp_atmos_master.sh[33]output_grids= -23: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -23: + interp_atmos_master.sh[35]gridopt=grid0p25 -23: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_24_0p25' -23: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_24 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_24_0p25 -23: 1:0:d=2021032312:HGT:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -23: 2:17685:d=2021032312:PRES:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -23: 3:35408:d=2021032312:VWSH:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -23: 4:46756:d=2021032312:UGRD:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 5:58042:d=2021032312:VGRD:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 6:69324:d=2021032312:TMP:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 7:85141:d=2021032312:HGT:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 8:103790:d=2021032312:PRES:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 9:122396:d=2021032312:VWSH:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 10:134659:d=2021032312:UGRD:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 11:145971:d=2021032312:VGRD:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 12:157293:d=2021032312:TMP:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 13:173101:d=2021032312:HGT:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 14:191613:d=2021032312:PRES:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 15:210117:d=2021032312:VWSH:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 16:222468:d=2021032312:UGRD:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 17:233447:d=2021032312:VGRD:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 18:244440:d=2021032312:TMP:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 19:259389:d=2021032312:HGT:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 20:277315:d=2021032312:PRES:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 21:295014:d=2021032312:VWSH:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 22:307120:d=2021032312:UGRD:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 23:318177:d=2021032312:VGRD:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 24:329210:d=2021032312:TMP:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 25:344287:d=2021032312:HGT:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 26:362125:d=2021032312:PRES:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 27:379767:d=2021032312:VWSH:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: + interp_atmos_master.sh[47]export err=0 -23: + interp_atmos_master.sh[47]err=0 -23: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -23: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -23: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_24_0p25 -23: + product_functions.sh[5]local filename=pgb2bfile_f026_24_0p25 -23: + product_functions.sh[6]wgrib2 pgb2bfile_f026_24_0p25 -not_if :RH: -grib pgb2bfile_f026_24_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_24_0p25.new -23: 1:0:d=2021032312:HGT:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -23: 2:574597:d=2021032312:PRES:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -23: 3:1152408:d=2021032312:VWSH:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -23: 4:1494754:d=2021032312:UGRD:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 5:1822321:d=2021032312:VGRD:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 6:2148173:d=2021032312:TMP:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 7:2621515:d=2021032312:HGT:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 8:3206689:d=2021032312:PRES:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 9:3797280:d=2021032312:VWSH:PV=1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 10:4157028:d=2021032312:UGRD:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 11:4484743:d=2021032312:VGRD:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 12:4808629:d=2021032312:TMP:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 13:5283508:d=2021032312:HGT:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 14:5871330:d=2021032312:PRES:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 15:6442782:d=2021032312:VWSH:PV=-1e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 16:6800963:d=2021032312:UGRD:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 17:7121824:d=2021032312:VGRD:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 18:7439835:d=2021032312:TMP:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 19:7889142:d=2021032312:HGT:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 20:8455652:d=2021032312:PRES:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 21:8998837:d=2021032312:VWSH:PV=1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 22:9352977:d=2021032312:UGRD:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 23:9676335:d=2021032312:VGRD:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 24:9994945:d=2021032312:TMP:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 25:10443107:d=2021032312:HGT:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 26:11012136:d=2021032312:PRES:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: 27:11554880:d=2021032312:VWSH:PV=-1.5e-06 (Km^2/kg/s) surface:26 hour fcst: -23: + product_functions.sh[10]rc=0 -23: + product_functions.sh[11](( rc == 0 )) -23: + product_functions.sh[11]mv pgb2bfile_f026_24_0p25.new pgb2bfile_f026_24_0p25 -23: + product_functions.sh[12]return 0 -23: + interp_atmos_master.sh[56]export err=0 -23: + interp_atmos_master.sh[56]err=0 -23: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -23: ++ interp_atmos_master.sh[62]wc -l -23: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_24_0p25 -match 'LAND|ICEC' -23: + interp_atmos_master.sh[62]var_count=0 -23: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -23: + interp_atmos_master.sh[73]exit 0 - 4: + bash[8]'[' -z '' ']' - 4: + bash[9]case "$-" in - 4: + bash[12]__lmod_vx=x - 4: + bash[16]'[' -n x ']' - 4: + bash[16]set +x - 4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 4: Shell debugging restarted - 4: + bash[224]unset __lmod_vx - 4: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_5 - 4: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_5 - 4: + interp_atmos_master.sh[9]grid_string=0p25 - 4: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 4: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 4: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 4: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 4: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 4: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 4: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 4: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 4: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 4: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 4: + interp_atmos_master.sh[31]IFS=: - 4: + interp_atmos_master.sh[31]read -ra grids - 4: + interp_atmos_master.sh[33]output_grids= - 4: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 4: + interp_atmos_master.sh[35]gridopt=grid0p25 - 4: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_5_0p25' - 4: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_5 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_5_0p25 - 4: 1:0:d=2021032312:TMP:175 mb:26 hour fcst: - 4: 2:17546:d=2021032312:RH:175 mb:26 hour fcst: - 4: 3:30719:d=2021032312:TCDC:175 mb:26 hour fcst: - 4: 4:35596:d=2021032312:VVEL:175 mb:26 hour fcst: - 4: 5:63970:d=2021032312:DZDT:175 mb:26 hour fcst: - 4: 6:90026:d=2021032312:UGRD:175 mb:26 hour fcst: - 4: 7:103025:d=2021032312:VGRD:175 mb:26 hour fcst: - 4: 8:123733:d=2021032312:ABSV:175 mb:26 hour fcst: - 4: 9:143140:d=2021032312:CLMR:175 mb:26 hour fcst: - 4: 10:143319:d=2021032312:ICMR:175 mb:26 hour fcst: - 4: 11:150004:d=2021032312:RWMR:175 mb:26 hour fcst: - 4: 12:150183:d=2021032312:SNMR:175 mb:26 hour fcst: - 4: 13:154557:d=2021032312:GRLE:175 mb:26 hour fcst: - 4: 14:154975:d=2021032312:HGT:225 mb:26 hour fcst: - 4: + interp_atmos_master.sh[47]export err=0 - 4: + interp_atmos_master.sh[47]err=0 - 4: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 4: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 4: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_5_0p25 - 4: + product_functions.sh[5]local filename=pgb2bfile_f026_5_0p25 - 4: + product_functions.sh[6]wgrib2 pgb2bfile_f026_5_0p25 -not_if :RH: -grib pgb2bfile_f026_5_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_5_0p25.new - 4: 1:0:d=2021032312:TMP:175 mb:26 hour fcst: - 4: 2:396250:d=2021032312:RH:175 mb:26 hour fcst: - 4: 3:697173:d=2021032312:TCDC:175 mb:26 hour fcst: - 4: 4:795792:d=2021032312:VVEL:175 mb:26 hour fcst: - 4: 5:1576869:d=2021032312:DZDT:175 mb:26 hour fcst: - 4: 6:2292498:d=2021032312:UGRD:175 mb:26 hour fcst: - 4: 7:2808888:d=2021032312:VGRD:175 mb:26 hour fcst: - 4: 8:3327163:d=2021032312:ABSV:175 mb:26 hour fcst: - 4: 9:3774904:d=2021032312:CLMR:175 mb:26 hour fcst: - 4: 10:3775083:d=2021032312:ICMR:175 mb:26 hour fcst: - 4: 11:3926426:d=2021032312:RWMR:175 mb:26 hour fcst: - 4: 12:3926605:d=2021032312:SNMR:175 mb:26 hour fcst: - 4: 13:4007397:d=2021032312:GRLE:175 mb:26 hour fcst: - 4: 14:4012288:d=2021032312:HGT:225 mb:26 hour fcst: - 4: + product_functions.sh[10]rc=0 - 4: + product_functions.sh[11](( rc == 0 )) - 4: + product_functions.sh[11]mv pgb2bfile_f026_5_0p25.new pgb2bfile_f026_5_0p25 - 4: + product_functions.sh[12]return 0 - 4: + interp_atmos_master.sh[56]export err=0 - 4: + interp_atmos_master.sh[56]err=0 - 4: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 4: ++ interp_atmos_master.sh[62]wc -l - 4: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_5_0p25 -match 'LAND|ICEC' - 4: + interp_atmos_master.sh[62]var_count=0 - 4: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 4: + interp_atmos_master.sh[73]exit 0 -22: + bash[8]'[' -z '' ']' -22: + bash[9]case "$-" in -22: + bash[12]__lmod_vx=x -22: + bash[16]'[' -n x ']' -22: + bash[16]set +x -22: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -22: Shell debugging restarted -22: + bash[224]unset __lmod_vx -22: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_23 -22: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_23 -22: + interp_atmos_master.sh[9]grid_string=0p25 -22: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -22: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -22: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -22: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -22: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -22: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -22: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -22: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -22: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -22: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -22: + interp_atmos_master.sh[31]IFS=: -22: + interp_atmos_master.sh[31]read -ra grids -22: + interp_atmos_master.sh[33]output_grids= -22: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -22: + interp_atmos_master.sh[35]gridopt=grid0p25 -22: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_23_0p25' -22: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_23 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_23_0p25 -22: 1:0:d=2021032312:TMP:180-150 mb above ground:26 hour fcst: -22: 2:21145:d=2021032312:RH:180-150 mb above ground:26 hour fcst: -22: 3:41151:d=2021032312:SPFH:180-150 mb above ground:26 hour fcst: -22: 4:70562:d=2021032312:UGRD:180-150 mb above ground:26 hour fcst: -22: 5:91616:d=2021032312:VGRD:180-150 mb above ground:26 hour fcst: -22: 6:113290:d=2021032312:UGRD:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -22: 7:123945:d=2021032312:VGRD:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -22: 8:134489:d=2021032312:TMP:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -22: 9:149157:d=2021032312:HGT:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -22: 10:165899:d=2021032312:PRES:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -22: 11:182785:d=2021032312:VWSH:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -22: 12:193782:d=2021032312:UGRD:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -22: 13:204845:d=2021032312:VGRD:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -22: 14:219296:d=2021032312:TMP:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -22: + interp_atmos_master.sh[47]export err=0 -22: + interp_atmos_master.sh[47]err=0 -22: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -22: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -22: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_23_0p25 -22: + product_functions.sh[5]local filename=pgb2bfile_f026_23_0p25 -22: + product_functions.sh[6]wgrib2 pgb2bfile_f026_23_0p25 -not_if :RH: -grib pgb2bfile_f026_23_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_23_0p25.new -22: 1:0:d=2021032312:TMP:180-150 mb above ground:26 hour fcst: -22: 2:507653:d=2021032312:RH:180-150 mb above ground:26 hour fcst: -22: 3:962447:d=2021032312:SPFH:180-150 mb above ground:26 hour fcst: -22: 4:1757924:d=2021032312:UGRD:180-150 mb above ground:26 hour fcst: -22: 5:2260720:d=2021032312:VGRD:180-150 mb above ground:26 hour fcst: -22: 6:2790432:d=2021032312:UGRD:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -22: 7:3104840:d=2021032312:VGRD:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -22: 8:3412471:d=2021032312:TMP:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -22: 9:3877187:d=2021032312:HGT:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -22: 10:4422559:d=2021032312:PRES:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -22: 11:4976764:d=2021032312:VWSH:PV=5e-07 (Km^2/kg/s) surface:26 hour fcst: -22: 12:5311395:d=2021032312:UGRD:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -22: 13:5775240:d=2021032312:VGRD:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -22: 14:6233535:d=2021032312:TMP:PV=-5e-07 (Km^2/kg/s) surface:26 hour fcst: -22: + product_functions.sh[10]rc=0 -22: + product_functions.sh[11](( rc == 0 )) -22: + product_functions.sh[11]mv pgb2bfile_f026_23_0p25.new pgb2bfile_f026_23_0p25 -22: + product_functions.sh[12]return 0 -22: + interp_atmos_master.sh[56]export err=0 -22: + interp_atmos_master.sh[56]err=0 -22: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -22: ++ interp_atmos_master.sh[62]wc -l -22: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_23_0p25 -match 'LAND|ICEC' -22: + interp_atmos_master.sh[62]var_count=0 -22: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -22: + interp_atmos_master.sh[73]exit 0 - 1: + bash[8]'[' -z '' ']' - 1: + bash[9]case "$-" in - 1: + bash[12]__lmod_vx=x - 1: + bash[16]'[' -n x ']' - 1: + bash[16]set +x - 1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 1: Shell debugging restarted - 1: + bash[224]unset __lmod_vx - 1: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_2 - 1: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_2 - 1: + interp_atmos_master.sh[9]grid_string=0p25 - 1: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 1: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 1: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 1: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 1: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 1: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 1: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 1: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 1: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 1: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 1: + interp_atmos_master.sh[31]IFS=: - 1: + interp_atmos_master.sh[31]read -ra grids - 1: + interp_atmos_master.sh[33]output_grids= - 1: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 1: + interp_atmos_master.sh[35]gridopt=grid0p25 - 1: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_2_0p25' - 1: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_2 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_2_0p25 - 1: 1:0:d=2021032312:HGT:3 mb:26 hour fcst: - 1: 2:19941:d=2021032312:TMP:3 mb:26 hour fcst: - 1: 3:36351:d=2021032312:RH:3 mb:26 hour fcst: - 1: 4:42413:d=2021032312:UGRD:3 mb:26 hour fcst: - 1: 5:53944:d=2021032312:VGRD:3 mb:26 hour fcst: - 1: 6:71906:d=2021032312:ABSV:3 mb:26 hour fcst: - 1: 7:88799:d=2021032312:O3MR:3 mb:26 hour fcst: - 1: 8:112131:d=2021032312:HGT:5 mb:26 hour fcst: - 1: 9:129870:d=2021032312:TMP:5 mb:26 hour fcst: - 1: 10:146504:d=2021032312:RH:5 mb:26 hour fcst: - 1: 11:155681:d=2021032312:UGRD:5 mb:26 hour fcst: - 1: 12:167180:d=2021032312:VGRD:5 mb:26 hour fcst: - 1: 13:186035:d=2021032312:ABSV:5 mb:26 hour fcst: - 1: 14:203647:d=2021032312:O3MR:5 mb:26 hour fcst: - 1: + interp_atmos_master.sh[47]export err=0 - 1: + interp_atmos_master.sh[47]err=0 - 1: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 1: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 1: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_2_0p25 - 1: + product_functions.sh[5]local filename=pgb2bfile_f026_2_0p25 - 1: + product_functions.sh[6]wgrib2 pgb2bfile_f026_2_0p25 -not_if :RH: -grib pgb2bfile_f026_2_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_2_0p25.new - 1: 1:0:d=2021032312:HGT:3 mb:26 hour fcst: - 1: 2:483724:d=2021032312:TMP:3 mb:26 hour fcst: - 1: 3:846925:d=2021032312:RH:3 mb:26 hour fcst: - 1: 4:1016229:d=2021032312:UGRD:3 mb:26 hour fcst: - 1: 5:1466780:d=2021032312:VGRD:3 mb:26 hour fcst: - 1: 6:1878734:d=2021032312:ABSV:3 mb:26 hour fcst: - 1: 7:2243315:d=2021032312:O3MR:3 mb:26 hour fcst: - 1: 8:2843710:d=2021032312:HGT:5 mb:26 hour fcst: - 1: 9:3239714:d=2021032312:TMP:5 mb:26 hour fcst: - 1: 10:3605136:d=2021032312:RH:5 mb:26 hour fcst: - 1: 11:3856123:d=2021032312:UGRD:5 mb:26 hour fcst: - 1: 12:4304424:d=2021032312:VGRD:5 mb:26 hour fcst: - 1: 13:4731303:d=2021032312:ABSV:5 mb:26 hour fcst: - 1: 14:5116722:d=2021032312:O3MR:5 mb:26 hour fcst: - 1: + product_functions.sh[10]rc=0 - 1: + product_functions.sh[11](( rc == 0 )) - 1: + product_functions.sh[11]mv pgb2bfile_f026_2_0p25.new pgb2bfile_f026_2_0p25 - 1: + product_functions.sh[12]return 0 - 1: + interp_atmos_master.sh[56]export err=0 - 1: + interp_atmos_master.sh[56]err=0 - 1: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 1: ++ interp_atmos_master.sh[62]wc -l - 1: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_2_0p25 -match 'LAND|ICEC' - 1: + interp_atmos_master.sh[62]var_count=0 - 1: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 1: + interp_atmos_master.sh[73]exit 0 - 7: + bash[8]'[' -z '' ']' - 7: + bash[9]case "$-" in - 7: + bash[12]__lmod_vx=x - 7: + bash[16]'[' -n x ']' - 7: + bash[16]set +x - 7: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 7: Shell debugging restarted - 7: + bash[224]unset __lmod_vx - 7: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_8 - 7: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_8 - 7: + interp_atmos_master.sh[9]grid_string=0p25 - 7: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 7: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 7: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 7: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 7: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 7: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 7: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 7: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 7: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 7: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 7: + interp_atmos_master.sh[31]IFS=: - 7: + interp_atmos_master.sh[31]read -ra grids - 7: + interp_atmos_master.sh[33]output_grids= - 7: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 7: + interp_atmos_master.sh[35]gridopt=grid0p25 - 7: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_8_0p25' - 7: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_8 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_8_0p25 - 7: 1:0:d=2021032312:TMP:325 mb:26 hour fcst: - 7: 2:18342:d=2021032312:RH:325 mb:26 hour fcst: - 7: 3:37279:d=2021032312:TCDC:325 mb:26 hour fcst: - 7: 4:50422:d=2021032312:VVEL:325 mb:26 hour fcst: - 7: 5:74865:d=2021032312:DZDT:325 mb:26 hour fcst: - 7: 6:103076:d=2021032312:UGRD:325 mb:26 hour fcst: - 7: 7:117727:d=2021032312:VGRD:325 mb:26 hour fcst: - 7: 8:132991:d=2021032312:ABSV:325 mb:26 hour fcst: - 7: 9:155256:d=2021032312:CLMR:325 mb:26 hour fcst: - 7: 10:155435:d=2021032312:ICMR:325 mb:26 hour fcst: - 7: 11:172073:d=2021032312:RWMR:325 mb:26 hour fcst: - 7: 12:172252:d=2021032312:SNMR:325 mb:26 hour fcst: - 7: 13:184213:d=2021032312:GRLE:325 mb:26 hour fcst: - 7: 14:184628:d=2021032312:HGT:375 mb:26 hour fcst: - 7: + interp_atmos_master.sh[47]export err=0 - 7: + interp_atmos_master.sh[47]err=0 - 7: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 7: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 7: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_8_0p25 - 7: + product_functions.sh[5]local filename=pgb2bfile_f026_8_0p25 - 7: + product_functions.sh[6]wgrib2 pgb2bfile_f026_8_0p25 -not_if :RH: -grib pgb2bfile_f026_8_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_8_0p25.new - 7: 1:0:d=2021032312:TMP:325 mb:26 hour fcst: - 7: 2:407206:d=2021032312:RH:325 mb:26 hour fcst: - 7: 3:833278:d=2021032312:TCDC:325 mb:26 hour fcst: - 7: 4:1076809:d=2021032312:VVEL:325 mb:26 hour fcst: - 7: 5:1720692:d=2021032312:DZDT:325 mb:26 hour fcst: - 7: 6:2512975:d=2021032312:UGRD:325 mb:26 hour fcst: - 7: 7:2827487:d=2021032312:VGRD:325 mb:26 hour fcst: - 7: 8:3160424:d=2021032312:ABSV:325 mb:26 hour fcst: - 7: 9:3710430:d=2021032312:CLMR:325 mb:26 hour fcst: - 7: 10:3710609:d=2021032312:ICMR:325 mb:26 hour fcst: - 7: 11:4078174:d=2021032312:RWMR:325 mb:26 hour fcst: - 7: 12:4078353:d=2021032312:SNMR:325 mb:26 hour fcst: - 7: 13:4323307:d=2021032312:GRLE:325 mb:26 hour fcst: - 7: 14:4328197:d=2021032312:HGT:375 mb:26 hour fcst: - 7: + product_functions.sh[10]rc=0 - 7: + product_functions.sh[11](( rc == 0 )) - 7: + product_functions.sh[11]mv pgb2bfile_f026_8_0p25.new pgb2bfile_f026_8_0p25 - 7: + product_functions.sh[12]return 0 - 7: + interp_atmos_master.sh[56]export err=0 - 7: + interp_atmos_master.sh[56]err=0 - 7: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 7: ++ interp_atmos_master.sh[62]wc -l - 7: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_8_0p25 -match 'LAND|ICEC' - 7: + interp_atmos_master.sh[62]var_count=0 - 7: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 7: + interp_atmos_master.sh[73]exit 0 -21: + bash[8]'[' -z '' ']' -21: + bash[9]case "$-" in -21: + bash[12]__lmod_vx=x -21: + bash[16]'[' -n x ']' -21: + bash[16]set +x -21: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -21: Shell debugging restarted -21: + bash[224]unset __lmod_vx -21: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_22 -21: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_22 -21: + interp_atmos_master.sh[9]grid_string=0p25 -21: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -21: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -21: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -21: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -21: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -21: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -21: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -21: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -21: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -21: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -21: + interp_atmos_master.sh[31]IFS=: -21: + interp_atmos_master.sh[31]read -ra grids -21: + interp_atmos_master.sh[33]output_grids= -21: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -21: + interp_atmos_master.sh[35]gridopt=grid0p25 -21: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_22_0p25' -21: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_22 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_22_0p25 -21: 1:0:d=2021032312:RH:90-60 mb above ground:26 hour fcst: -21: 2:19065:d=2021032312:SPFH:90-60 mb above ground:26 hour fcst: -21: 3:47970:d=2021032312:UGRD:90-60 mb above ground:26 hour fcst: -21: 4:69369:d=2021032312:VGRD:90-60 mb above ground:26 hour fcst: -21: 5:91377:d=2021032312:TMP:120-90 mb above ground:26 hour fcst: -21: 6:112303:d=2021032312:RH:120-90 mb above ground:26 hour fcst: -21: 7:131641:d=2021032312:SPFH:120-90 mb above ground:26 hour fcst: -21: 8:160599:d=2021032312:UGRD:120-90 mb above ground:26 hour fcst: -21: 9:181834:d=2021032312:VGRD:120-90 mb above ground:26 hour fcst: -21: 10:203601:d=2021032312:TMP:150-120 mb above ground:26 hour fcst: -21: 11:224508:d=2021032312:RH:150-120 mb above ground:26 hour fcst: -21: 12:244139:d=2021032312:SPFH:150-120 mb above ground:26 hour fcst: -21: 13:273185:d=2021032312:UGRD:150-120 mb above ground:26 hour fcst: -21: 14:294224:d=2021032312:VGRD:150-120 mb above ground:26 hour fcst: -21: + interp_atmos_master.sh[47]export err=0 -21: + interp_atmos_master.sh[47]err=0 -21: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -21: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -21: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_22_0p25 -21: + product_functions.sh[5]local filename=pgb2bfile_f026_22_0p25 -21: + product_functions.sh[6]wgrib2 pgb2bfile_f026_22_0p25 -not_if :RH: -grib pgb2bfile_f026_22_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_22_0p25.new -21: 1:0:d=2021032312:RH:90-60 mb above ground:26 hour fcst: -21: 2:426588:d=2021032312:SPFH:90-60 mb above ground:26 hour fcst: -21: 3:1215879:d=2021032312:UGRD:90-60 mb above ground:26 hour fcst: -21: 4:1731505:d=2021032312:VGRD:90-60 mb above ground:26 hour fcst: -21: 5:2278887:d=2021032312:TMP:120-90 mb above ground:26 hour fcst: -21: 6:2779427:d=2021032312:RH:120-90 mb above ground:26 hour fcst: -21: 7:3215422:d=2021032312:SPFH:120-90 mb above ground:26 hour fcst: -21: 8:4006161:d=2021032312:UGRD:120-90 mb above ground:26 hour fcst: -21: 9:4514872:d=2021032312:VGRD:120-90 mb above ground:26 hour fcst: -21: 10:5054718:d=2021032312:TMP:150-120 mb above ground:26 hour fcst: -21: 11:5556962:d=2021032312:RH:150-120 mb above ground:26 hour fcst: -21: 12:6002396:d=2021032312:SPFH:150-120 mb above ground:26 hour fcst: -21: 13:6793979:d=2021032312:UGRD:150-120 mb above ground:26 hour fcst: -21: 14:7297939:d=2021032312:VGRD:150-120 mb above ground:26 hour fcst: -21: + product_functions.sh[10]rc=0 -21: + product_functions.sh[11](( rc == 0 )) -21: + product_functions.sh[11]mv pgb2bfile_f026_22_0p25.new pgb2bfile_f026_22_0p25 -21: + product_functions.sh[12]return 0 -21: + interp_atmos_master.sh[56]export err=0 -21: + interp_atmos_master.sh[56]err=0 -21: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -21: ++ interp_atmos_master.sh[62]wc -l -21: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_22_0p25 -match 'LAND|ICEC' -21: + interp_atmos_master.sh[62]var_count=0 -21: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -21: + interp_atmos_master.sh[73]exit 0 -12: + bash[8]'[' -z '' ']' -12: + bash[9]case "$-" in -12: + bash[12]__lmod_vx=x -12: + bash[16]'[' -n x ']' -12: + bash[16]set +x -12: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -12: Shell debugging restarted -12: + bash[224]unset __lmod_vx -12: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_13 -12: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_13 -12: + interp_atmos_master.sh[9]grid_string=0p25 -12: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -12: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -12: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -12: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -12: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -12: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -12: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -12: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -12: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -12: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -12: + interp_atmos_master.sh[31]IFS=: -12: + interp_atmos_master.sh[31]read -ra grids -12: + interp_atmos_master.sh[33]output_grids= -12: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -12: + interp_atmos_master.sh[35]gridopt=grid0p25 -12: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_13_0p25' -12: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_13 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_13_0p25 -12: 1:0:d=2021032312:TMP:575 mb:26 hour fcst: -12: 2:18945:d=2021032312:RH:575 mb:26 hour fcst: -12: 3:38816:d=2021032312:TCDC:575 mb:26 hour fcst: -12: 4:51018:d=2021032312:VVEL:575 mb:26 hour fcst: -12: 5:76234:d=2021032312:DZDT:575 mb:26 hour fcst: -12: 6:103682:d=2021032312:UGRD:575 mb:26 hour fcst: -12: 7:124628:d=2021032312:VGRD:575 mb:26 hour fcst: -12: 8:146101:d=2021032312:ABSV:575 mb:26 hour fcst: -12: 9:166944:d=2021032312:CLMR:575 mb:26 hour fcst: -12: 10:171203:d=2021032312:ICMR:575 mb:26 hour fcst: -12: 11:187196:d=2021032312:RWMR:575 mb:26 hour fcst: -12: 12:189551:d=2021032312:SNMR:575 mb:26 hour fcst: -12: 13:204109:d=2021032312:GRLE:575 mb:26 hour fcst: -12: 14:206137:d=2021032312:HGT:625 mb:26 hour fcst: -12: + interp_atmos_master.sh[47]export err=0 -12: + interp_atmos_master.sh[47]err=0 -12: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -12: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -12: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_13_0p25 -12: + product_functions.sh[5]local filename=pgb2bfile_f026_13_0p25 -12: + product_functions.sh[6]wgrib2 pgb2bfile_f026_13_0p25 -not_if :RH: -grib pgb2bfile_f026_13_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_13_0p25.new -12: 1:0:d=2021032312:TMP:575 mb:26 hour fcst: -12: 2:424271:d=2021032312:RH:575 mb:26 hour fcst: -12: 3:878510:d=2021032312:TCDC:575 mb:26 hour fcst: -12: 4:1108423:d=2021032312:VVEL:575 mb:26 hour fcst: -12: 5:1780300:d=2021032312:DZDT:575 mb:26 hour fcst: -12: 6:2541000:d=2021032312:UGRD:575 mb:26 hour fcst: -12: 7:3050077:d=2021032312:VGRD:575 mb:26 hour fcst: -12: 8:3589084:d=2021032312:ABSV:575 mb:26 hour fcst: -12: 9:4075012:d=2021032312:CLMR:575 mb:26 hour fcst: -12: 10:4164536:d=2021032312:ICMR:575 mb:26 hour fcst: -12: 11:4518726:d=2021032312:RWMR:575 mb:26 hour fcst: -12: 12:4562398:d=2021032312:SNMR:575 mb:26 hour fcst: -12: 13:4877800:d=2021032312:GRLE:575 mb:26 hour fcst: -12: 14:4914782:d=2021032312:HGT:625 mb:26 hour fcst: -12: + product_functions.sh[10]rc=0 -12: + product_functions.sh[11](( rc == 0 )) -12: + product_functions.sh[11]mv pgb2bfile_f026_13_0p25.new pgb2bfile_f026_13_0p25 -12: + product_functions.sh[12]return 0 -12: + interp_atmos_master.sh[56]export err=0 -12: + interp_atmos_master.sh[56]err=0 -12: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -12: ++ interp_atmos_master.sh[62]wc -l -12: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_13_0p25 -match 'LAND|ICEC' -12: + interp_atmos_master.sh[62]var_count=0 -12: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -12: + interp_atmos_master.sh[73]exit 0 -19: + bash[8]'[' -z '' ']' -19: + bash[9]case "$-" in -19: + bash[12]__lmod_vx=x -19: + bash[16]'[' -n x ']' -19: + bash[16]set +x -19: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -19: Shell debugging restarted -19: + bash[224]unset __lmod_vx -19: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_20 -19: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_20 -19: + interp_atmos_master.sh[9]grid_string=0p25 -19: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -19: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -19: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -19: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -19: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -19: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -19: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -19: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -19: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -19: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -19: + interp_atmos_master.sh[31]IFS=: -19: + interp_atmos_master.sh[31]read -ra grids -19: + interp_atmos_master.sh[33]output_grids= -19: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -19: + interp_atmos_master.sh[35]gridopt=grid0p25 -19: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_20_0p25' -19: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_20 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_20_0p25 -19: 1:0:d=2021032312:SOILL:0.1-0.4 m below ground:26 hour fcst: -19: 2:8892:d=2021032312:SOILL:0.4-1 m below ground:26 hour fcst: -19: 3:17815:d=2021032312:SOILL:1-2 m below ground:26 hour fcst: -19: 4:26892:d=2021032312:CNWAT:surface:26 hour fcst: -19: 5:32414:d=2021032312:ICETK:surface:26 hour fcst: -19: 6:35822:d=2021032312:DUVB:surface:24-26 hour ave fcst: -19: 7:53252:d=2021032312:CDUVB:surface:24-26 hour ave fcst: -19: 8:66972:d=2021032312:TMP:305 m above mean sea level:26 hour fcst: -19: 9:85177:d=2021032312:UGRD:305 m above mean sea level:26 hour fcst: -19: 10:104340:d=2021032312:VGRD:305 m above mean sea level:26 hour fcst: -19: 11:123983:d=2021032312:TMP:457 m above mean sea level:26 hour fcst: -19: 12:143391:d=2021032312:UGRD:457 m above mean sea level:26 hour fcst: -19: 13:163633:d=2021032312:VGRD:457 m above mean sea level:26 hour fcst: -19: 14:184196:d=2021032312:TMP:610 m above mean sea level:26 hour fcst: -19: + interp_atmos_master.sh[47]export err=0 -19: + interp_atmos_master.sh[47]err=0 -19: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -19: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -19: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_20_0p25 -19: + product_functions.sh[5]local filename=pgb2bfile_f026_20_0p25 -19: + product_functions.sh[6]wgrib2 pgb2bfile_f026_20_0p25 -not_if :RH: -grib pgb2bfile_f026_20_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_20_0p25.new -19: 1:0:d=2021032312:SOILL:0.1-0.4 m below ground:26 hour fcst: -19: 2:294038:d=2021032312:SOILL:0.4-1 m below ground:26 hour fcst: -19: 3:589652:d=2021032312:SOILL:1-2 m below ground:26 hour fcst: -19: 4:885774:d=2021032312:CNWAT:surface:26 hour fcst: -19: 5:1071486:d=2021032312:ICETK:surface:26 hour fcst: -19: 6:1124846:d=2021032312:DUVB:surface:24-26 hour ave fcst: -19: 7:1568929:d=2021032312:CDUVB:surface:24-26 hour ave fcst: -19: 8:1899174:d=2021032312:TMP:305 m above mean sea level:26 hour fcst: -19: 9:2411464:d=2021032312:UGRD:305 m above mean sea level:26 hour fcst: -19: 10:2971604:d=2021032312:VGRD:305 m above mean sea level:26 hour fcst: -19: 11:3547753:d=2021032312:TMP:457 m above mean sea level:26 hour fcst: -19: 12:4091619:d=2021032312:UGRD:457 m above mean sea level:26 hour fcst: -19: 13:4677396:d=2021032312:VGRD:457 m above mean sea level:26 hour fcst: -19: 14:5282055:d=2021032312:TMP:610 m above mean sea level:26 hour fcst: -19: + product_functions.sh[10]rc=0 -19: + product_functions.sh[11](( rc == 0 )) -19: + product_functions.sh[11]mv pgb2bfile_f026_20_0p25.new pgb2bfile_f026_20_0p25 -19: + product_functions.sh[12]return 0 -19: + interp_atmos_master.sh[56]export err=0 -19: + interp_atmos_master.sh[56]err=0 -19: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -19: ++ interp_atmos_master.sh[62]wc -l -19: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_20_0p25 -match 'LAND|ICEC' -19: + interp_atmos_master.sh[62]var_count=0 -19: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -19: + interp_atmos_master.sh[73]exit 0 - 8: + bash[8]'[' -z '' ']' - 8: + bash[9]case "$-" in - 8: + bash[12]__lmod_vx=x - 8: + bash[16]'[' -n x ']' - 8: + bash[16]set +x - 8: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 8: Shell debugging restarted - 8: + bash[224]unset __lmod_vx - 8: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_9 - 8: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_9 - 8: + interp_atmos_master.sh[9]grid_string=0p25 - 8: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 8: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 8: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 8: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 8: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 8: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 8: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 8: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 8: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 8: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 8: + interp_atmos_master.sh[31]IFS=: - 8: + interp_atmos_master.sh[31]read -ra grids - 8: + interp_atmos_master.sh[33]output_grids= - 8: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 8: + interp_atmos_master.sh[35]gridopt=grid0p25 - 8: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_9_0p25' - 8: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_9 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_9_0p25 - 8: 1:0:d=2021032312:TMP:375 mb:26 hour fcst: - 8: 2:18325:d=2021032312:RH:375 mb:26 hour fcst: - 8: 3:37489:d=2021032312:TCDC:375 mb:26 hour fcst: - 8: 4:50699:d=2021032312:VVEL:375 mb:26 hour fcst: - 8: 5:75531:d=2021032312:DZDT:375 mb:26 hour fcst: - 8: 6:103631:d=2021032312:UGRD:375 mb:26 hour fcst: - 8: 7:118113:d=2021032312:VGRD:375 mb:26 hour fcst: - 8: 8:133211:d=2021032312:ABSV:375 mb:26 hour fcst: - 8: 9:155348:d=2021032312:CLMR:375 mb:26 hour fcst: - 8: 10:155604:d=2021032312:ICMR:375 mb:26 hour fcst: - 8: 11:172693:d=2021032312:RWMR:375 mb:26 hour fcst: - 8: 12:172950:d=2021032312:SNMR:375 mb:26 hour fcst: - 8: 13:186077:d=2021032312:GRLE:375 mb:26 hour fcst: - 8: 14:186691:d=2021032312:HGT:425 mb:26 hour fcst: - 8: + interp_atmos_master.sh[47]export err=0 - 8: + interp_atmos_master.sh[47]err=0 - 8: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 8: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 8: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_9_0p25 - 8: + product_functions.sh[5]local filename=pgb2bfile_f026_9_0p25 - 8: + product_functions.sh[6]wgrib2 pgb2bfile_f026_9_0p25 -not_if :RH: -grib pgb2bfile_f026_9_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_9_0p25.new - 8: 1:0:d=2021032312:TMP:375 mb:26 hour fcst: - 8: 2:410047:d=2021032312:RH:375 mb:26 hour fcst: - 8: 3:843801:d=2021032312:TCDC:375 mb:26 hour fcst: - 8: 4:1084128:d=2021032312:VVEL:375 mb:26 hour fcst: - 8: 5:1742990:d=2021032312:DZDT:375 mb:26 hour fcst: - 8: 6:2537028:d=2021032312:UGRD:375 mb:26 hour fcst: - 8: 7:2846810:d=2021032312:VGRD:375 mb:26 hour fcst: - 8: 8:3174736:d=2021032312:ABSV:375 mb:26 hour fcst: - 8: 9:3717833:d=2021032312:CLMR:375 mb:26 hour fcst: - 8: 10:3718715:d=2021032312:ICMR:375 mb:26 hour fcst: - 8: 11:4095124:d=2021032312:RWMR:375 mb:26 hour fcst: - 8: 12:4095974:d=2021032312:SNMR:375 mb:26 hour fcst: - 8: 13:4369108:d=2021032312:GRLE:375 mb:26 hour fcst: - 8: 14:4379334:d=2021032312:HGT:425 mb:26 hour fcst: - 8: + product_functions.sh[10]rc=0 - 8: + product_functions.sh[11](( rc == 0 )) - 8: + product_functions.sh[11]mv pgb2bfile_f026_9_0p25.new pgb2bfile_f026_9_0p25 - 8: + product_functions.sh[12]return 0 - 8: + interp_atmos_master.sh[56]export err=0 - 8: + interp_atmos_master.sh[56]err=0 - 8: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 8: ++ interp_atmos_master.sh[62]wc -l - 8: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_9_0p25 -match 'LAND|ICEC' - 8: + interp_atmos_master.sh[62]var_count=0 - 8: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 8: + interp_atmos_master.sh[73]exit 0 -16: + bash[8]'[' -z '' ']' -16: + bash[9]case "$-" in -16: + bash[12]__lmod_vx=x -16: + bash[16]'[' -n x ']' -16: + bash[16]set +x -16: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -16: Shell debugging restarted -16: + bash[224]unset __lmod_vx -16: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_17 -16: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_17 -16: + interp_atmos_master.sh[9]grid_string=0p25 -16: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -16: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -16: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -16: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -16: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -16: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -16: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -16: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -16: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -16: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -16: + interp_atmos_master.sh[31]IFS=: -16: + interp_atmos_master.sh[31]read -ra grids -16: + interp_atmos_master.sh[33]output_grids= -16: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -16: + interp_atmos_master.sh[35]gridopt=grid0p25 -16: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_17_0p25' -16: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_17 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_17_0p25 -16: 1:0:d=2021032312:TMP:775 mb:26 hour fcst: -16: 2:20096:d=2021032312:RH:775 mb:26 hour fcst: -16: 3:40529:d=2021032312:TCDC:775 mb:26 hour fcst: -16: 4:53106:d=2021032312:VVEL:775 mb:26 hour fcst: -16: 5:78862:d=2021032312:DZDT:775 mb:26 hour fcst: -16: 6:105977:d=2021032312:UGRD:775 mb:26 hour fcst: -16: 7:126841:d=2021032312:VGRD:775 mb:26 hour fcst: -16: 8:148568:d=2021032312:ABSV:775 mb:26 hour fcst: -16: 9:169741:d=2021032312:CLMR:775 mb:26 hour fcst: -16: 10:180011:d=2021032312:ICMR:775 mb:26 hour fcst: -16: 11:191277:d=2021032312:RWMR:775 mb:26 hour fcst: -16: 12:198045:d=2021032312:SNMR:775 mb:26 hour fcst: -16: 13:211281:d=2021032312:GRLE:775 mb:26 hour fcst: -16: 14:213815:d=2021032312:HGT:825 mb:26 hour fcst: -16: + interp_atmos_master.sh[47]export err=0 -16: + interp_atmos_master.sh[47]err=0 -16: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -16: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -16: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_17_0p25 -16: + product_functions.sh[5]local filename=pgb2bfile_f026_17_0p25 -16: + product_functions.sh[6]wgrib2 pgb2bfile_f026_17_0p25 -not_if :RH: -grib pgb2bfile_f026_17_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_17_0p25.new -16: 1:0:d=2021032312:TMP:775 mb:26 hour fcst: -16: 2:469556:d=2021032312:RH:775 mb:26 hour fcst: -16: 3:949463:d=2021032312:TCDC:775 mb:26 hour fcst: -16: 4:1204238:d=2021032312:VVEL:775 mb:26 hour fcst: -16: 5:1890990:d=2021032312:DZDT:775 mb:26 hour fcst: -16: 6:2638691:d=2021032312:UGRD:775 mb:26 hour fcst: -16: 7:3144830:d=2021032312:VGRD:775 mb:26 hour fcst: -16: 8:3679920:d=2021032312:ABSV:775 mb:26 hour fcst: -16: 9:4184537:d=2021032312:CLMR:775 mb:26 hour fcst: -16: 10:4409152:d=2021032312:ICMR:775 mb:26 hour fcst: -16: 11:4679767:d=2021032312:RWMR:775 mb:26 hour fcst: -16: 12:4805543:d=2021032312:SNMR:775 mb:26 hour fcst: -16: 13:5101218:d=2021032312:GRLE:775 mb:26 hour fcst: -16: 14:5146381:d=2021032312:HGT:825 mb:26 hour fcst: -16: + product_functions.sh[10]rc=0 -16: + product_functions.sh[11](( rc == 0 )) -16: + product_functions.sh[11]mv pgb2bfile_f026_17_0p25.new pgb2bfile_f026_17_0p25 -16: + product_functions.sh[12]return 0 -16: + interp_atmos_master.sh[56]export err=0 -16: + interp_atmos_master.sh[56]err=0 -16: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -16: ++ interp_atmos_master.sh[62]wc -l -16: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_17_0p25 -match 'LAND|ICEC' -16: + interp_atmos_master.sh[62]var_count=0 -16: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -16: + interp_atmos_master.sh[73]exit 0 -18: + bash[8]'[' -z '' ']' -18: + bash[9]case "$-" in -18: + bash[12]__lmod_vx=x -18: + bash[16]'[' -n x ']' -18: + bash[16]set +x -18: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -18: Shell debugging restarted -18: + bash[224]unset __lmod_vx -18: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_19 -18: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_19 -18: + interp_atmos_master.sh[9]grid_string=0p25 -18: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -18: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -18: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -18: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -18: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -18: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -18: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -18: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -18: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -18: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -18: + interp_atmos_master.sh[31]IFS=: -18: + interp_atmos_master.sh[31]read -ra grids -18: + interp_atmos_master.sh[33]output_grids= -18: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -18: + interp_atmos_master.sh[35]gridopt=grid0p25 -18: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_19_0p25' -18: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_19 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_19_0p25 -18: 1:0:d=2021032312:TMP:875 mb:26 hour fcst: -18: 2:20930:d=2021032312:RH:875 mb:26 hour fcst: -18: 3:40923:d=2021032312:TCDC:875 mb:26 hour fcst: -18: 4:55160:d=2021032312:VVEL:875 mb:26 hour fcst: -18: 5:80504:d=2021032312:DZDT:875 mb:26 hour fcst: -18: 6:106887:d=2021032312:UGRD:875 mb:26 hour fcst: -18: 7:128190:d=2021032312:VGRD:875 mb:26 hour fcst: -18: 8:150041:d=2021032312:ABSV:875 mb:26 hour fcst: -18: 9:171289:d=2021032312:CLMR:875 mb:26 hour fcst: -18: 10:183516:d=2021032312:ICMR:875 mb:26 hour fcst: -18: 11:192677:d=2021032312:RWMR:875 mb:26 hour fcst: -18: 12:204001:d=2021032312:SNMR:875 mb:26 hour fcst: -18: 13:215000:d=2021032312:GRLE:875 mb:26 hour fcst: -18: 14:217397:d=2021032312:SOILL:0-0.1 m below ground:26 hour fcst: -18: + interp_atmos_master.sh[47]export err=0 -18: + interp_atmos_master.sh[47]err=0 -18: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -18: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -18: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_19_0p25 -18: + product_functions.sh[5]local filename=pgb2bfile_f026_19_0p25 -18: + product_functions.sh[6]wgrib2 pgb2bfile_f026_19_0p25 -not_if :RH: -grib pgb2bfile_f026_19_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_19_0p25.new -18: 1:0:d=2021032312:TMP:875 mb:26 hour fcst: -18: 2:494205:d=2021032312:RH:875 mb:26 hour fcst: -18: 3:955145:d=2021032312:TCDC:875 mb:26 hour fcst: -18: 4:1246147:d=2021032312:VVEL:875 mb:26 hour fcst: -18: 5:1918873:d=2021032312:DZDT:875 mb:26 hour fcst: -18: 6:2643147:d=2021032312:UGRD:875 mb:26 hour fcst: -18: 7:3157034:d=2021032312:VGRD:875 mb:26 hour fcst: -18: 8:3700995:d=2021032312:ABSV:875 mb:26 hour fcst: -18: 9:4206994:d=2021032312:CLMR:875 mb:26 hour fcst: -18: 10:4473649:d=2021032312:ICMR:875 mb:26 hour fcst: -18: 11:4692508:d=2021032312:RWMR:875 mb:26 hour fcst: -18: 12:4914867:d=2021032312:SNMR:875 mb:26 hour fcst: -18: 13:5167038:d=2021032312:GRLE:875 mb:26 hour fcst: -18: 14:5210799:d=2021032312:SOILL:0-0.1 m below ground:26 hour fcst: -18: + product_functions.sh[10]rc=0 -18: + product_functions.sh[11](( rc == 0 )) -18: + product_functions.sh[11]mv pgb2bfile_f026_19_0p25.new pgb2bfile_f026_19_0p25 -18: + product_functions.sh[12]return 0 -18: + interp_atmos_master.sh[56]export err=0 -18: + interp_atmos_master.sh[56]err=0 -18: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -18: ++ interp_atmos_master.sh[62]wc -l -18: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_19_0p25 -match 'LAND|ICEC' -18: + interp_atmos_master.sh[62]var_count=0 -18: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -18: + interp_atmos_master.sh[73]exit 0 -13: + bash[8]'[' -z '' ']' -13: + bash[9]case "$-" in -13: + bash[12]__lmod_vx=x -13: + bash[16]'[' -n x ']' -13: + bash[16]set +x -13: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -13: Shell debugging restarted -13: + bash[224]unset __lmod_vx -13: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_14 -13: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_14 -13: + interp_atmos_master.sh[9]grid_string=0p25 -13: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -13: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -13: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -13: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -13: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -13: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -13: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -13: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -13: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -13: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -13: + interp_atmos_master.sh[31]IFS=: -13: + interp_atmos_master.sh[31]read -ra grids -13: + interp_atmos_master.sh[33]output_grids= -13: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -13: + interp_atmos_master.sh[35]gridopt=grid0p25 -13: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_14_0p25' -13: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_14 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_14_0p25 -13: 1:0:d=2021032312:TMP:625 mb:26 hour fcst: -13: 2:19145:d=2021032312:RH:625 mb:26 hour fcst: -13: 3:39232:d=2021032312:TCDC:625 mb:26 hour fcst: -13: 4:51049:d=2021032312:VVEL:625 mb:26 hour fcst: -13: 5:76530:d=2021032312:DZDT:625 mb:26 hour fcst: -13: 6:103940:d=2021032312:UGRD:625 mb:26 hour fcst: -13: 7:124941:d=2021032312:VGRD:625 mb:26 hour fcst: -13: 8:146593:d=2021032312:ABSV:625 mb:26 hour fcst: -13: 9:167375:d=2021032312:CLMR:625 mb:26 hour fcst: -13: 10:172433:d=2021032312:ICMR:625 mb:26 hour fcst: -13: 11:185468:d=2021032312:RWMR:625 mb:26 hour fcst: -13: 12:189707:d=2021032312:SNMR:625 mb:26 hour fcst: -13: 13:204975:d=2021032312:GRLE:625 mb:26 hour fcst: -13: 14:207403:d=2021032312:HGT:675 mb:26 hour fcst: -13: + interp_atmos_master.sh[47]export err=0 -13: + interp_atmos_master.sh[47]err=0 -13: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -13: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -13: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_14_0p25 -13: + product_functions.sh[5]local filename=pgb2bfile_f026_14_0p25 -13: + product_functions.sh[6]wgrib2 pgb2bfile_f026_14_0p25 -not_if :RH: -grib pgb2bfile_f026_14_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_14_0p25.new -13: 1:0:d=2021032312:TMP:625 mb:26 hour fcst: -13: 2:427795:d=2021032312:RH:625 mb:26 hour fcst: -13: 3:896855:d=2021032312:TCDC:625 mb:26 hour fcst: -13: 4:1120268:d=2021032312:VVEL:625 mb:26 hour fcst: -13: 5:1795981:d=2021032312:DZDT:625 mb:26 hour fcst: -13: 6:2554023:d=2021032312:UGRD:625 mb:26 hour fcst: -13: 7:3060011:d=2021032312:VGRD:625 mb:26 hour fcst: -13: 8:3595820:d=2021032312:ABSV:625 mb:26 hour fcst: -13: 9:4079126:d=2021032312:CLMR:625 mb:26 hour fcst: -13: 10:4179898:d=2021032312:ICMR:625 mb:26 hour fcst: -13: 11:4467148:d=2021032312:RWMR:625 mb:26 hour fcst: -13: 12:4555027:d=2021032312:SNMR:625 mb:26 hour fcst: -13: 13:4891167:d=2021032312:GRLE:625 mb:26 hour fcst: -13: 14:4935907:d=2021032312:HGT:675 mb:26 hour fcst: -13: + product_functions.sh[10]rc=0 -13: + product_functions.sh[11](( rc == 0 )) -13: + product_functions.sh[11]mv pgb2bfile_f026_14_0p25.new pgb2bfile_f026_14_0p25 -13: + product_functions.sh[12]return 0 -13: + interp_atmos_master.sh[56]export err=0 -13: + interp_atmos_master.sh[56]err=0 -13: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -13: ++ interp_atmos_master.sh[62]wc -l -13: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_14_0p25 -match 'LAND|ICEC' -13: + interp_atmos_master.sh[62]var_count=0 -13: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -13: + interp_atmos_master.sh[73]exit 0 -10: + bash[8]'[' -z '' ']' -10: + bash[9]case "$-" in -10: + bash[12]__lmod_vx=x -10: + bash[16]'[' -n x ']' -10: + bash[16]set +x -10: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -10: Shell debugging restarted -10: + bash[224]unset __lmod_vx -10: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_11 -10: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_11 -10: + interp_atmos_master.sh[9]grid_string=0p25 -10: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -10: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -10: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -10: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -10: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -10: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -10: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -10: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -10: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -10: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -10: + interp_atmos_master.sh[31]IFS=: -10: + interp_atmos_master.sh[31]read -ra grids -10: + interp_atmos_master.sh[33]output_grids= -10: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -10: + interp_atmos_master.sh[35]gridopt=grid0p25 -10: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_11_0p25' -10: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_11 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_11_0p25 -10: 1:0:d=2021032312:TMP:475 mb:26 hour fcst: -10: 2:18476:d=2021032312:RH:475 mb:26 hour fcst: -10: 3:38134:d=2021032312:TCDC:475 mb:26 hour fcst: -10: 4:50647:d=2021032312:VVEL:475 mb:26 hour fcst: -10: 5:75654:d=2021032312:DZDT:475 mb:26 hour fcst: -10: 6:103609:d=2021032312:UGRD:475 mb:26 hour fcst: -10: 7:125187:d=2021032312:VGRD:475 mb:26 hour fcst: -10: 8:147344:d=2021032312:ABSV:475 mb:26 hour fcst: -10: 9:168625:d=2021032312:CLMR:475 mb:26 hour fcst: -10: 10:169807:d=2021032312:ICMR:475 mb:26 hour fcst: -10: 11:186162:d=2021032312:RWMR:475 mb:26 hour fcst: -10: 12:186845:d=2021032312:SNMR:475 mb:26 hour fcst: -10: 13:200807:d=2021032312:GRLE:475 mb:26 hour fcst: -10: 14:202322:d=2021032312:HGT:525 mb:26 hour fcst: -10: + interp_atmos_master.sh[47]export err=0 -10: + interp_atmos_master.sh[47]err=0 -10: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -10: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -10: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_11_0p25 -10: + product_functions.sh[5]local filename=pgb2bfile_f026_11_0p25 -10: + product_functions.sh[6]wgrib2 pgb2bfile_f026_11_0p25 -not_if :RH: -grib pgb2bfile_f026_11_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_11_0p25.new -10: 1:0:d=2021032312:TMP:475 mb:26 hour fcst: -10: 2:422880:d=2021032312:RH:475 mb:26 hour fcst: -10: 3:872998:d=2021032312:TCDC:475 mb:26 hour fcst: -10: 4:1110834:d=2021032312:VVEL:475 mb:26 hour fcst: -10: 5:1777810:d=2021032312:DZDT:475 mb:26 hour fcst: -10: 6:2560268:d=2021032312:UGRD:475 mb:26 hour fcst: -10: 7:3089346:d=2021032312:VGRD:475 mb:26 hour fcst: -10: 8:3646788:d=2021032312:ABSV:475 mb:26 hour fcst: -10: 9:4164838:d=2021032312:CLMR:475 mb:26 hour fcst: -10: 10:4186476:d=2021032312:ICMR:475 mb:26 hour fcst: -10: 11:4548457:d=2021032312:RWMR:475 mb:26 hour fcst: -10: 12:4559021:d=2021032312:SNMR:475 mb:26 hour fcst: -10: 13:4859525:d=2021032312:GRLE:475 mb:26 hour fcst: -10: 14:4880167:d=2021032312:HGT:525 mb:26 hour fcst: -10: + product_functions.sh[10]rc=0 -10: + product_functions.sh[11](( rc == 0 )) -10: + product_functions.sh[11]mv pgb2bfile_f026_11_0p25.new pgb2bfile_f026_11_0p25 -10: + product_functions.sh[12]return 0 -10: + interp_atmos_master.sh[56]export err=0 -10: + interp_atmos_master.sh[56]err=0 -10: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -10: ++ interp_atmos_master.sh[62]wc -l -10: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_11_0p25 -match 'LAND|ICEC' -10: + interp_atmos_master.sh[62]var_count=0 -10: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -10: + interp_atmos_master.sh[73]exit 0 -20: + bash[8]'[' -z '' ']' -20: + bash[9]case "$-" in -20: + bash[12]__lmod_vx=x -20: + bash[16]'[' -n x ']' -20: + bash[16]set +x -20: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -20: Shell debugging restarted -20: + bash[224]unset __lmod_vx -20: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_21 -20: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_21 -20: + interp_atmos_master.sh[9]grid_string=0p25 -20: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -20: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -20: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -20: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -20: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -20: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -20: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -20: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -20: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -20: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -20: + interp_atmos_master.sh[31]IFS=: -20: + interp_atmos_master.sh[31]read -ra grids -20: + interp_atmos_master.sh[33]output_grids= -20: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -20: + interp_atmos_master.sh[35]gridopt=grid0p25 -20: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_21_0p25' -20: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_21 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_21_0p25 -20: 1:0:d=2021032312:UGRD:610 m above mean sea level:26 hour fcst: -20: 2:20821:d=2021032312:VGRD:610 m above mean sea level:26 hour fcst: -20: 3:42017:d=2021032312:TMP:914 m above mean sea level:26 hour fcst: -20: 4:62917:d=2021032312:UGRD:914 m above mean sea level:26 hour fcst: -20: 5:84274:d=2021032312:VGRD:914 m above mean sea level:26 hour fcst: -20: 6:106080:d=2021032312:TMP:4572 m above mean sea level:26 hour fcst: -20: 7:127277:d=2021032312:UGRD:4572 m above mean sea level:26 hour fcst: -20: 8:150480:d=2021032312:VGRD:4572 m above mean sea level:26 hour fcst: -20: 9:174477:d=2021032312:TMP:60-30 mb above ground:26 hour fcst: -20: 10:195348:d=2021032312:RH:60-30 mb above ground:26 hour fcst: -20: 11:213985:d=2021032312:SPFH:60-30 mb above ground:26 hour fcst: -20: 12:242763:d=2021032312:UGRD:60-30 mb above ground:26 hour fcst: -20: 13:264372:d=2021032312:VGRD:60-30 mb above ground:26 hour fcst: -20: 14:286527:d=2021032312:TMP:90-60 mb above ground:26 hour fcst: -20: + interp_atmos_master.sh[47]export err=0 -20: + interp_atmos_master.sh[47]err=0 -20: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -20: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -20: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_21_0p25 -20: + product_functions.sh[5]local filename=pgb2bfile_f026_21_0p25 -20: + product_functions.sh[6]wgrib2 pgb2bfile_f026_21_0p25 -not_if :RH: -grib pgb2bfile_f026_21_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_21_0p25.new -20: 1:0:d=2021032312:UGRD:610 m above mean sea level:26 hour fcst: -20: 2:593181:d=2021032312:VGRD:610 m above mean sea level:26 hour fcst: -20: 3:1211109:d=2021032312:TMP:914 m above mean sea level:26 hour fcst: -20: 4:1797074:d=2021032312:UGRD:914 m above mean sea level:26 hour fcst: -20: 5:2400084:d=2021032312:VGRD:914 m above mean sea level:26 hour fcst: -20: 6:3030802:d=2021032312:TMP:4572 m above mean sea level:26 hour fcst: -20: 7:3603601:d=2021032312:UGRD:4572 m above mean sea level:26 hour fcst: -20: 8:4248307:d=2021032312:VGRD:4572 m above mean sea level:26 hour fcst: -20: 9:4922932:d=2021032312:TMP:60-30 mb above ground:26 hour fcst: -20: 10:5408704:d=2021032312:RH:60-30 mb above ground:26 hour fcst: -20: 11:5819516:d=2021032312:SPFH:60-30 mb above ground:26 hour fcst: -20: 12:6604564:d=2021032312:UGRD:60-30 mb above ground:26 hour fcst: -20: 13:7127305:d=2021032312:VGRD:60-30 mb above ground:26 hour fcst: -20: 14:7678861:d=2021032312:TMP:90-60 mb above ground:26 hour fcst: -20: + product_functions.sh[10]rc=0 -20: + product_functions.sh[11](( rc == 0 )) -20: + product_functions.sh[11]mv pgb2bfile_f026_21_0p25.new pgb2bfile_f026_21_0p25 -20: + product_functions.sh[12]return 0 -20: + interp_atmos_master.sh[56]export err=0 -20: + interp_atmos_master.sh[56]err=0 -20: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -20: ++ interp_atmos_master.sh[62]wc -l -20: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_21_0p25 -match 'LAND|ICEC' -20: + interp_atmos_master.sh[62]var_count=0 -20: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -20: + interp_atmos_master.sh[73]exit 0 -14: + bash[8]'[' -z '' ']' -14: + bash[9]case "$-" in -14: + bash[12]__lmod_vx=x -14: + bash[16]'[' -n x ']' -14: + bash[16]set +x -14: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -14: Shell debugging restarted -14: + bash[224]unset __lmod_vx -14: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_15 -14: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_15 -14: + interp_atmos_master.sh[9]grid_string=0p25 -14: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -14: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -14: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -14: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -14: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -14: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -14: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -14: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -14: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -14: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -14: + interp_atmos_master.sh[31]IFS=: -14: + interp_atmos_master.sh[31]read -ra grids -14: + interp_atmos_master.sh[33]output_grids= -14: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -14: + interp_atmos_master.sh[35]gridopt=grid0p25 -14: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_15_0p25' -14: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_15 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_15_0p25 -14: 1:0:d=2021032312:TMP:675 mb:26 hour fcst: -14: 2:19500:d=2021032312:RH:675 mb:26 hour fcst: -14: 3:39771:d=2021032312:TCDC:675 mb:26 hour fcst: -14: 4:51368:d=2021032312:VVEL:675 mb:26 hour fcst: -14: 5:77026:d=2021032312:DZDT:675 mb:26 hour fcst: -14: 6:104427:d=2021032312:UGRD:675 mb:26 hour fcst: -14: 7:125481:d=2021032312:VGRD:675 mb:26 hour fcst: -14: 8:147141:d=2021032312:ABSV:675 mb:26 hour fcst: -14: 9:168233:d=2021032312:CLMR:675 mb:26 hour fcst: -14: 10:173822:d=2021032312:ICMR:675 mb:26 hour fcst: -14: 11:185572:d=2021032312:RWMR:675 mb:26 hour fcst: -14: 12:190648:d=2021032312:SNMR:675 mb:26 hour fcst: -14: 13:205149:d=2021032312:GRLE:675 mb:26 hour fcst: -14: 14:207035:d=2021032312:HGT:725 mb:26 hour fcst: -14: + interp_atmos_master.sh[47]export err=0 -14: + interp_atmos_master.sh[47]err=0 -14: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -14: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -14: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_15_0p25 -14: + product_functions.sh[5]local filename=pgb2bfile_f026_15_0p25 -14: + product_functions.sh[6]wgrib2 pgb2bfile_f026_15_0p25 -not_if :RH: -grib pgb2bfile_f026_15_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_15_0p25.new -14: 1:0:d=2021032312:TMP:675 mb:26 hour fcst: -14: 2:440712:d=2021032312:RH:675 mb:26 hour fcst: -14: 3:912895:d=2021032312:TCDC:675 mb:26 hour fcst: -14: 4:1139714:d=2021032312:VVEL:675 mb:26 hour fcst: -14: 5:1820841:d=2021032312:DZDT:675 mb:26 hour fcst: -14: 6:2576385:d=2021032312:UGRD:675 mb:26 hour fcst: -14: 7:3080619:d=2021032312:VGRD:675 mb:26 hour fcst: -14: 8:3615282:d=2021032312:ABSV:675 mb:26 hour fcst: -14: 9:4107458:d=2021032312:CLMR:675 mb:26 hour fcst: -14: 10:4225110:d=2021032312:ICMR:675 mb:26 hour fcst: -14: 11:4494670:d=2021032312:RWMR:675 mb:26 hour fcst: -14: 12:4593655:d=2021032312:SNMR:675 mb:26 hour fcst: -14: 13:4912411:d=2021032312:GRLE:675 mb:26 hour fcst: -14: 14:4941290:d=2021032312:HGT:725 mb:26 hour fcst: -14: + product_functions.sh[10]rc=0 -14: + product_functions.sh[11](( rc == 0 )) -14: + product_functions.sh[11]mv pgb2bfile_f026_15_0p25.new pgb2bfile_f026_15_0p25 -14: + product_functions.sh[12]return 0 -14: + interp_atmos_master.sh[56]export err=0 -14: + interp_atmos_master.sh[56]err=0 -14: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -14: ++ interp_atmos_master.sh[62]wc -l -14: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_15_0p25 -match 'LAND|ICEC' -14: + interp_atmos_master.sh[62]var_count=0 -14: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -14: + interp_atmos_master.sh[73]exit 0 - 5: + bash[8]'[' -z '' ']' - 5: + bash[9]case "$-" in - 5: + bash[12]__lmod_vx=x - 5: + bash[16]'[' -n x ']' - 5: + bash[16]set +x - 5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) - 5: Shell debugging restarted - 5: + bash[224]unset __lmod_vx - 5: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_6 - 5: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_6 - 5: + interp_atmos_master.sh[9]grid_string=0p25 - 5: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' - 5: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' - 5: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' - 5: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' - 5: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' - 5: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' - 5: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' - 5: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' - 5: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' - 5: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh - 5: + interp_atmos_master.sh[31]IFS=: - 5: + interp_atmos_master.sh[31]read -ra grids - 5: + interp_atmos_master.sh[33]output_grids= - 5: + interp_atmos_master.sh[34]for grid in "${grids[@]}" - 5: + interp_atmos_master.sh[35]gridopt=grid0p25 - 5: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_6_0p25' - 5: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_6 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_6_0p25 - 5: 1:0:d=2021032312:TMP:225 mb:26 hour fcst: - 5: 2:18613:d=2021032312:RH:225 mb:26 hour fcst: - 5: 3:35289:d=2021032312:TCDC:225 mb:26 hour fcst: - 5: 4:44119:d=2021032312:VVEL:225 mb:26 hour fcst: - 5: 5:71656:d=2021032312:DZDT:225 mb:26 hour fcst: - 5: 6:98402:d=2021032312:UGRD:225 mb:26 hour fcst: - 5: 7:112201:d=2021032312:VGRD:225 mb:26 hour fcst: - 5: 8:126550:d=2021032312:ABSV:225 mb:26 hour fcst: - 5: 9:147370:d=2021032312:CLMR:225 mb:26 hour fcst: - 5: 10:147549:d=2021032312:ICMR:225 mb:26 hour fcst: - 5: 11:159049:d=2021032312:RWMR:225 mb:26 hour fcst: - 5: 12:159228:d=2021032312:SNMR:225 mb:26 hour fcst: - 5: 13:166081:d=2021032312:GRLE:225 mb:26 hour fcst: - 5: 14:166356:d=2021032312:HGT:275 mb:26 hour fcst: - 5: + interp_atmos_master.sh[47]export err=0 - 5: + interp_atmos_master.sh[47]err=0 - 5: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] - 5: + interp_atmos_master.sh[54]for grid in "${grids[@]}" - 5: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_6_0p25 - 5: + product_functions.sh[5]local filename=pgb2bfile_f026_6_0p25 - 5: + product_functions.sh[6]wgrib2 pgb2bfile_f026_6_0p25 -not_if :RH: -grib pgb2bfile_f026_6_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_6_0p25.new - 5: 1:0:d=2021032312:TMP:225 mb:26 hour fcst: - 5: 2:419959:d=2021032312:RH:225 mb:26 hour fcst: - 5: 3:805079:d=2021032312:TCDC:225 mb:26 hour fcst: - 5: 4:981628:d=2021032312:VVEL:225 mb:26 hour fcst: - 5: 5:1742528:d=2021032312:DZDT:225 mb:26 hour fcst: - 5: 6:2492293:d=2021032312:UGRD:225 mb:26 hour fcst: - 5: 7:2803192:d=2021032312:VGRD:225 mb:26 hour fcst: - 5: 8:3123770:d=2021032312:ABSV:225 mb:26 hour fcst: - 5: 9:3628094:d=2021032312:CLMR:225 mb:26 hour fcst: - 5: 10:3628273:d=2021032312:ICMR:225 mb:26 hour fcst: - 5: 11:3878521:d=2021032312:RWMR:225 mb:26 hour fcst: - 5: 12:3878700:d=2021032312:SNMR:225 mb:26 hour fcst: - 5: 13:4009582:d=2021032312:GRLE:225 mb:26 hour fcst: - 5: 14:4010891:d=2021032312:HGT:275 mb:26 hour fcst: - 5: + product_functions.sh[10]rc=0 - 5: + product_functions.sh[11](( rc == 0 )) - 5: + product_functions.sh[11]mv pgb2bfile_f026_6_0p25.new pgb2bfile_f026_6_0p25 - 5: + product_functions.sh[12]return 0 - 5: + interp_atmos_master.sh[56]export err=0 - 5: + interp_atmos_master.sh[56]err=0 - 5: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] - 5: ++ interp_atmos_master.sh[62]wc -l - 5: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_6_0p25 -match 'LAND|ICEC' - 5: + interp_atmos_master.sh[62]var_count=0 - 5: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] - 5: + interp_atmos_master.sh[73]exit 0 -11: + bash[8]'[' -z '' ']' -11: + bash[9]case "$-" in -11: + bash[12]__lmod_vx=x -11: + bash[16]'[' -n x ']' -11: + bash[16]set +x -11: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -11: Shell debugging restarted -11: + bash[224]unset __lmod_vx -11: + interp_atmos_master.sh[7]input_file=tmpfileb_f026_12 -11: + interp_atmos_master.sh[8]output_file_prefix=pgb2bfile_f026_12 -11: + interp_atmos_master.sh[9]grid_string=0p25 -11: + interp_atmos_master.sh[12]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -11: + interp_atmos_master.sh[13]interp_winds='-new_grid_winds earth' -11: + interp_atmos_master.sh[14]interp_bilinear='-new_grid_interpolation bilinear' -11: + interp_atmos_master.sh[15]interp_neighbor='-if :(CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -11: + interp_atmos_master.sh[16]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -11: + interp_atmos_master.sh[17]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -11: + interp_atmos_master.sh[21]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -11: + interp_atmos_master.sh[23]grid0p50='latlon 0:720:0.5 90:361:-0.5' -11: + interp_atmos_master.sh[25]grid1p00='latlon 0:360:1.0 90:181:-1.0' -11: + interp_atmos_master.sh[28]source /work2/noaa/global/mterry/global-workflow_forked/ush/product_functions.sh -11: + interp_atmos_master.sh[31]IFS=: -11: + interp_atmos_master.sh[31]read -ra grids -11: + interp_atmos_master.sh[33]output_grids= -11: + interp_atmos_master.sh[34]for grid in "${grids[@]}" -11: + interp_atmos_master.sh[35]gridopt=grid0p25 -11: + interp_atmos_master.sh[36]output_grids=' -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_12_0p25' -11: + interp_atmos_master.sh[40]wgrib2 tmpfileb_f026_12 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:1440:0.25 90:721:-0.25 pgb2bfile_f026_12_0p25 -11: 1:0:d=2021032312:TMP:525 mb:26 hour fcst: -11: 2:18793:d=2021032312:RH:525 mb:26 hour fcst: -11: 3:38469:d=2021032312:TCDC:525 mb:26 hour fcst: -11: 4:50709:d=2021032312:VVEL:525 mb:26 hour fcst: -11: 5:75812:d=2021032312:DZDT:525 mb:26 hour fcst: -11: 6:103408:d=2021032312:UGRD:525 mb:26 hour fcst: -11: 7:124588:d=2021032312:VGRD:525 mb:26 hour fcst: -11: 8:146531:d=2021032312:ABSV:525 mb:26 hour fcst: -11: 9:167521:d=2021032312:CLMR:525 mb:26 hour fcst: -11: 10:169837:d=2021032312:ICMR:525 mb:26 hour fcst: -11: 11:187068:d=2021032312:RWMR:525 mb:26 hour fcst: -11: 12:188197:d=2021032312:SNMR:525 mb:26 hour fcst: -11: 13:202612:d=2021032312:GRLE:525 mb:26 hour fcst: -11: 14:204260:d=2021032312:HGT:575 mb:26 hour fcst: -11: + interp_atmos_master.sh[47]export err=0 -11: + interp_atmos_master.sh[47]err=0 -11: + interp_atmos_master.sh[48][[ 0 -ne 0 ]] -11: + interp_atmos_master.sh[54]for grid in "${grids[@]}" -11: + interp_atmos_master.sh[55]trim_rh pgb2bfile_f026_12_0p25 -11: + product_functions.sh[5]local filename=pgb2bfile_f026_12_0p25 -11: + product_functions.sh[6]wgrib2 pgb2bfile_f026_12_0p25 -not_if :RH: -grib pgb2bfile_f026_12_0p25.new -if :RH: -rpn '10:*:0.5:+:floor:1000:min:10:/' -set_grib_type same -set_scaling -1 0 -grib_out pgb2bfile_f026_12_0p25.new -11: 1:0:d=2021032312:TMP:525 mb:26 hour fcst: -11: 2:429878:d=2021032312:RH:525 mb:26 hour fcst: -11: 3:878829:d=2021032312:TCDC:525 mb:26 hour fcst: -11: 4:1109397:d=2021032312:VVEL:525 mb:26 hour fcst: -11: 5:1780483:d=2021032312:DZDT:525 mb:26 hour fcst: -11: 6:2557754:d=2021032312:UGRD:525 mb:26 hour fcst: -11: 7:3075023:d=2021032312:VGRD:525 mb:26 hour fcst: -11: 8:3620494:d=2021032312:ABSV:525 mb:26 hour fcst: -11: 9:4119873:d=2021032312:CLMR:525 mb:26 hour fcst: -11: 10:4167159:d=2021032312:ICMR:525 mb:26 hour fcst: -11: 11:4554118:d=2021032312:RWMR:525 mb:26 hour fcst: -11: 12:4571830:d=2021032312:SNMR:525 mb:26 hour fcst: -11: 13:4884934:d=2021032312:GRLE:525 mb:26 hour fcst: -11: 14:4910511:d=2021032312:HGT:575 mb:26 hour fcst: -11: + product_functions.sh[10]rc=0 -11: + product_functions.sh[11](( rc == 0 )) -11: + product_functions.sh[11]mv pgb2bfile_f026_12_0p25.new pgb2bfile_f026_12_0p25 -11: + product_functions.sh[12]return 0 -11: + interp_atmos_master.sh[56]export err=0 -11: + interp_atmos_master.sh[56]err=0 -11: + interp_atmos_master.sh[57][[ 0 -ne 0 ]] -11: ++ interp_atmos_master.sh[62]wc -l -11: ++ interp_atmos_master.sh[62]wgrib2 pgb2bfile_f026_12_0p25 -match 'LAND|ICEC' -11: + interp_atmos_master.sh[62]var_count=0 -11: + interp_atmos_master.sh[63][[ 0 -eq 2 ]] -11: + interp_atmos_master.sh[73]exit 0 -+ run_mpmd.sh[113]exit 0 -+ run_mpmd.sh[1]postamble run_mpmd.sh 1753756111 0 -+ preamble.sh[62]set +x -End run_mpmd.sh at 02:28:46 with error code 0 (time elapsed: 00:00:15) -+ exglobal_atmos_products.sh[142]true -+ exglobal_atmos_products.sh[143]export err=0 -+ exglobal_atmos_products.sh[143]err=0 -+ exglobal_atmos_products.sh[144][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[149]mv mpmd.out mpmd_2.out -+ exglobal_atmos_products.sh[153]echo 'Concatenating processor-specific grib2 files into a single product file' -Concatenating processor-specific grib2 files into a single product file -+ exglobal_atmos_products.sh[154](( iproc = 1 )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_1_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_1_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_1 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_2_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_2_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_2 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_3_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_3_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_3 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_4_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_4_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_4 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_5_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_5_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_5 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_6_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_6_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_6 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_7_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_7_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_7 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_8_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_8_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_8 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_9_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_9_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_9 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_10_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_10_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_10 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_11_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_11_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_11 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_12_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_12_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_12 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_13_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_13_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_13 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_14_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_14_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_14 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_15_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_15_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_15 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_16_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_16_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_16 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_17_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_17_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_17 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_18_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_18_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_18 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_19_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_19_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_19 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_20_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_20_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_20 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_21_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_21_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_21 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_22_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_22_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_22 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_23_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_23_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_23 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[155]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[156]cat pgb2bfile_f026_24_0p25 -+ exglobal_atmos_products.sh[157]rm -f pgb2bfile_f026_24_0p25 -+ exglobal_atmos_products.sh[160]rm -f tmpfileb_f026_24 -+ exglobal_atmos_products.sh[154](( iproc++ )) -+ exglobal_atmos_products.sh[154](( iproc <= nproc )) -+ exglobal_atmos_products.sh[164]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[165]prod_dir=COMOUT_ATMOS_GRIB_0p25 -+ exglobal_atmos_products.sh[166]cpfs pgb2bfile_f026_0p25 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f026 -+ cpfs[3]'[' 2 -ne 2 ']' -+ cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f026 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f026 = ./ ']' -+ cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f026 ']' -+ cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f026 -+ cpfs[16]cp pgb2bfile_f026_0p25 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f026.cptmp -+ cpfs[18]'[' 0 -ne 0 ']' -+ cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f026.cptmp -+ cpfs[23]'[' 0 -ne 0 ']' -+ cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f026.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2b.0p25.f026 -+ cpfs[28]'[' 0 -ne 0 ']' -+ exglobal_atmos_products.sh[167]wgrib2 -s pgb2bfile_f026_0p25 -+ exglobal_atmos_products.sh[170]echo 'Finished processing nset = 2' -Finished processing nset = 2 -+ exglobal_atmos_products.sh[154](( nset++ )) -+ exglobal_atmos_products.sh[154](( nset <= downset )) -+ exglobal_atmos_products.sh[177]FLUX_FILE=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.sfluxgrbf026.grib2 -+ exglobal_atmos_products.sh[178][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.sfluxgrbf026.grib2 ]] -+ exglobal_atmos_products.sh[179]wgrib2 -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.sfluxgrbf026.grib2 -+ exglobal_atmos_products.sh[185][[ YES == \Y\E\S ]] -+ exglobal_atmos_products.sh[188]input_file=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.sfluxgrbf026.grib2 -+ exglobal_atmos_products.sh[189]output_file_prefix=sflux_f026 -+ exglobal_atmos_products.sh[190]grid_string=1p00 -+ exglobal_atmos_products.sh[191]/work2/noaa/global/mterry/global-workflow_forked/ush/interp_atmos_sflux.sh /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.sfluxgrbf026.grib2 sflux_f026 1p00 -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ interp_atmos_sflux.sh[6]input_file=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.sfluxgrbf026.grib2 -+ interp_atmos_sflux.sh[7]output_file_prefix=sflux_f026 -+ interp_atmos_sflux.sh[8]grid_string=1p00 -+ interp_atmos_sflux.sh[11]defaults='-set_grib_type same -set_bitmap 1 -set_grib_max_bits 16' -+ interp_atmos_sflux.sh[12]interp_winds='-new_grid_winds earth' -+ interp_atmos_sflux.sh[13]interp_bilinear='-new_grid_interpolation bilinear' -+ interp_atmos_sflux.sh[14]interp_neighbor='-if :(LAND|CSNOW|CRAIN|CFRZR|CICEP|ICSEV): -new_grid_interpolation neighbor -fi' -+ interp_atmos_sflux.sh[15]interp_budget='-if :(APCP|ACPCP|PRATE|CPRAT|DZDT): -new_grid_interpolation budget -fi' -+ interp_atmos_sflux.sh[16]increased_bits='-if :(APCP|ACPCP|PRATE|CPRAT): -set_grib_max_bits 25 -fi' -+ interp_atmos_sflux.sh[20]grid0p25='latlon 0:1440:0.25 90:721:-0.25' -+ interp_atmos_sflux.sh[22]grid0p50='latlon 0:720:0.5 90:361:-0.5' -+ interp_atmos_sflux.sh[24]grid1p00='latlon 0:360:1.0 90:181:-1.0' -+ interp_atmos_sflux.sh[27]IFS=: -+ interp_atmos_sflux.sh[27]read -ra grids -+ interp_atmos_sflux.sh[29]output_grids= -+ interp_atmos_sflux.sh[30]for grid in "${grids[@]}" -+ interp_atmos_sflux.sh[31]gridopt=grid1p00 -+ interp_atmos_sflux.sh[32]output_grids=' -new_grid latlon 0:360:1.0 90:181:-1.0 sflux_f026_1p00' -+ interp_atmos_sflux.sh[36]wgrib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/atmos/master/gfs.t12z.sfluxgrbf026.grib2 -set_grib_type same -set_bitmap 1 -set_grib_max_bits 16 -new_grid_winds earth -new_grid_interpolation bilinear -if ':(LAND|CSNOW|CRAIN|CFRZR|CICEP|ICSEV):' -new_grid_interpolation neighbor -fi -if ':(APCP|ACPCP|PRATE|CPRAT|DZDT):' -new_grid_interpolation budget -fi -if ':(APCP|ACPCP|PRATE|CPRAT):' -set_grib_max_bits 25 -fi -new_grid latlon 0:360:1.0 90:181:-1.0 sflux_f026_1p00 -1:0:d=2021032312:HGT:1 hybrid level:26 hour fcst: -2:11758:d=2021032312:TMP:1 hybrid level:26 hour fcst: -3:32624:d=2021032312:SPFH:1 hybrid level:26 hour fcst: -4:61439:d=2021032312:UGRD:1 hybrid level:26 hour fcst: -5:81752:d=2021032312:VGRD:1 hybrid level:26 hour fcst: -6:102514:d=2021032312:PRES:surface:26 hour fcst: -7:128219:d=2021032312:HGT:surface:26 hour fcst: -8:144289:d=2021032312:TMP:surface:26 hour fcst: -9:160925:d=2021032312:TSOIL:0-0.1 m below ground:26 hour fcst: -10:172893:d=2021032312:SOILW:0-0.1 m below ground:26 hour fcst: -11:181810:d=2021032312:SOILL:0-0.1 m below ground:26 hour fcst: -12:190610:d=2021032312:TSOIL:0.1-0.4 m below ground:26 hour fcst: -13:206103:d=2021032312:SOILW:0.1-0.4 m below ground:26 hour fcst: -14:215117:d=2021032312:SOILL:0.1-0.4 m below ground:26 hour fcst: -15:224009:d=2021032312:TSOIL:0.4-1 m below ground:26 hour fcst: -16:238458:d=2021032312:SOILW:0.4-1 m below ground:26 hour fcst: -17:247591:d=2021032312:SOILL:0.4-1 m below ground:26 hour fcst: -18:256514:d=2021032312:TSOIL:1-2 m below ground:26 hour fcst: -19:270804:d=2021032312:SOILW:1-2 m below ground:26 hour fcst: -20:279861:d=2021032312:SOILL:1-2 m below ground:26 hour fcst: -21:288938:d=2021032312:CISOILM:0-2 m below ground:26 hour fcst: -22:299446:d=2021032312:CNWAT:surface:26 hour fcst: -23:304968:d=2021032312:WEASD:surface:26 hour fcst: -24:317793:d=2021032312:SNOWC:surface:24-26 hour ave fcst: -25:322732:d=2021032312:SNOD:surface:26 hour fcst: -26:337111:d=2021032312:PEVPR:surface:26 hour fcst: -27:352903:d=2021032312:ICETK:surface:26 hour fcst: -28:356311:d=2021032312:ACOND:surface:26 hour fcst: -29:373844:d=2021032312:EVCW:surface:24-26 hour ave fcst: -30:383547:d=2021032312:EVBS:surface:24-26 hour ave fcst: -31:397974:d=2021032312:TRANS:surface:24-26 hour ave fcst: -32:406747:d=2021032312:SBSNO:surface:24-26 hour ave fcst: -33:414761:d=2021032312:var discipline=2 center=7 local_table=1 parmcat=0 parm=233:surface:26 hour fcst: -34:422782:d=2021032312:var discipline=2 center=7 local_table=1 parmcat=0 parm=235:surface:26 hour fcst: -35:426967:d=2021032312:var discipline=2 center=7 local_table=1 parmcat=0 parm=236:surface:26 hour fcst: -36:433657:d=2021032312:var discipline=2 center=7 local_table=1 parmcat=0 parm=237:surface:24-26 hour acc fcst: -37:439523:d=2021032312:var discipline=2 center=7 local_table=1 parmcat=0 parm=238:surface:24-26 hour acc fcst: -38:445602:d=2021032312:var discipline=2 center=7 local_table=1 parmcat=0 parm=239:surface:24-26 hour acc fcst: -39:451980:d=2021032312:var discipline=2 center=7 local_table=1 parmcat=0 parm=235:surface:24-26 hour ave fcst: -40:456207:d=2021032312:TMP:2 m above ground:26 hour fcst: -41:477027:d=2021032312:SPFH:2 m above ground:26 hour fcst: -42:505841:d=2021032312:TMAX:2 m above ground:24-26 hour max fcst: -43:526577:d=2021032312:TMIN:2 m above ground:24-26 hour min fcst: -44:547369:d=2021032312:QMAX:2 m above ground:24-26 hour max fcst: -45:564589:d=2021032312:QMIN:2 m above ground:24-26 hour min fcst: -46:581844:d=2021032312:UGRD:10 m above ground:26 hour fcst: -47:602389:d=2021032312:VGRD:10 m above ground:26 hour fcst: -48:623307:d=2021032312:CPOFP:surface:26 hour fcst: -49:643653:d=2021032312:CPRAT:surface:24-26 hour ave fcst: -50:667296:d=2021032312:PRATE:surface:24-26 hour ave fcst: -51:692427:d=2021032312:SSRUN:surface:24-26 hour acc fcst: -52:699479:d=2021032312:WATR:surface:24-26 hour acc fcst: -53:712223:d=2021032312:LHTFL:surface:24-26 hour ave fcst: -54:734022:d=2021032312:SHTFL:surface:24-26 hour ave fcst: -55:754817:d=2021032312:GFLUX:surface:24-26 hour ave fcst: -56:766882:d=2021032312:SNOHF:surface:24-26 hour ave fcst: -57:770362:d=2021032312:UFLX:surface:24-26 hour ave fcst: -58:786775:d=2021032312:VFLX:surface:24-26 hour ave fcst: -59:803275:d=2021032312:SFCR:surface:26 hour fcst: -60:821690:d=2021032312:FRICV:surface:26 hour fcst: -61:843429:d=2021032312:U-GWD:surface:24-26 hour ave fcst: -62:858344:d=2021032312:V-GWD:surface:24-26 hour ave fcst: -63:873309:d=2021032312:SHTFL:surface:26 hour fcst: -64:894059:d=2021032312:LHTFL:surface:26 hour fcst: -65:915860:d=2021032312:SFEXC:surface:26 hour fcst: -66:932608:d=2021032312:VEG:surface:26 hour fcst: -67:940642:d=2021032312:GFLUX:surface:26 hour fcst: -68:952604:d=2021032312:VGTYP:surface:26 hour fcst: -69:961137:d=2021032312:SOTYP:surface:26 hour fcst: -70:969096:d=2021032312:SLTYP:surface:26 hour fcst: -71:979308:d=2021032312:WILT:surface:26 hour fcst: -72:989218:d=2021032312:FLDCP:surface:26 hour fcst: -73:999424:d=2021032312:SUNSD:surface:26 hour fcst: -74:1009311:d=2021032312:PEVPR:surface:24-26 hour ave fcst: -75:1025055:d=2021032312:PWAT:entire atmosphere (considered as a single layer):26 hour fcst: -76:1051969:d=2021032312:LCDC:low cloud layer:24-26 hour ave fcst: -77:1070665:d=2021032312:MCDC:middle cloud layer:24-26 hour ave fcst: -78:1086596:d=2021032312:HCDC:high cloud layer:24-26 hour ave fcst: -79:1103020:d=2021032312:TCDC:entire atmosphere (considered as a single layer):24-26 hour ave fcst: -80:1121082:d=2021032312:PRES:convective cloud bottom level:26 hour fcst: -81:1134720:d=2021032312:PRES:low cloud bottom level:24-26 hour ave fcst: -82:1156329:d=2021032312:PRES:middle cloud bottom level:24-26 hour ave fcst: -83:1173132:d=2021032312:PRES:high cloud bottom level:24-26 hour ave fcst: -84:1190568:d=2021032312:PRES:convective cloud top level:26 hour fcst: -85:1205894:d=2021032312:PRES:low cloud top level:24-26 hour ave fcst: -86:1227412:d=2021032312:PRES:middle cloud top level:24-26 hour ave fcst: -87:1244057:d=2021032312:PRES:high cloud top level:24-26 hour ave fcst: -88:1261988:d=2021032312:TMP:low cloud top level:24-26 hour ave fcst: -89:1278214:d=2021032312:TMP:middle cloud top level:24-26 hour ave fcst: -90:1291516:d=2021032312:TMP:high cloud top level:24-26 hour ave fcst: -91:1306457:d=2021032312:TCDC:convective cloud layer:26 hour fcst: -92:1326901:d=2021032312:TCDC:boundary layer cloud layer:24-26 hour ave fcst: -93:1343038:d=2021032312:CWORK:entire atmosphere (considered as a single layer):24-26 hour ave fcst: -94:1351608:d=2021032312:DSWRF:surface:24-26 hour ave fcst: -95:1370292:d=2021032312:DUVB:surface:24-26 hour ave fcst: -96:1387722:d=2021032312:CDUVB:surface:24-26 hour ave fcst: -97:1401442:d=2021032312:DLWRF:surface:24-26 hour ave fcst: -98:1423250:d=2021032312:USWRF:surface:24-26 hour ave fcst: -99:1441418:d=2021032312:ULWRF:surface:24-26 hour ave fcst: -100:1460301:d=2021032312:USWRF:top of atmosphere:24-26 hour ave fcst: -101:1479307:d=2021032312:ULWRF:top of atmosphere:24-26 hour ave fcst: -102:1500399:d=2021032312:DSWRF:surface:26 hour fcst: -103:1517551:d=2021032312:DLWRF:surface:26 hour fcst: -104:1539418:d=2021032312:USWRF:surface:26 hour fcst: -105:1555984:d=2021032312:ULWRF:surface:26 hour fcst: -106:1574726:d=2021032312:CSDSF:surface:24-26 hour ave fcst: -107:1588985:d=2021032312:CSUSF:surface:24-26 hour ave fcst: -108:1605390:d=2021032312:CSUSF:top of atmosphere:24-26 hour ave fcst: -109:1622090:d=2021032312:DSWRF:top of atmosphere:24-26 hour ave fcst: -110:1631079:d=2021032312:CSDLF:surface:24-26 hour ave fcst: -111:1649289:d=2021032312:CSULF:surface:24-26 hour ave fcst: -112:1668008:d=2021032312:CSULF:top of atmosphere:24-26 hour ave fcst: -113:1684781:d=2021032312:VBDSF:surface:24-26 hour ave fcst: -114:1702109:d=2021032312:VDDSF:surface:24-26 hour ave fcst: -115:1720787:d=2021032312:NBDSF:surface:24-26 hour ave fcst: -116:1739140:d=2021032312:NDDSF:surface:24-26 hour ave fcst: -117:1758096:d=2021032312:HPBL:surface:26 hour fcst: -118:1789171:d=2021032312:LAND:surface:26 hour fcst: -119:1791112:d=2021032312:ICEC:surface:26 hour fcst: -120:1794251:d=2021032312:ALBDO:surface:24-26 hour ave fcst: -121:1809866:d=2021032312:FDNSSTMP:surface:26 hour fcst: -+ interp_atmos_sflux.sh[43]export err=0 -+ interp_atmos_sflux.sh[43]err=0 -+ interp_atmos_sflux.sh[44][[ 0 -ne 0 ]] -+ interp_atmos_sflux.sh[49]exit 0 -+ exglobal_atmos_products.sh[191]true -+ exglobal_atmos_products.sh[192]export err=0 -+ exglobal_atmos_products.sh[192]err=0 -+ exglobal_atmos_products.sh[193][[ 0 -ne 0 ]] -+ exglobal_atmos_products.sh[198]IFS=: -+ exglobal_atmos_products.sh[198]read -ra grids -+ exglobal_atmos_products.sh[199]for grid in "${grids[@]}" -+ exglobal_atmos_products.sh[200]prod_dir=COMOUT_ATMOS_GRIB_1p00 -+ exglobal_atmos_products.sh[201]cpfs sflux_f026_1p00 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f026 -+ cpfs[3]'[' 2 -ne 2 ']' -+ cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f026 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f026 = ./ ']' -+ cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f026 ']' -+ cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f026 -+ cpfs[16]cp sflux_f026_1p00 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f026.cptmp -+ cpfs[18]'[' 0 -ne 0 ']' -+ cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f026.cptmp -+ cpfs[23]'[' 0 -ne 0 ']' -+ cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f026.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/1p00/gfs.t12z.flux.1p00.f026 -+ cpfs[28]'[' 0 -ne 0 ']' -+ exglobal_atmos_products.sh[202]wgrib2 -s sflux_f026_1p00 -+ exglobal_atmos_products.sh[207][[ YES == \Y\E\S ]] -+ exglobal_atmos_products.sh[208]grp= -+ exglobal_atmos_products.sh[209][[ 26 -gt 0 ]] -+ exglobal_atmos_products.sh[209][[ 26 -le 180 ]] -+ exglobal_atmos_products.sh[211]wgrib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.pgrb2.0p25.f026 -d 597 -grib /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/atmos/grib2/0p25/gfs.t12z.wgne.f026 -597:256252481:d=2021032312:APCP:surface:24-26 hour acc fcst: -+ exglobal_atmos_products.sh[219][[ NO == \Y\E\S ]] -+ exglobal_atmos_products.sh[290]exit 0 -+ JGLOBAL_ATMOS_PRODUCTS[27]true -+ JGLOBAL_ATMOS_PRODUCTS[28]export err=0 -+ JGLOBAL_ATMOS_PRODUCTS[28]err=0 -+ JGLOBAL_ATMOS_PRODUCTS[29][[ 0 -ne 0 ]] -+ JGLOBAL_ATMOS_PRODUCTS[40][[ -e OUTPUT.559401 ]] -+ JGLOBAL_ATMOS_PRODUCTS[47]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312 -+ JGLOBAL_ATMOS_PRODUCTS[48][[ NO == \N\O ]] -+ JGLOBAL_ATMOS_PRODUCTS[49]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/atmos_products_f026.533601 -+ JGLOBAL_ATMOS_PRODUCTS[53]exit 0 -+ JGLOBAL_ATMOS_PRODUCTS[1]postamble /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_ATMOS_PRODUCTS 1753756051 0 -+ preamble.sh[62]set +x -End /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_ATMOS_PRODUCTS at 02:28:52 with error code 0 (time elapsed: 00:01:21) -+ status=0 -+ [[ 0 -ne 0 ]] -+ exit 0 diff --git a/ci/error_logs/HECULES_PR_LOGS/C48_S2SW_logs_2021032312_gfs_cleanup.log b/ci/error_logs/HECULES_PR_LOGS/C48_S2SW_logs_2021032312_gfs_cleanup.log deleted file mode 100644 index 7c074199..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/C48_S2SW_logs_2021032312_gfs_cleanup.log +++ /dev/null @@ -1,1152 +0,0 @@ -+ . /work2/noaa/global/mterry/global-workflow_forked/ush/load_fv3gfs_modules.sh -++ [[ hxB == *x* ]] -++ set_x=YES -++ [[ NO == \N\O ]] -++ echo 'Loading modules quietly...' -Loading modules quietly... -++ set +x -Running "module reset". Resetting modules to system default. The following $MODULEPATH directories have been removed: None - -Currently Loaded Modules: - 1) contrib/0.1 43) ip/4.3.0 - 2) intel-oneapi-compilers/2023.1.0 44) grib-util/1.3.0 - 3) stack-intel/2021.9.0 45) g2tmpl/1.13.0 - 4) intel-oneapi-mpi/2021.9.0 46) gsi-ncdiag/1.1.2 - 5) stack-intel-oneapi-mpi/2021.9.0 47) crtm-fix/2.4.0.1_emc - 6) intel-oneapi-mkl/2023.1.0 48) git-lfs/3.1.2 - 7) zlib/1.2.13 49) crtm/2.4.0.1 - 8) pigz/2.7 50) openblas/0.3.24 - 9) zstd/1.5.2 51) py-setuptools/63.4.3 - 10) tar/1.34 52) py-numpy/1.23.4 - 11) gettext/0.21.1 53) bufr/11.7.0 - 12) libxcrypt/4.4.35 54) gmake/4.2.1 - 13) sqlite/3.43.2 55) wgrib2/2.0.8 - 14) util-linux-uuid/2.38.1 56) py-f90nml/1.4.3 - 15) python/3.11.6 57) py-cftime/1.0.3.4 - 16) libjpeg/2.1.0 58) py-netcdf4/1.5.8 - 17) jasper/2.0.32 59) libyaml/0.2.5 - 18) libpng/1.6.37 60) py-pyyaml/6.0 - 19) openjpeg/2.3.1 61) py-markupsafe/2.1.3 - 20) eccodes/2.32.0 62) py-jinja2/3.1.2 - 21) fftw/3.3.10 63) py-bottleneck/1.3.7 - 22) nghttp2/1.57.0 64) py-numexpr/2.8.4 - 23) curl/8.4.0 65) py-et-xmlfile/1.0.1 - 24) proj/9.2.1 66) py-openpyxl/3.1.2 - 25) udunits/2.2.28 67) py-pytz/2023.3 - 26) cdo/2.2.0 68) py-pyxlsb/1.0.10 - 27) hdf5/1.14.0 69) py-xlrd/2.0.1 - 28) snappy/1.1.10 70) py-xlsxwriter/3.1.7 - 29) c-blosc/1.21.5 71) py-xlwt/1.3.0 - 30) netcdf-c/4.9.2 72) py-pandas/1.5.3 - 31) netcdf-fortran/4.6.1 73) py-six/1.16.0 - 32) parallel-netcdf/1.12.2 74) py-python-dateutil/2.8.2 - 33) parallelio/2.5.10 75) g2c/1.8.0 - 34) esmf/8.6.0 76) netcdf-cxx4/4.3.1 - 35) antlr/2.7.7 77) met/9.1.3 - 36) gsl/2.7.1 78) metplus/3.1.1 - 37) nco/5.0.6 79) py-packaging/23.1 - 38) bacio/2.4.1 80) py-xarray/2023.7.0 - 39) w3emc/2.10.0 81) prepobs/1.1.0 - 40) prod_util/2.1.1 82) fit2obs/1.1.7.1 - 41) g2/3.4.5 83) globus-cli/3.35.2 - 42) sp/2.5.0 84) module_base.hercules - - - -++ [[ -d /work2/noaa/global/mterry/global-workflow_forked/sorc/wxflow/src ]] -++ PYTHONPATH=/apps/other/globus-cli-3.35.2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:/work2/noaa/global/mterry/global-workflow_forked/sorc/wxflow/src -++ PYTHONPATH=/apps/other/globus-cli-3.35.2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:/work2/noaa/global/mterry/global-workflow_forked/sorc/wxflow/src:/work2/noaa/global/mterry/global-workflow_forked/ush/python -++ export PYTHONPATH -++ ulimit -S -s 16384 -++ unset ulimit_s -+ status=0 -+ [[ 0 -ne 0 ]] -+ export job=cleanup -+ job=cleanup -+ export jobid=cleanup.3752306 -+ jobid=cleanup.3752306 -+ /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_CLEANUP -Begin /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_CLEANUP at Tue Jul 29 03:16:26 UTC 2025 -++ jjob_header.sh[46]OPTIND=1 -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[50]env_job=cleanup -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[49]read -ra configs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[61]shift 4 -++ jjob_header.sh[63][[ -z cleanup ]] -++ jjob_header.sh[71]export DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/cleanup.3752306 -++ jjob_header.sh[71]DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/cleanup.3752306 -++ jjob_header.sh[72][[ YES == \Y\E\S ]] -++ jjob_header.sh[73]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/cleanup.3752306 -++ jjob_header.sh[75]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/cleanup.3752306 -++ jjob_header.sh[76]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/cleanup.3752306 -++ jjob_header.sh[85]export pid=3752576 -++ jjob_header.sh[85]pid=3752576 -++ jjob_header.sh[86]export pgmout=OUTPUT.3752576 -++ jjob_header.sh[86]pgmout=OUTPUT.3752576 -++ jjob_header.sh[87]export pgmerr=errfile -++ jjob_header.sh[87]pgmerr=errfile -++ jjob_header.sh[90]export pgm= -++ jjob_header.sh[90]pgm= -++ jjob_header.sh[96]export cycle=t12z -++ jjob_header.sh[96]cycle=t12z -++ jjob_header.sh[97]setpdy.sh -+ setpdy.sh[20]'[' /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/cleanup.3752306 == /home/mterry ']' -+ setpdy.sh[25][[ ! t12z =~ t??z ]] -+ setpdy.sh[30]case $# in -+ setpdy.sh[31]dates_before_PDY=7 -+ setpdy.sh[32]dates_after_PDY=7 -+ setpdy.sh[50]COMDATEROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+ setpdy.sh[53]'[' -z 20210323 ']' -+ setpdy.sh[57]sed 's/[0-9]\{8\}/20210323/' /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z -sed: can't read /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z: No such file or directory -++ jjob_header.sh[97]true -++ jjob_header.sh[98]source ./PDY -/work2/noaa/global/mterry/global-workflow_forked/ush/jjob_header.sh: line 98: ./PDY: No such file or directory -++ jjob_header.sh[98]true -++ jjob_header.sh[104]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[104]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.base -+++ config.base[6]echo 'BEGIN: config.base' -BEGIN: config.base -+++ config.base[9]export machine=HERCULES -+++ config.base[9]machine=HERCULES -+++ config.base[12]export RUN_ENVIR=emc -+++ config.base[12]RUN_ENVIR=emc -+++ config.base[15]export ACCOUNT=fv3-cpu -+++ config.base[15]ACCOUNT=fv3-cpu -+++ config.base[16]export QUEUE=batch -+++ config.base[16]QUEUE=batch -+++ config.base[17]export QUEUE_SERVICE=batch -+++ config.base[17]QUEUE_SERVICE=batch -+++ config.base[18]export QUEUE_DTN=batch -+++ config.base[18]QUEUE_DTN=batch -+++ config.base[19]export PARTITION_BATCH=hercules -+++ config.base[19]PARTITION_BATCH=hercules -+++ config.base[20]export PARTITION_SERVICE=service -+++ config.base[20]PARTITION_SERVICE=service -+++ config.base[21]export PARTITION_DTN= -+++ config.base[21]PARTITION_DTN= -+++ config.base[22]export RESERVATION= -+++ config.base[22]RESERVATION= -+++ config.base[23]export CLUSTERS= -+++ config.base[23]CLUSTERS= -+++ config.base[24]export CLUSTERS_SERVICE= -+++ config.base[24]CLUSTERS_SERVICE= -+++ config.base[25]export CLUSTERS_DTN= -+++ config.base[25]CLUSTERS_DTN= -+++ config.base[28]export HPSS_PROJECT=emc-global -+++ config.base[28]HPSS_PROJECT=emc-global -+++ config.base[31]export HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[31]HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[32]export EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[32]EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[33]export FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[33]FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[34]export PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[34]PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[35]export SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[35]SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[36]export USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[36]USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[38]export FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[38]FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[39]export FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[39]FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[40]export FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[40]FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[41]export FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[41]FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[42]export FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[42]FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[43]export FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[43]FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[44]export FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[44]FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[45]export FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[45]FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[50]export PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[50]PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[51]export COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[51]COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[52]export COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[52]COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[53]export DMPDIR=/work/noaa/rstprod/dump -+++ config.base[53]DMPDIR=/work/noaa/rstprod/dump -+++ config.base[57]export COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[57]COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[58]export COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[58]COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[59]export COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[59]COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[62]export HOMEDIR=/work2/noaa/global/mterry -+++ config.base[62]HOMEDIR=/work2/noaa/global/mterry -+++ config.base[63]export STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[63]STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]export PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[65]export NOSCRUB=/work2/noaa/global/mterry -+++ config.base[65]NOSCRUB=/work2/noaa/global/mterry -+++ config.base[68]export BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[68]BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[71]export BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[71]BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[74]export DO_PREP_SFC=NO -+++ config.base[74]DO_PREP_SFC=NO -+++ config.base[77]export DO_GOES=NO -+++ config.base[77]DO_GOES=NO -+++ config.base[78]export DO_BUFRSND=NO -+++ config.base[78]DO_BUFRSND=NO -+++ config.base[79]export DO_GEMPAK=NO -+++ config.base[79]DO_GEMPAK=NO -+++ config.base[80]export DO_AWIPS=NO -+++ config.base[80]DO_AWIPS=NO -+++ config.base[81]export DO_NPOESS=NO -+++ config.base[81]DO_NPOESS=NO -+++ config.base[82]export DO_TRACKER=YES -+++ config.base[82]DO_TRACKER=YES -+++ config.base[83]export DO_GENESIS=YES -+++ config.base[83]DO_GENESIS=YES -+++ config.base[84]export DO_GENESIS_FSU=NO -+++ config.base[84]DO_GENESIS_FSU=NO -+++ config.base[85]export DO_VERFOZN=YES -+++ config.base[85]DO_VERFOZN=YES -+++ config.base[86]export DO_VERFRAD=YES -+++ config.base[86]DO_VERFRAD=YES -+++ config.base[87]export DO_VMINMON=YES -+++ config.base[87]DO_VMINMON=YES -+++ config.base[88]export DO_ANLSTAT=NO -+++ config.base[88]DO_ANLSTAT=NO -+++ config.base[91]export MODE=forecast-only -+++ config.base[91]MODE=forecast-only -+++ config.base[92]export DO_TEST_MODE=YES -+++ config.base[92]DO_TEST_MODE=YES -+++ config.base[101]export FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[101]FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[102]export HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[102]HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[103]export HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v1.2.0 -+++ config.base[103]HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v1.2.0 -+++ config.base[106]export NMV=/bin/mv -+++ config.base[106]NMV=/bin/mv -+++ config.base[107]export 'NLN=/bin/ln -sf' -+++ config.base[107]NLN='/bin/ln -sf' -+++ config.base[108]export VERBOSE=YES -+++ config.base[108]VERBOSE=YES -+++ config.base[109]export KEEPDATA=NO -+++ config.base[109]KEEPDATA=NO -+++ config.base[110]export DEBUG_POSTSCRIPT=NO -+++ config.base[110]DEBUG_POSTSCRIPT=NO -+++ config.base[111]export CHGRP_RSTPROD=YES -+++ config.base[111]CHGRP_RSTPROD=YES -+++ config.base[112]export 'CHGRP_CMD=chgrp rstprod' -+++ config.base[112]CHGRP_CMD='chgrp rstprod' -+++ config.base[113]export NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[113]NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[114]export NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[114]NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[117]export BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[117]BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[120]export SDATE=2021032312 -+++ config.base[120]SDATE=2021032312 -+++ config.base[121]export EDATE=2021032312 -+++ config.base[121]EDATE=2021032312 -+++ config.base[122]export EXP_WARM_START=.false. -+++ config.base[122]EXP_WARM_START=.false. -+++ config.base[123]export assim_freq=6 -+++ config.base[123]assim_freq=6 -+++ config.base[124]export PSLOT=C48_S2SW -+++ config.base[124]PSLOT=C48_S2SW -+++ config.base[125]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[125]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[126]export ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[126]ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[127]export DUMP_SUFFIX= -+++ config.base[127]DUMP_SUFFIX= -+++ config.base[128][[ 2021032312 -ge 2019092100 ]] -+++ config.base[128][[ 2021032312 -le 2019110700 ]] -+++ config.base[131]export ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[131]ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[132]export ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[132]ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[133]export FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[133]FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[136]export envir=prod -+++ config.base[136]envir=prod -+++ config.base[137]export NET=gfs -+++ config.base[137]NET=gfs -+++ config.base[138]export RUN=gfs -+++ config.base[138]RUN=gfs -+++ config.base[141]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.com -++++ config.com[4]echo 'BEGIN: config.com' -BEGIN: config.com -++++ config.com[38][[ emc == \n\c\o ]] -++++ config.com[43]COM_OBSPROC_TMPL='${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos' -++++ config.com[44]COM_RTOFS_TMPL='${DMPDIR}' -++++ config.com[45]COM_TCVITAL_TMPL='${DMPDIR}/${RUN}.${YMD}/${HH}/atmos' -++++ config.com[47]declare -rx 'COM_OBS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/obs' -++++ config.com[48]declare -rx COM_OBSPROC_TMPL COM_RTOFS_TMPL -++++ config.com[50]COM_BASE='${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}' -++++ config.com[52]declare -rx 'COM_TOP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}' -++++ config.com[54]declare -rx 'COM_CONF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf' -++++ config.com[55]declare -rx 'COM_OBS_JEDI=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi' -++++ config.com[57]declare -rx 'COM_ATMOS_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input' -++++ config.com[58]declare -rx 'COM_ATMOS_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart' -++++ config.com[59]declare -rx 'COM_ATMOS_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos' -++++ config.com[60]declare -rx 'COM_SNOW_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow' -++++ config.com[61]declare -rx 'COM_SNOW_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon' -++++ config.com[62]declare -rx 'COM_ATMOS_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history' -++++ config.com[63]declare -rx 'COM_ATMOS_MASTER_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master' -++++ config.com[64]declare -rx 'COM_ATMOS_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2' -++++ config.com[65]declare -rx 'COM_ATMOS_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}' -++++ config.com[66]declare -rx 'COM_ATMOS_BUFR_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr' -++++ config.com[67]declare -rx 'COM_ATMOS_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}' -++++ config.com[68]declare -rx 'COM_ATMOS_GENESIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital' -++++ config.com[69]declare -rx 'COM_ATMOS_TRACK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks' -++++ config.com[70]declare -rx 'COM_ATMOS_GOES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim' -++++ config.com[71]declare -rx 'COM_ATMOS_IMAGERY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery' -++++ config.com[72]declare -rx 'COM_ATMOS_OZNMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon' -++++ config.com[73]declare -rx 'COM_ATMOS_RADMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon' -++++ config.com[74]declare -rx 'COM_ATMOS_MINMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon' -++++ config.com[75]declare -rx 'COM_ATMOS_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon' -++++ config.com[76]declare -rx 'COM_ATMOS_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo' -++++ config.com[78]declare -rx 'COM_WAVE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart' -++++ config.com[79]declare -rx 'COM_WAVE_PREP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep' -++++ config.com[80]declare -rx 'COM_WAVE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history' -++++ config.com[81]declare -rx 'COM_WAVE_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded' -++++ config.com[82]declare -rx 'COM_WAVE_GRID_RES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}' -++++ config.com[83]declare -rx 'COM_WAVE_STATION_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station' -++++ config.com[84]declare -rx 'COM_WAVE_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak' -++++ config.com[85]declare -rx 'COM_WAVE_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo' -++++ config.com[87]declare -rx 'COM_OCEAN_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history' -++++ config.com[88]declare -rx 'COM_OCEAN_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart' -++++ config.com[89]declare -rx 'COM_OCEAN_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input' -++++ config.com[90]declare -rx 'COM_OCEAN_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean' -++++ config.com[91]declare -rx 'COM_OCEAN_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon' -++++ config.com[92]declare -rx 'COM_OCEAN_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf' -++++ config.com[93]declare -rx 'COM_OCEAN_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean' -++++ config.com[94]declare -rx 'COM_OCEAN_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf' -++++ config.com[95]declare -rx 'COM_OCEAN_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2' -++++ config.com[96]declare -rx 'COM_OCEAN_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}' -++++ config.com[98]declare -rx 'COM_ICE_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice' -++++ config.com[99]declare -rx 'COM_ICE_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf' -++++ config.com[100]declare -rx 'COM_ICE_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon' -++++ config.com[101]declare -rx 'COM_ICE_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice' -++++ config.com[102]declare -rx 'COM_ICE_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input' -++++ config.com[103]declare -rx 'COM_ICE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history' -++++ config.com[104]declare -rx 'COM_ICE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart' -++++ config.com[105]declare -rx 'COM_ICE_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf' -++++ config.com[106]declare -rx 'COM_ICE_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2' -++++ config.com[107]declare -rx 'COM_ICE_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}' -++++ config.com[109]declare -rx 'COM_CHEM_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history' -++++ config.com[110]declare -rx 'COM_CHEM_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem' -++++ config.com[111]declare -rx 'COM_CHEM_BMAT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix' -++++ config.com[112]declare -rx 'COM_CHEM_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon' -++++ config.com[114]declare -rx 'COM_MED_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart' -+++ config.base[143]export LOGSCRIPT= -+++ config.base[143]LOGSCRIPT= -+++ config.base[145]export 'REDOUT=1>' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.cleanup -+++ config.cleanup[4]echo 'BEGIN: config.cleanup' -BEGIN: config.cleanup -+++ config.cleanup[7]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources cleanup -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=cleanup -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[1151]walltime=00:15:00 -++++ config.resources[1152]ntasks=1 -++++ config.resources[1153]tasks_per_node=1 -++++ config.resources[1154]threads_per_task=1 -++++ config.resources[1155]memory=4096M -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n '' ]] -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export memory -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.cleanup[9]export CLEANUP_COM=YES -+++ config.cleanup[9]CLEANUP_COM=YES -+++ config.cleanup[12]export RMOLDSTD=144 -+++ config.cleanup[12]RMOLDSTD=144 -+++ config.cleanup[13]export RMOLDEND=24 -+++ config.cleanup[13]RMOLDEND=24 -+++ config.cleanup[15][[ NO == \Y\E\S ]] -+++ config.cleanup[23]case ${RUN} in -+++ config.cleanup[24]exclude_string='*prepbufr*, *cnvstat*, *atmanl.nc' -+++ config.cleanup[28]export exclude_string -+++ config.cleanup[30]echo 'END: config.cleanup' -END: config.cleanup -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env cleanup -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=cleanup -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 1029208 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 4194304 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 1 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 1 ]] -+++ HERCULES.env[34]max_threads_per_task=80 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 80 ]] -+++ HERCULES.env[40][[ 1 -gt 80 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 1' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[363]echo 'WARNING: The job step cleanup does not specify Hercules-specific resources' -WARNING: The job step cleanup does not specify Hercules-specific resources -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_CLEANUP[5]/work2/noaa/global/mterry/global-workflow_forked/scripts/exglobal_cleanup.sh -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ exglobal_cleanup.sh[4]echo 'Begin Cleanup /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312!' -Begin Cleanup /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312! -+ exglobal_cleanup.sh[8]DATAfcst=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/gfsfcst.2021032312 -+ exglobal_cleanup.sh[9][[ -d /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/gfsfcst.2021032312 ]] -+ exglobal_cleanup.sh[9]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/gfsfcst.2021032312 -+ exglobal_cleanup.sh[11]rm -rf '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/gfsefcs*2021032312' -+ exglobal_cleanup.sh[14][[ YES == NO ]] -++ exglobal_cleanup.sh[24]date --utc +%Y%m%d%H -d '20210323 12 -24 hours' -+ exglobal_cleanup.sh[24]last_date=2021032212 -++ exglobal_cleanup.sh[25]date --utc +%Y%m%d%H -d '20210323 12 -144 hours' -+ exglobal_cleanup.sh[25]first_date=2021031712 -++ exglobal_cleanup.sh[26]date --utc +%Y%m%d%H -d '20210323 12 -48 hours' -+ exglobal_cleanup.sh[26]last_rtofs=2021032112 -+ exglobal_cleanup.sh[50](( current_date=first_date )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[52]current_PDY=20210317 -+ exglobal_cleanup.sh[53]current_cyc=12 -+ exglobal_cleanup.sh[54]rtofs_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/rtofs.20210317 -+ exglobal_cleanup.sh[55]rocotolog=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021031712.log -+ exglobal_cleanup.sh[56][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021031712.log ]] -++ exglobal_cleanup.sh[50]date --utc +%Y%m%d%H -d '20210317 12 +6 hours' -+ exglobal_cleanup.sh[50](( current_date=2021031718 )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[52]current_PDY=20210317 -+ exglobal_cleanup.sh[53]current_cyc=18 -+ exglobal_cleanup.sh[54]rtofs_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/rtofs.20210317 -+ exglobal_cleanup.sh[55]rocotolog=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021031718.log -+ exglobal_cleanup.sh[56][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021031718.log ]] -++ exglobal_cleanup.sh[50]date --utc +%Y%m%d%H -d '20210317 18 +6 hours' -+ exglobal_cleanup.sh[50](( current_date=2021031800 )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[52]current_PDY=20210318 -+ exglobal_cleanup.sh[53]current_cyc=00 -+ exglobal_cleanup.sh[54]rtofs_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/rtofs.20210318 -+ exglobal_cleanup.sh[55]rocotolog=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021031800.log -+ exglobal_cleanup.sh[56][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021031800.log ]] -++ exglobal_cleanup.sh[50]date --utc +%Y%m%d%H -d '20210318 00 +6 hours' -+ exglobal_cleanup.sh[50](( current_date=2021031806 )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[52]current_PDY=20210318 -+ exglobal_cleanup.sh[53]current_cyc=06 -+ exglobal_cleanup.sh[54]rtofs_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/rtofs.20210318 -+ exglobal_cleanup.sh[55]rocotolog=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021031806.log -+ exglobal_cleanup.sh[56][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021031806.log ]] -++ exglobal_cleanup.sh[50]date --utc +%Y%m%d%H -d '20210318 06 +6 hours' -+ exglobal_cleanup.sh[50](( current_date=2021031812 )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[52]current_PDY=20210318 -+ exglobal_cleanup.sh[53]current_cyc=12 -+ exglobal_cleanup.sh[54]rtofs_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/rtofs.20210318 -+ exglobal_cleanup.sh[55]rocotolog=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021031812.log -+ exglobal_cleanup.sh[56][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021031812.log ]] -++ exglobal_cleanup.sh[50]date --utc +%Y%m%d%H -d '20210318 12 +6 hours' -+ exglobal_cleanup.sh[50](( current_date=2021031818 )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[52]current_PDY=20210318 -+ exglobal_cleanup.sh[53]current_cyc=18 -+ exglobal_cleanup.sh[54]rtofs_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/rtofs.20210318 -+ exglobal_cleanup.sh[55]rocotolog=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021031818.log -+ exglobal_cleanup.sh[56][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021031818.log ]] -++ exglobal_cleanup.sh[50]date --utc +%Y%m%d%H -d '20210318 18 +6 hours' -+ exglobal_cleanup.sh[50](( current_date=2021031900 )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[52]current_PDY=20210319 -+ exglobal_cleanup.sh[53]current_cyc=00 -+ exglobal_cleanup.sh[54]rtofs_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/rtofs.20210319 -+ exglobal_cleanup.sh[55]rocotolog=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021031900.log -+ exglobal_cleanup.sh[56][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021031900.log ]] -++ exglobal_cleanup.sh[50]date --utc +%Y%m%d%H -d '20210319 00 +6 hours' -+ exglobal_cleanup.sh[50](( current_date=2021031906 )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[52]current_PDY=20210319 -+ exglobal_cleanup.sh[53]current_cyc=06 -+ exglobal_cleanup.sh[54]rtofs_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/rtofs.20210319 -+ exglobal_cleanup.sh[55]rocotolog=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021031906.log -+ exglobal_cleanup.sh[56][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021031906.log ]] -++ exglobal_cleanup.sh[50]date --utc +%Y%m%d%H -d '20210319 06 +6 hours' -+ exglobal_cleanup.sh[50](( current_date=2021031912 )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[52]current_PDY=20210319 -+ exglobal_cleanup.sh[53]current_cyc=12 -+ exglobal_cleanup.sh[54]rtofs_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/rtofs.20210319 -+ exglobal_cleanup.sh[55]rocotolog=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021031912.log -+ exglobal_cleanup.sh[56][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021031912.log ]] -++ exglobal_cleanup.sh[50]date --utc +%Y%m%d%H -d '20210319 12 +6 hours' -+ exglobal_cleanup.sh[50](( current_date=2021031918 )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[52]current_PDY=20210319 -+ exglobal_cleanup.sh[53]current_cyc=18 -+ exglobal_cleanup.sh[54]rtofs_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/rtofs.20210319 -+ exglobal_cleanup.sh[55]rocotolog=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021031918.log -+ exglobal_cleanup.sh[56][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021031918.log ]] -++ exglobal_cleanup.sh[50]date --utc +%Y%m%d%H -d '20210319 18 +6 hours' -+ exglobal_cleanup.sh[50](( current_date=2021032000 )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[52]current_PDY=20210320 -+ exglobal_cleanup.sh[53]current_cyc=00 -+ exglobal_cleanup.sh[54]rtofs_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/rtofs.20210320 -+ exglobal_cleanup.sh[55]rocotolog=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032000.log -+ exglobal_cleanup.sh[56][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032000.log ]] -++ exglobal_cleanup.sh[50]date --utc +%Y%m%d%H -d '20210320 00 +6 hours' -+ exglobal_cleanup.sh[50](( current_date=2021032006 )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[52]current_PDY=20210320 -+ exglobal_cleanup.sh[53]current_cyc=06 -+ exglobal_cleanup.sh[54]rtofs_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/rtofs.20210320 -+ exglobal_cleanup.sh[55]rocotolog=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032006.log -+ exglobal_cleanup.sh[56][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032006.log ]] -++ exglobal_cleanup.sh[50]date --utc +%Y%m%d%H -d '20210320 06 +6 hours' -+ exglobal_cleanup.sh[50](( current_date=2021032012 )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[52]current_PDY=20210320 -+ exglobal_cleanup.sh[53]current_cyc=12 -+ exglobal_cleanup.sh[54]rtofs_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/rtofs.20210320 -+ exglobal_cleanup.sh[55]rocotolog=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032012.log -+ exglobal_cleanup.sh[56][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032012.log ]] -++ exglobal_cleanup.sh[50]date --utc +%Y%m%d%H -d '20210320 12 +6 hours' -+ exglobal_cleanup.sh[50](( current_date=2021032018 )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[52]current_PDY=20210320 -+ exglobal_cleanup.sh[53]current_cyc=18 -+ exglobal_cleanup.sh[54]rtofs_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/rtofs.20210320 -+ exglobal_cleanup.sh[55]rocotolog=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032018.log -+ exglobal_cleanup.sh[56][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032018.log ]] -++ exglobal_cleanup.sh[50]date --utc +%Y%m%d%H -d '20210320 18 +6 hours' -+ exglobal_cleanup.sh[50](( current_date=2021032100 )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[52]current_PDY=20210321 -+ exglobal_cleanup.sh[53]current_cyc=00 -+ exglobal_cleanup.sh[54]rtofs_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/rtofs.20210321 -+ exglobal_cleanup.sh[55]rocotolog=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032100.log -+ exglobal_cleanup.sh[56][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032100.log ]] -++ exglobal_cleanup.sh[50]date --utc +%Y%m%d%H -d '20210321 00 +6 hours' -+ exglobal_cleanup.sh[50](( current_date=2021032106 )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[52]current_PDY=20210321 -+ exglobal_cleanup.sh[53]current_cyc=06 -+ exglobal_cleanup.sh[54]rtofs_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/rtofs.20210321 -+ exglobal_cleanup.sh[55]rocotolog=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032106.log -+ exglobal_cleanup.sh[56][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032106.log ]] -++ exglobal_cleanup.sh[50]date --utc +%Y%m%d%H -d '20210321 06 +6 hours' -+ exglobal_cleanup.sh[50](( current_date=2021032112 )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[52]current_PDY=20210321 -+ exglobal_cleanup.sh[53]current_cyc=12 -+ exglobal_cleanup.sh[54]rtofs_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/rtofs.20210321 -+ exglobal_cleanup.sh[55]rocotolog=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032112.log -+ exglobal_cleanup.sh[56][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032112.log ]] -++ exglobal_cleanup.sh[50]date --utc +%Y%m%d%H -d '20210321 12 +6 hours' -+ exglobal_cleanup.sh[50](( current_date=2021032118 )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[52]current_PDY=20210321 -+ exglobal_cleanup.sh[53]current_cyc=18 -+ exglobal_cleanup.sh[54]rtofs_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/rtofs.20210321 -+ exglobal_cleanup.sh[55]rocotolog=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032118.log -+ exglobal_cleanup.sh[56][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032118.log ]] -++ exglobal_cleanup.sh[50]date --utc +%Y%m%d%H -d '20210321 18 +6 hours' -+ exglobal_cleanup.sh[50](( current_date=2021032200 )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[52]current_PDY=20210322 -+ exglobal_cleanup.sh[53]current_cyc=00 -+ exglobal_cleanup.sh[54]rtofs_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/rtofs.20210322 -+ exglobal_cleanup.sh[55]rocotolog=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032200.log -+ exglobal_cleanup.sh[56][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032200.log ]] -++ exglobal_cleanup.sh[50]date --utc +%Y%m%d%H -d '20210322 00 +6 hours' -+ exglobal_cleanup.sh[50](( current_date=2021032206 )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[52]current_PDY=20210322 -+ exglobal_cleanup.sh[53]current_cyc=06 -+ exglobal_cleanup.sh[54]rtofs_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/rtofs.20210322 -+ exglobal_cleanup.sh[55]rocotolog=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032206.log -+ exglobal_cleanup.sh[56][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032206.log ]] -++ exglobal_cleanup.sh[50]date --utc +%Y%m%d%H -d '20210322 06 +6 hours' -+ exglobal_cleanup.sh[50](( current_date=2021032212 )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[52]current_PDY=20210322 -+ exglobal_cleanup.sh[53]current_cyc=12 -+ exglobal_cleanup.sh[54]rtofs_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/rtofs.20210322 -+ exglobal_cleanup.sh[55]rocotolog=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032212.log -+ exglobal_cleanup.sh[56][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/logs/2021032212.log ]] -++ exglobal_cleanup.sh[50]date --utc +%Y%m%d%H -d '20210322 12 +6 hours' -+ exglobal_cleanup.sh[50](( current_date=2021032218 )) -+ exglobal_cleanup.sh[50](( current_date <= last_date )) -+ exglobal_cleanup.sh[76][[ gfs == \g\f\s ]] -+ exglobal_cleanup.sh[77]fhmax=156 -++ exglobal_cleanup.sh[78]date --utc +%Y%m%d%H -d '20210323 12 -156 hours' -+ exglobal_cleanup.sh[78]RDATE=2021031700 -+ exglobal_cleanup.sh[79]verify_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210317 -+ exglobal_cleanup.sh[80][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210317 ]] -++ exglobal_cleanup.sh[84]date --utc +%Y%m%d%H -d '20210323 12 -120 hours' -+ exglobal_cleanup.sh[84]touch_date=2021031812 -+ exglobal_cleanup.sh[85](( touch_date < 2021032312 )) -+ exglobal_cleanup.sh[86]touch_PDY=20210318 -+ exglobal_cleanup.sh[87]touch_cyc=12 -+ exglobal_cleanup.sh[88]touch_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210318/12 -+ exglobal_cleanup.sh[89][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210318/12 ]] -++ exglobal_cleanup.sh[92]date --utc +%Y%m%d%H -d '20210318 12 +6 hours' -+ exglobal_cleanup.sh[92]touch_date=2021031818 -+ exglobal_cleanup.sh[85](( touch_date < 2021032312 )) -+ exglobal_cleanup.sh[86]touch_PDY=20210318 -+ exglobal_cleanup.sh[87]touch_cyc=18 -+ exglobal_cleanup.sh[88]touch_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210318/18 -+ exglobal_cleanup.sh[89][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210318/18 ]] -++ exglobal_cleanup.sh[92]date --utc +%Y%m%d%H -d '20210318 18 +6 hours' -+ exglobal_cleanup.sh[92]touch_date=2021031900 -+ exglobal_cleanup.sh[85](( touch_date < 2021032312 )) -+ exglobal_cleanup.sh[86]touch_PDY=20210319 -+ exglobal_cleanup.sh[87]touch_cyc=00 -+ exglobal_cleanup.sh[88]touch_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210319/00 -+ exglobal_cleanup.sh[89][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210319/00 ]] -++ exglobal_cleanup.sh[92]date --utc +%Y%m%d%H -d '20210319 00 +6 hours' -+ exglobal_cleanup.sh[92]touch_date=2021031906 -+ exglobal_cleanup.sh[85](( touch_date < 2021032312 )) -+ exglobal_cleanup.sh[86]touch_PDY=20210319 -+ exglobal_cleanup.sh[87]touch_cyc=06 -+ exglobal_cleanup.sh[88]touch_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210319/06 -+ exglobal_cleanup.sh[89][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210319/06 ]] -++ exglobal_cleanup.sh[92]date --utc +%Y%m%d%H -d '20210319 06 +6 hours' -+ exglobal_cleanup.sh[92]touch_date=2021031912 -+ exglobal_cleanup.sh[85](( touch_date < 2021032312 )) -+ exglobal_cleanup.sh[86]touch_PDY=20210319 -+ exglobal_cleanup.sh[87]touch_cyc=12 -+ exglobal_cleanup.sh[88]touch_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210319/12 -+ exglobal_cleanup.sh[89][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210319/12 ]] -++ exglobal_cleanup.sh[92]date --utc +%Y%m%d%H -d '20210319 12 +6 hours' -+ exglobal_cleanup.sh[92]touch_date=2021031918 -+ exglobal_cleanup.sh[85](( touch_date < 2021032312 )) -+ exglobal_cleanup.sh[86]touch_PDY=20210319 -+ exglobal_cleanup.sh[87]touch_cyc=18 -+ exglobal_cleanup.sh[88]touch_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210319/18 -+ exglobal_cleanup.sh[89][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210319/18 ]] -++ exglobal_cleanup.sh[92]date --utc +%Y%m%d%H -d '20210319 18 +6 hours' -+ exglobal_cleanup.sh[92]touch_date=2021032000 -+ exglobal_cleanup.sh[85](( touch_date < 2021032312 )) -+ exglobal_cleanup.sh[86]touch_PDY=20210320 -+ exglobal_cleanup.sh[87]touch_cyc=00 -+ exglobal_cleanup.sh[88]touch_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210320/00 -+ exglobal_cleanup.sh[89][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210320/00 ]] -++ exglobal_cleanup.sh[92]date --utc +%Y%m%d%H -d '20210320 00 +6 hours' -+ exglobal_cleanup.sh[92]touch_date=2021032006 -+ exglobal_cleanup.sh[85](( touch_date < 2021032312 )) -+ exglobal_cleanup.sh[86]touch_PDY=20210320 -+ exglobal_cleanup.sh[87]touch_cyc=06 -+ exglobal_cleanup.sh[88]touch_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210320/06 -+ exglobal_cleanup.sh[89][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210320/06 ]] -++ exglobal_cleanup.sh[92]date --utc +%Y%m%d%H -d '20210320 06 +6 hours' -+ exglobal_cleanup.sh[92]touch_date=2021032012 -+ exglobal_cleanup.sh[85](( touch_date < 2021032312 )) -+ exglobal_cleanup.sh[86]touch_PDY=20210320 -+ exglobal_cleanup.sh[87]touch_cyc=12 -+ exglobal_cleanup.sh[88]touch_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210320/12 -+ exglobal_cleanup.sh[89][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210320/12 ]] -++ exglobal_cleanup.sh[92]date --utc +%Y%m%d%H -d '20210320 12 +6 hours' -+ exglobal_cleanup.sh[92]touch_date=2021032018 -+ exglobal_cleanup.sh[85](( touch_date < 2021032312 )) -+ exglobal_cleanup.sh[86]touch_PDY=20210320 -+ exglobal_cleanup.sh[87]touch_cyc=18 -+ exglobal_cleanup.sh[88]touch_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210320/18 -+ exglobal_cleanup.sh[89][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210320/18 ]] -++ exglobal_cleanup.sh[92]date --utc +%Y%m%d%H -d '20210320 18 +6 hours' -+ exglobal_cleanup.sh[92]touch_date=2021032100 -+ exglobal_cleanup.sh[85](( touch_date < 2021032312 )) -+ exglobal_cleanup.sh[86]touch_PDY=20210321 -+ exglobal_cleanup.sh[87]touch_cyc=00 -+ exglobal_cleanup.sh[88]touch_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210321/00 -+ exglobal_cleanup.sh[89][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210321/00 ]] -++ exglobal_cleanup.sh[92]date --utc +%Y%m%d%H -d '20210321 00 +6 hours' -+ exglobal_cleanup.sh[92]touch_date=2021032106 -+ exglobal_cleanup.sh[85](( touch_date < 2021032312 )) -+ exglobal_cleanup.sh[86]touch_PDY=20210321 -+ exglobal_cleanup.sh[87]touch_cyc=06 -+ exglobal_cleanup.sh[88]touch_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210321/06 -+ exglobal_cleanup.sh[89][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210321/06 ]] -++ exglobal_cleanup.sh[92]date --utc +%Y%m%d%H -d '20210321 06 +6 hours' -+ exglobal_cleanup.sh[92]touch_date=2021032112 -+ exglobal_cleanup.sh[85](( touch_date < 2021032312 )) -+ exglobal_cleanup.sh[86]touch_PDY=20210321 -+ exglobal_cleanup.sh[87]touch_cyc=12 -+ exglobal_cleanup.sh[88]touch_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210321/12 -+ exglobal_cleanup.sh[89][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210321/12 ]] -++ exglobal_cleanup.sh[92]date --utc +%Y%m%d%H -d '20210321 12 +6 hours' -+ exglobal_cleanup.sh[92]touch_date=2021032118 -+ exglobal_cleanup.sh[85](( touch_date < 2021032312 )) -+ exglobal_cleanup.sh[86]touch_PDY=20210321 -+ exglobal_cleanup.sh[87]touch_cyc=18 -+ exglobal_cleanup.sh[88]touch_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210321/18 -+ exglobal_cleanup.sh[89][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210321/18 ]] -++ exglobal_cleanup.sh[92]date --utc +%Y%m%d%H -d '20210321 18 +6 hours' -+ exglobal_cleanup.sh[92]touch_date=2021032200 -+ exglobal_cleanup.sh[85](( touch_date < 2021032312 )) -+ exglobal_cleanup.sh[86]touch_PDY=20210322 -+ exglobal_cleanup.sh[87]touch_cyc=00 -+ exglobal_cleanup.sh[88]touch_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210322/00 -+ exglobal_cleanup.sh[89][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210322/00 ]] -++ exglobal_cleanup.sh[92]date --utc +%Y%m%d%H -d '20210322 00 +6 hours' -+ exglobal_cleanup.sh[92]touch_date=2021032206 -+ exglobal_cleanup.sh[85](( touch_date < 2021032312 )) -+ exglobal_cleanup.sh[86]touch_PDY=20210322 -+ exglobal_cleanup.sh[87]touch_cyc=06 -+ exglobal_cleanup.sh[88]touch_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210322/06 -+ exglobal_cleanup.sh[89][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210322/06 ]] -++ exglobal_cleanup.sh[92]date --utc +%Y%m%d%H -d '20210322 06 +6 hours' -+ exglobal_cleanup.sh[92]touch_date=2021032212 -+ exglobal_cleanup.sh[85](( touch_date < 2021032312 )) -+ exglobal_cleanup.sh[86]touch_PDY=20210322 -+ exglobal_cleanup.sh[87]touch_cyc=12 -+ exglobal_cleanup.sh[88]touch_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210322/12 -+ exglobal_cleanup.sh[89][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210322/12 ]] -++ exglobal_cleanup.sh[92]date --utc +%Y%m%d%H -d '20210322 12 +6 hours' -+ exglobal_cleanup.sh[92]touch_date=2021032218 -+ exglobal_cleanup.sh[85](( touch_date < 2021032312 )) -+ exglobal_cleanup.sh[86]touch_PDY=20210322 -+ exglobal_cleanup.sh[87]touch_cyc=18 -+ exglobal_cleanup.sh[88]touch_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210322/18 -+ exglobal_cleanup.sh[89][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210322/18 ]] -++ exglobal_cleanup.sh[92]date --utc +%Y%m%d%H -d '20210322 18 +6 hours' -+ exglobal_cleanup.sh[92]touch_date=2021032300 -+ exglobal_cleanup.sh[85](( touch_date < 2021032312 )) -+ exglobal_cleanup.sh[86]touch_PDY=20210323 -+ exglobal_cleanup.sh[87]touch_cyc=00 -+ exglobal_cleanup.sh[88]touch_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210323/00 -+ exglobal_cleanup.sh[89][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210323/00 ]] -++ exglobal_cleanup.sh[92]date --utc +%Y%m%d%H -d '20210323 00 +6 hours' -+ exglobal_cleanup.sh[92]touch_date=2021032306 -+ exglobal_cleanup.sh[85](( touch_date < 2021032312 )) -+ exglobal_cleanup.sh[86]touch_PDY=20210323 -+ exglobal_cleanup.sh[87]touch_cyc=06 -+ exglobal_cleanup.sh[88]touch_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210323/06 -+ exglobal_cleanup.sh[89][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/vrfyarch/gfs.20210323/06 ]] -++ exglobal_cleanup.sh[92]date --utc +%Y%m%d%H -d '20210323 06 +6 hours' -+ exglobal_cleanup.sh[92]touch_date=2021032312 -+ exglobal_cleanup.sh[85](( touch_date < 2021032312 )) -++ exglobal_cleanup.sh[97]date --utc +%Y%m%d%H -d '20210323 12 -144 hours' -+ exglobal_cleanup.sh[97]GDATE=2021031712 -++ exglobal_cleanup.sh[98]date --utc +%Y%m%d%H -d '20210323 12 -120 hours' -+ exglobal_cleanup.sh[98]RDATE=2021031812 -+ exglobal_cleanup.sh[99](( GDATE < RDATE )) -+ exglobal_cleanup.sh[100]RDATE=2021031712 -+ exglobal_cleanup.sh[102]deletion_target=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210317 -+ exglobal_cleanup.sh[103][[ -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210317 ]] -+ exglobal_cleanup.sh[106]sync -+ exglobal_cleanup.sh[106]sleep 1 -+ JGLOBAL_CLEANUP[5]true -+ JGLOBAL_CLEANUP[6]export err=0 -+ JGLOBAL_CLEANUP[6]err=0 -+ JGLOBAL_CLEANUP[7][[ 0 -ne 0 ]] -+ JGLOBAL_CLEANUP[18]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312 -+ JGLOBAL_CLEANUP[19]echo 'Cleanup /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312 completed!' -Cleanup /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312 completed! -+ JGLOBAL_CLEANUP[21]exit 0 -+ JGLOBAL_CLEANUP[1]postamble /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_CLEANUP 1753758986 0 -+ preamble.sh[62]set +x -End /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_CLEANUP at 03:16:44 with error code 0 (time elapsed: 00:00:18) -+ status=0 -+ exit 0 diff --git a/ci/error_logs/HECULES_PR_LOGS/C48_S2SW_logs_2021032312_gfs_ice_prod_f024.log b/ci/error_logs/HECULES_PR_LOGS/C48_S2SW_logs_2021032312_gfs_ice_prod_f024.log deleted file mode 100644 index d0094f91..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/C48_S2SW_logs_2021032312_gfs_ice_prod_f024.log +++ /dev/null @@ -1,1027 +0,0 @@ -+ . /work2/noaa/global/mterry/global-workflow_forked/ush/load_fv3gfs_modules.sh -++ [[ hxB == *x* ]] -++ set_x=YES -++ [[ NO == \N\O ]] -++ echo 'Loading modules quietly...' -Loading modules quietly... -++ set +x -Running "module reset". Resetting modules to system default. The following $MODULEPATH directories have been removed: None - -Currently Loaded Modules: - 1) contrib/0.1 43) ip/4.3.0 - 2) intel-oneapi-compilers/2023.1.0 44) grib-util/1.3.0 - 3) stack-intel/2021.9.0 45) g2tmpl/1.13.0 - 4) intel-oneapi-mpi/2021.9.0 46) gsi-ncdiag/1.1.2 - 5) stack-intel-oneapi-mpi/2021.9.0 47) crtm-fix/2.4.0.1_emc - 6) intel-oneapi-mkl/2023.1.0 48) git-lfs/3.1.2 - 7) zlib/1.2.13 49) crtm/2.4.0.1 - 8) pigz/2.7 50) openblas/0.3.24 - 9) zstd/1.5.2 51) py-setuptools/63.4.3 - 10) tar/1.34 52) py-numpy/1.23.4 - 11) gettext/0.21.1 53) bufr/11.7.0 - 12) libxcrypt/4.4.35 54) gmake/4.2.1 - 13) sqlite/3.43.2 55) wgrib2/2.0.8 - 14) util-linux-uuid/2.38.1 56) py-f90nml/1.4.3 - 15) python/3.11.6 57) py-cftime/1.0.3.4 - 16) libjpeg/2.1.0 58) py-netcdf4/1.5.8 - 17) jasper/2.0.32 59) libyaml/0.2.5 - 18) libpng/1.6.37 60) py-pyyaml/6.0 - 19) openjpeg/2.3.1 61) py-markupsafe/2.1.3 - 20) eccodes/2.32.0 62) py-jinja2/3.1.2 - 21) fftw/3.3.10 63) py-bottleneck/1.3.7 - 22) nghttp2/1.57.0 64) py-numexpr/2.8.4 - 23) curl/8.4.0 65) py-et-xmlfile/1.0.1 - 24) proj/9.2.1 66) py-openpyxl/3.1.2 - 25) udunits/2.2.28 67) py-pytz/2023.3 - 26) cdo/2.2.0 68) py-pyxlsb/1.0.10 - 27) hdf5/1.14.0 69) py-xlrd/2.0.1 - 28) snappy/1.1.10 70) py-xlsxwriter/3.1.7 - 29) c-blosc/1.21.5 71) py-xlwt/1.3.0 - 30) netcdf-c/4.9.2 72) py-pandas/1.5.3 - 31) netcdf-fortran/4.6.1 73) py-six/1.16.0 - 32) parallel-netcdf/1.12.2 74) py-python-dateutil/2.8.2 - 33) parallelio/2.5.10 75) g2c/1.8.0 - 34) esmf/8.6.0 76) netcdf-cxx4/4.3.1 - 35) antlr/2.7.7 77) met/9.1.3 - 36) gsl/2.7.1 78) metplus/3.1.1 - 37) nco/5.0.6 79) py-packaging/23.1 - 38) bacio/2.4.1 80) py-xarray/2023.7.0 - 39) w3emc/2.10.0 81) prepobs/1.1.0 - 40) prod_util/2.1.1 82) fit2obs/1.1.7.1 - 41) g2/3.4.5 83) globus-cli/3.35.2 - 42) sp/2.5.0 84) module_base.hercules - - - -++ [[ -d /work2/noaa/global/mterry/global-workflow_forked/sorc/wxflow/src ]] -++ PYTHONPATH=/apps/other/globus-cli-3.35.2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:/work2/noaa/global/mterry/global-workflow_forked/sorc/wxflow/src -++ PYTHONPATH=/apps/other/globus-cli-3.35.2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:/work2/noaa/global/mterry/global-workflow_forked/sorc/wxflow/src:/work2/noaa/global/mterry/global-workflow_forked/ush/python -++ export PYTHONPATH -++ ulimit -S -s 16384 -++ unset ulimit_s -+ status=0 -+ (( status != 0 )) -+ export job=oceanice_products -+ job=oceanice_products -+ IFS=', ' -+ read -r -a fhr_list -+ export FORECAST_HOUR jobid -+ for FORECAST_HOUR in "${fhr_list[@]}" -++ printf %03d 24 -+ fhr3=024 -+ jobid=oceanice_products_ice_f024.615012 -+ /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_OCEANICE_PRODUCTS -Begin /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_OCEANICE_PRODUCTS at Tue Jul 29 02:34:23 UTC 2025 -++ jjob_header.sh[46]OPTIND=1 -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[50]env_job=oceanice_products -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[49]read -ra configs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[61]shift 4 -++ jjob_header.sh[63][[ -z oceanice_products ]] -++ jjob_header.sh[71]export DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ice_f024.615012 -++ jjob_header.sh[71]DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ice_f024.615012 -++ jjob_header.sh[72][[ YES == \Y\E\S ]] -++ jjob_header.sh[73]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ice_f024.615012 -++ jjob_header.sh[75]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ice_f024.615012 -++ jjob_header.sh[76]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ice_f024.615012 -++ jjob_header.sh[85]export pid=615344 -++ jjob_header.sh[85]pid=615344 -++ jjob_header.sh[86]export pgmout=OUTPUT.615344 -++ jjob_header.sh[86]pgmout=OUTPUT.615344 -++ jjob_header.sh[87]export pgmerr=errfile -++ jjob_header.sh[87]pgmerr=errfile -++ jjob_header.sh[90]export pgm= -++ jjob_header.sh[90]pgm= -++ jjob_header.sh[96]export cycle=t12z -++ jjob_header.sh[96]cycle=t12z -++ jjob_header.sh[97]setpdy.sh -+ setpdy.sh[20]'[' /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ice_f024.615012 == /home/mterry ']' -+ setpdy.sh[25][[ ! t12z =~ t??z ]] -+ setpdy.sh[30]case $# in -+ setpdy.sh[31]dates_before_PDY=7 -+ setpdy.sh[32]dates_after_PDY=7 -+ setpdy.sh[50]COMDATEROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+ setpdy.sh[53]'[' -z 20210323 ']' -+ setpdy.sh[57]sed 's/[0-9]\{8\}/20210323/' /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z -sed: can't read /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z: No such file or directory -++ jjob_header.sh[97]true -++ jjob_header.sh[98]source ./PDY -/work2/noaa/global/mterry/global-workflow_forked/ush/jjob_header.sh: line 98: ./PDY: No such file or directory -++ jjob_header.sh[98]true -++ jjob_header.sh[104]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[104]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.base -+++ config.base[6]echo 'BEGIN: config.base' -BEGIN: config.base -+++ config.base[9]export machine=HERCULES -+++ config.base[9]machine=HERCULES -+++ config.base[12]export RUN_ENVIR=emc -+++ config.base[12]RUN_ENVIR=emc -+++ config.base[15]export ACCOUNT=fv3-cpu -+++ config.base[15]ACCOUNT=fv3-cpu -+++ config.base[16]export QUEUE=batch -+++ config.base[16]QUEUE=batch -+++ config.base[17]export QUEUE_SERVICE=batch -+++ config.base[17]QUEUE_SERVICE=batch -+++ config.base[18]export QUEUE_DTN=batch -+++ config.base[18]QUEUE_DTN=batch -+++ config.base[19]export PARTITION_BATCH=hercules -+++ config.base[19]PARTITION_BATCH=hercules -+++ config.base[20]export PARTITION_SERVICE=service -+++ config.base[20]PARTITION_SERVICE=service -+++ config.base[21]export PARTITION_DTN= -+++ config.base[21]PARTITION_DTN= -+++ config.base[22]export RESERVATION= -+++ config.base[22]RESERVATION= -+++ config.base[23]export CLUSTERS= -+++ config.base[23]CLUSTERS= -+++ config.base[24]export CLUSTERS_SERVICE= -+++ config.base[24]CLUSTERS_SERVICE= -+++ config.base[25]export CLUSTERS_DTN= -+++ config.base[25]CLUSTERS_DTN= -+++ config.base[28]export HPSS_PROJECT=emc-global -+++ config.base[28]HPSS_PROJECT=emc-global -+++ config.base[31]export HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[31]HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[32]export EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[32]EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[33]export FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[33]FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[34]export PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[34]PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[35]export SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[35]SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[36]export USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[36]USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[38]export FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[38]FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[39]export FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[39]FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[40]export FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[40]FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[41]export FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[41]FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[42]export FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[42]FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[43]export FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[43]FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[44]export FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[44]FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[45]export FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[45]FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[50]export PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[50]PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[51]export COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[51]COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[52]export COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[52]COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[53]export DMPDIR=/work/noaa/rstprod/dump -+++ config.base[53]DMPDIR=/work/noaa/rstprod/dump -+++ config.base[57]export COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[57]COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[58]export COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[58]COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[59]export COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[59]COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[62]export HOMEDIR=/work2/noaa/global/mterry -+++ config.base[62]HOMEDIR=/work2/noaa/global/mterry -+++ config.base[63]export STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[63]STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]export PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[65]export NOSCRUB=/work2/noaa/global/mterry -+++ config.base[65]NOSCRUB=/work2/noaa/global/mterry -+++ config.base[68]export BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[68]BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[71]export BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[71]BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[74]export DO_PREP_SFC=NO -+++ config.base[74]DO_PREP_SFC=NO -+++ config.base[77]export DO_GOES=NO -+++ config.base[77]DO_GOES=NO -+++ config.base[78]export DO_BUFRSND=NO -+++ config.base[78]DO_BUFRSND=NO -+++ config.base[79]export DO_GEMPAK=NO -+++ config.base[79]DO_GEMPAK=NO -+++ config.base[80]export DO_AWIPS=NO -+++ config.base[80]DO_AWIPS=NO -+++ config.base[81]export DO_NPOESS=NO -+++ config.base[81]DO_NPOESS=NO -+++ config.base[82]export DO_TRACKER=YES -+++ config.base[82]DO_TRACKER=YES -+++ config.base[83]export DO_GENESIS=YES -+++ config.base[83]DO_GENESIS=YES -+++ config.base[84]export DO_GENESIS_FSU=NO -+++ config.base[84]DO_GENESIS_FSU=NO -+++ config.base[85]export DO_VERFOZN=YES -+++ config.base[85]DO_VERFOZN=YES -+++ config.base[86]export DO_VERFRAD=YES -+++ config.base[86]DO_VERFRAD=YES -+++ config.base[87]export DO_VMINMON=YES -+++ config.base[87]DO_VMINMON=YES -+++ config.base[88]export DO_ANLSTAT=NO -+++ config.base[88]DO_ANLSTAT=NO -+++ config.base[91]export MODE=forecast-only -+++ config.base[91]MODE=forecast-only -+++ config.base[92]export DO_TEST_MODE=YES -+++ config.base[92]DO_TEST_MODE=YES -+++ config.base[101]export FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[101]FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[102]export HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[102]HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[103]export HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v1.2.0 -+++ config.base[103]HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v1.2.0 -+++ config.base[106]export NMV=/bin/mv -+++ config.base[106]NMV=/bin/mv -+++ config.base[107]export 'NLN=/bin/ln -sf' -+++ config.base[107]NLN='/bin/ln -sf' -+++ config.base[108]export VERBOSE=YES -+++ config.base[108]VERBOSE=YES -+++ config.base[109]export KEEPDATA=NO -+++ config.base[109]KEEPDATA=NO -+++ config.base[110]export DEBUG_POSTSCRIPT=NO -+++ config.base[110]DEBUG_POSTSCRIPT=NO -+++ config.base[111]export CHGRP_RSTPROD=YES -+++ config.base[111]CHGRP_RSTPROD=YES -+++ config.base[112]export 'CHGRP_CMD=chgrp rstprod' -+++ config.base[112]CHGRP_CMD='chgrp rstprod' -+++ config.base[113]export NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[113]NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[114]export NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[114]NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[117]export BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[117]BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[120]export SDATE=2021032312 -+++ config.base[120]SDATE=2021032312 -+++ config.base[121]export EDATE=2021032312 -+++ config.base[121]EDATE=2021032312 -+++ config.base[122]export EXP_WARM_START=.false. -+++ config.base[122]EXP_WARM_START=.false. -+++ config.base[123]export assim_freq=6 -+++ config.base[123]assim_freq=6 -+++ config.base[124]export PSLOT=C48_S2SW -+++ config.base[124]PSLOT=C48_S2SW -+++ config.base[125]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[125]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[126]export ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[126]ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[127]export DUMP_SUFFIX= -+++ config.base[127]DUMP_SUFFIX= -+++ config.base[128][[ 2021032312 -ge 2019092100 ]] -+++ config.base[128][[ 2021032312 -le 2019110700 ]] -+++ config.base[131]export ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[131]ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[132]export ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[132]ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[133]export FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[133]FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[136]export envir=prod -+++ config.base[136]envir=prod -+++ config.base[137]export NET=gfs -+++ config.base[137]NET=gfs -+++ config.base[138]export RUN=gfs -+++ config.base[138]RUN=gfs -+++ config.base[141]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.com -++++ config.com[4]echo 'BEGIN: config.com' -BEGIN: config.com -++++ config.com[38][[ emc == \n\c\o ]] -++++ config.com[43]COM_OBSPROC_TMPL='${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos' -++++ config.com[44]COM_RTOFS_TMPL='${DMPDIR}' -++++ config.com[45]COM_TCVITAL_TMPL='${DMPDIR}/${RUN}.${YMD}/${HH}/atmos' -++++ config.com[47]declare -rx 'COM_OBS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/obs' -++++ config.com[48]declare -rx COM_OBSPROC_TMPL COM_RTOFS_TMPL -++++ config.com[50]COM_BASE='${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}' -++++ config.com[52]declare -rx 'COM_TOP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}' -++++ config.com[54]declare -rx 'COM_CONF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf' -++++ config.com[55]declare -rx 'COM_OBS_JEDI=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi' -++++ config.com[57]declare -rx 'COM_ATMOS_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input' -++++ config.com[58]declare -rx 'COM_ATMOS_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart' -++++ config.com[59]declare -rx 'COM_ATMOS_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos' -++++ config.com[60]declare -rx 'COM_SNOW_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow' -++++ config.com[61]declare -rx 'COM_SNOW_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon' -++++ config.com[62]declare -rx 'COM_ATMOS_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history' -++++ config.com[63]declare -rx 'COM_ATMOS_MASTER_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master' -++++ config.com[64]declare -rx 'COM_ATMOS_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2' -++++ config.com[65]declare -rx 'COM_ATMOS_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}' -++++ config.com[66]declare -rx 'COM_ATMOS_BUFR_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr' -++++ config.com[67]declare -rx 'COM_ATMOS_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}' -++++ config.com[68]declare -rx 'COM_ATMOS_GENESIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital' -++++ config.com[69]declare -rx 'COM_ATMOS_TRACK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks' -++++ config.com[70]declare -rx 'COM_ATMOS_GOES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim' -++++ config.com[71]declare -rx 'COM_ATMOS_IMAGERY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery' -++++ config.com[72]declare -rx 'COM_ATMOS_OZNMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon' -++++ config.com[73]declare -rx 'COM_ATMOS_RADMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon' -++++ config.com[74]declare -rx 'COM_ATMOS_MINMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon' -++++ config.com[75]declare -rx 'COM_ATMOS_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon' -++++ config.com[76]declare -rx 'COM_ATMOS_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo' -++++ config.com[78]declare -rx 'COM_WAVE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart' -++++ config.com[79]declare -rx 'COM_WAVE_PREP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep' -++++ config.com[80]declare -rx 'COM_WAVE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history' -++++ config.com[81]declare -rx 'COM_WAVE_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded' -++++ config.com[82]declare -rx 'COM_WAVE_GRID_RES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}' -++++ config.com[83]declare -rx 'COM_WAVE_STATION_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station' -++++ config.com[84]declare -rx 'COM_WAVE_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak' -++++ config.com[85]declare -rx 'COM_WAVE_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo' -++++ config.com[87]declare -rx 'COM_OCEAN_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history' -++++ config.com[88]declare -rx 'COM_OCEAN_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart' -++++ config.com[89]declare -rx 'COM_OCEAN_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input' -++++ config.com[90]declare -rx 'COM_OCEAN_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean' -++++ config.com[91]declare -rx 'COM_OCEAN_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon' -++++ config.com[92]declare -rx 'COM_OCEAN_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf' -++++ config.com[93]declare -rx 'COM_OCEAN_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean' -++++ config.com[94]declare -rx 'COM_OCEAN_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf' -++++ config.com[95]declare -rx 'COM_OCEAN_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2' -++++ config.com[96]declare -rx 'COM_OCEAN_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}' -++++ config.com[98]declare -rx 'COM_ICE_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice' -++++ config.com[99]declare -rx 'COM_ICE_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf' -++++ config.com[100]declare -rx 'COM_ICE_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon' -++++ config.com[101]declare -rx 'COM_ICE_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice' -++++ config.com[102]declare -rx 'COM_ICE_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input' -++++ config.com[103]declare -rx 'COM_ICE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history' -++++ config.com[104]declare -rx 'COM_ICE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart' -++++ config.com[105]declare -rx 'COM_ICE_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf' -++++ config.com[106]declare -rx 'COM_ICE_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2' -++++ config.com[107]declare -rx 'COM_ICE_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}' -++++ config.com[109]declare -rx 'COM_CHEM_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history' -++++ config.com[110]declare -rx 'COM_CHEM_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem' -++++ config.com[111]declare -rx 'COM_CHEM_BMAT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix' -++++ config.com[112]declare -rx 'COM_CHEM_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon' -++++ config.com[114]declare -rx 'COM_MED_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart' -+++ config.base[143]export LOGSCRIPT= -+++ config.base[143]LOGSCRIPT= -+++ config.base[145]export 'REDOUT=1>' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.oceanice_products -+++ config.oceanice_products[5]echo 'BEGIN: config.oceanice_products' -BEGIN: config.oceanice_products -+++ config.oceanice_products[8]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources oceanice_products -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=oceanice_products -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[978]walltime=00:15:00 -++++ config.resources[979]ntasks=1 -++++ config.resources[980]tasks_per_node=1 -++++ config.resources[981]threads_per_task=1 -++++ config.resources[982]memory=96GB -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n '' ]] -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export memory -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.oceanice_products[11]export MAX_TASKS=25 -+++ config.oceanice_products[11]MAX_TASKS=25 -+++ config.oceanice_products[13]export OCEANICEPRODUCTS_CONFIG=/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml -+++ config.oceanice_products[13]OCEANICEPRODUCTS_CONFIG=/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml -+++ config.oceanice_products[16]export NFHRS_PER_GROUP=3 -+++ config.oceanice_products[16]NFHRS_PER_GROUP=3 -+++ config.oceanice_products[18]echo 'END: config.oceanice_products' -END: config.oceanice_products -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env oceanice_products -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=oceanice_products -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 100663296 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 1 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 1 ]] -+++ HERCULES.env[34]max_threads_per_task=80 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 80 ]] -+++ HERCULES.env[40][[ 1 -gt 80 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 1' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[286]export NTHREADS_OCNICEPOST=1 -+++ HERCULES.env[286]NTHREADS_OCNICEPOST=1 -+++ HERCULES.env[287]export 'APRUN_OCNICEPOST=srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1' -+++ HERCULES.env[287]APRUN_OCNICEPOST='srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1' -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_OCEANICE_PRODUCTS[11]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[11]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[11]declare_from_tmpl -rx COMIN_ICE_HISTORY:COM_ICE_HISTORY_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_ICE_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/ice/history -+ JGLOBAL_OCEANICE_PRODUCTS[12]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[12]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[12]declare_from_tmpl -rx COMOUT_ICE_GRIB:COM_ICE_GRIB_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_ICE_GRIB=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ice/grib2 -+ JGLOBAL_OCEANICE_PRODUCTS[13]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[13]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[13]declare_from_tmpl -rx COMOUT_ICE_NETCDF:COM_ICE_NETCDF_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_ICE_NETCDF=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ice/netcdf -+ JGLOBAL_OCEANICE_PRODUCTS[17]/work2/noaa/global/mterry/global-workflow_forked/scripts/exglobal_oceanice_products.py -2025-07-28 21:34:54,440 - INFO - root : BEGIN: __main__.main -2025-07-28 21:34:54,440 - DEBUG - root : ( ) -2025-07-28 21:34:54,448 - INFO - oceanice_products: BEGIN: OceanIceProducts.__init__ -2025-07-28 21:34:54,448 - DEBUG - oceanice_products: ( , {'SHELL': '/bin/bash', 'PDY': datetime.datetime(2021, 3, 23, 0, 0), 'DO_JEDIATMENS': False, 'sqlite_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4', 'COPYGB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/copygb', 'WORK_hercules': '/work/hercules', 'TOCGRIB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib2', 'G2_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/include_d', 'PIO_TYPENAME_VALID_VALUES': ['netcdf', 'netcdf4p', 'netcdf4c', 'pnetcdf'], 'WGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/wgrib', 'SERIAL_CXX': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icpc', 'FHMAX_WAV': 9, 'FIXlut': '/work2/noaa/global/mterry/global-workflow_forked/fix/lut', 'fftw_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj', 'snappy_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx', '_ModuleTable016_': 'bGxOYW1lID0gImdldHRleHQvMC4yMS4xIiwKbG9hZE9yZGVyID0gMTEsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ2V0dGV4dC8wLjIxLjEiLAp3ViA9ICIwMDAwMDAwMDAuMDAwMDAwMDIxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJnaXQtbGZzIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvZ2l0LWxmcy8zLjEuMi5sdWEiLApmdWxsTmFtZSA9ICJnaXQtbGZzLzMuMS4yIiwK', 'MAKE_NSSTBUFR': False, 'cdo_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e', 'curl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops', 'PIO_VERSION_MAJOR': 2, 'threads_per_task': 1, 'SLURM_JOB_USER': 'mterry', '_ModuleTable067_': 'NC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3NuYXBweS8xLjEuMTAubHVhIiwKZnVsbE5hbWUgPSAic25hcHB5LzEuMS4xMCIsCmxvYWRPcmRlciA9IDI4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInNuYXBweS8xLjEuMTAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAxLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKc3AgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45', 'COM_ICE_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}', 'COM_OCEAN_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}', '_ModuleTable022_': 'ZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9oZGY1LzEuMTQuMC5sdWEiLApmdWxsTmFtZSA9ICJoZGY1LzEuMTQuMCIsCmxvYWRPcmRlciA9IDI3LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImhkZjUvMS4xNC4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAxNC4qemZpbmFsIiwKfSwKWyJpbnRlbC1vbmVhcGktY29tcGlsZXJzIl0gPSB7CmZuID0gIi9hcHBzL3NwYWNrLW1hbmFnZWQvbW9kdWxlZmlsZXMvbGludXgtcm9ja3k5LXg4Nl82NC9Db3JlL2ludGVsLW9uZWFwaS1jb21waWxlcnMvMjAyMy4xLjAubHVhIiwKZnVsbE5hbWUgPSAiaW50ZWwtb25lYXBpLWNvbXBpbGVy', 'SLURM_TASKS_PER_NODE': 1, 'COM_ATMOS_OZNMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon', 'COM_ATMOS_GEMPAK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}', 'SLURM_JOB_UID': 9583, 'HISTCONTROL': 'ignoredups', 'SLURM_EXPORT_ENV': 'NONE', 'g2c_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp', 'COM_OCEAN_NETCDF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf', 'GRB2INDEX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grb2index', 'l4densvar': True, 'FHOUT_HF_GFS': 1, 'I_MPI_EXTRA_FILESYSTEM': 1, 'SLURM_TASK_PID': 614996, 'EXECprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/exec', 'g2tmpl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel', 'MPI_BUFS_PER_PROC': 2048, 'IAU_DELTHRS_ENKF': 6, 'COM_WAVE_WMO_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo', '_ModuleTable027_': 'LTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9saWJqcGVnLzIuMS4wLmx1YSIsCmZ1bGxOYW1lID0gImxpYmpwZWcvMi4xLjAiLApsb2FkT3JkZXIgPSAxNiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJsaWJqcGVnLzIuMS4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKbGlicG5nID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4w', 'NMEM_ENS': 0, '__LMOD_REF_COUNT_CLASSPATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/mpi.jar:1', '_ModuleTable050_': 'ZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbWFya3Vwc2FmZS8yLjEuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1tYXJrdXBzYWZlLzIuMS4zIiwKbG9hZE9yZGVyID0gNjEsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktbWFya3Vwc2FmZS8yLjEuMyIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDEuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LW5ldGNkZjQiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFj', 'err': 0, '_ModuleTable032_': 'CnVzZXJOYW1lID0gIm1ldHBsdXMvMy4xLjEiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJtb2R1bGVfYmFzZS5oZXJjdWxlcyJdID0gewpmbiA9ICIvd29yazIvbm9hYS9nbG9iYWwvbXRlcnJ5L2dsb2JhbC13b3JrZmxvd19mb3JrZWQvbW9kdWxlZmlsZXMvbW9kdWxlX2Jhc2UuaGVyY3VsZXMubHVhIiwKZnVsbE5hbWUgPSAibW9kdWxlX2Jhc2UuaGVyY3VsZXMiLApsb2FkT3JkZXIgPSA4NCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDAsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJtb2R1bGVfYmFzZS5oZXJjdWxlcyIsCndWID0gIk0uKnpmaW5hbCIsCn0sCm5jbyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGlj', 'CHGRP_CMD': 'chgrp rstprod', 'PKG_CONFIG_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib/pkgconfig:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/pkgconfig:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/lib/pkgconfig', 'SLURM_LOCALID': 0, 'MAX_TASKS': 25, '_ModuleTable040_': 'aW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9wYXJhbGxlbC1uZXRjZGYvMS4xMi4yLmx1YSIsCmZ1bGxOYW1lID0gInBhcmFsbGVsLW5ldGNkZi8xLjEyLjIiLApsb2FkT3JkZXIgPSAzMiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwYXJhbGxlbC1uZXRjZGYvMS4xMi4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAxMi4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCnBhcmFsbGVsaW8gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0', 'SLURM_SUBMIT_DIR': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', '_ModuleTable077_': 'LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvenN0ZC8xLjUuMi5sdWEiLApmdWxsTmFtZSA9ICJ6c3RkLzEuNS4yIiwKbG9hZE9yZGVyID0gOSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMywKc3RhY2tEZXB0aCA9IDQsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJ6c3RkLzEuNS4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwNS4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCn0sCm1wYXRoQSA9IHsKIi93b3JrMi9ub2FhL2dsb2JhbC9yb2xlLWdsb2JhbC9naXQvRml0Mk9icy92MS4xLjcuMS9tb2R1bGVmaWxlcyIKLCAiL3dvcmsyL25vYWEvZ2xvYmFsL3JvbGUtZ2xvYmFsL2dpdC9wcmVwb2JzL3YxLjEuMC9tb2R1bGVmaWxl', 'HOSTNAME': 'hercules-07-11', 'HISTSIZE': 1000, 'wgrib2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln', 'FRAC_GRID': True, 'G2_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/include_4', 'intel_oneapi_mpi_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse', 'SLURMD_NODENAME': 'hercules-07-11', '_ModuleTable080_': 'bGVzL2xpbnV4LXJvY2t5OS14ODZfNjQvb25lYXBpLzIwMjMuMS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL0NvcmUiCiwgIi93b3JrMi9ub2FhL2dsb2JhbC9tdGVycnkvZ2xvYmFsLXdvcmtmbG93X2ZvcmtlZC9tb2R1bGVmaWxlcyIKLCAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L0NvcmUiCiwgIi9hcHBzL290aGVyL21vZHVsZWZpbGVzIiwgIi9hcHBzL2NvbnRhaW5lcnMvbW9kdWxlZmlsZXMiCiwgIi9hcHBzL2xpY2Vuc2VkL21vZHVsZWZpbGVzIiwg', '__LMOD_REF_COUNT_OCL_ICD_FILENAMES': 'libintelocl_emu.so:1;libalteracl.so:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64/libintelocl.so:1', 'c_blosc_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn', 'FHOUT_OCN': 3, 'COM_OBS_JEDI': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi', 'eccodes_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj', '_ModuleTable037_': 'ay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9uZ2h0dHAyLzEuNTcuMC5sdWEiLApmdWxsTmFtZSA9ICJuZ2h0dHAyLzEuNTcuMCIsCmxvYWRPcmRlciA9IDIyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gNCwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm5naHR0cDIvMS41Ny4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDA1Ny4qemZpbmFsIiwKfSwKb3BlbmJsYXMgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2kt', 'netcdf_c_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy', '_ModuleTable070_': 'c3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzdGFjay1pbnRlbC8yMDIxLjkuMCIsCndWID0gIjAwMDAwMjAyMS4wMDAwMDAwMDkuKnpmaW5hbCIsCn0sClsic3RhY2staW50ZWwtb25lYXBpLW1waSJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3N0YWNrLWludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAubHVhIiwKZnVsbE5hbWUgPSAic3RhY2staW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMCIsCmxvYWRPcmRlciA9IDUsCnByb3BUID0g', 'SLURM_JOB_START_TIME': 1753756458, '_ModuleTable011_': 'b3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZXNtZi84LjYuMCIsCndWID0gIjAwMDAwMDAwOC4wMDAwMDAwMDYuKnpmaW5hbCIsCn0sCmZmdHcgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9mZnR3LzMuMy4xMC5sdWEiLApmdWxsTmFtZSA9ICJmZnR3LzMuMy4xMCIsCmxvYWRPcmRlciA9IDIxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwK', 'DO_NEST': False, 'STMP': '/work2/noaa/stmp/mterry/HERCULES', 'HYDRA_LAUNCHER_EXTRA_ARGS': '--external-launcher', 'ACL_BOARD_VENDOR_PATH': '/opt/Intel/OpenCLFPGA/oneAPI/Boards', '_ModuleTable060_': 'ZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1weXlhbWwvNi4wIiwKd1YgPSAiMDAwMDAwMDA2Lip6ZmluYWwiLAp9LApbInB5LXNldHVwdG9vbHMiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1zZXR1cHRvb2xzLzYzLjQuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1zZXR1cHRvb2xzLzYzLjQuMyIsCmxvYWRPcmRlciA9IDUxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIs', '_ModuleTable007_': 'ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9jcnRtLzIuNC4wLjEubHVhIiwKZnVsbE5hbWUgPSAiY3J0bS8yLjQuMC4xIiwKbG9hZE9yZGVyID0gNDksCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY3J0bS8yLjQuMC4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNC4wMDAwMDAwMDAuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbImNydG0tZml4Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45', 'G2TMPL_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib/libg2tmpl.a', 'DO_WAVE': True, 'SCRgfs': '/work2/noaa/global/mterry/global-workflow_forked/scripts', 'SP_INC8': 'include_8', '_ModuleTable002_': 'b2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9hbnRsci8yLjcuNy5sdWEiLApmdWxsTmFtZSA9ICJhbnRsci8yLjcuNyIsCmxvYWRPcmRlciA9IDM1LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImFudGxyLzIuNy43IiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4wMDAwMDAwMDcuKnpmaW5hbCIsCn0sCmJhY2lvID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2JhY2lvLzIuNC4xLmx1', 'MPICC': 'mpiicc', 'HDF5_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh', '_ModuleTable055_': 'L21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXBhY2thZ2luZy8yMy4xLmx1YSIsCmZ1bGxOYW1lID0gInB5LXBhY2thZ2luZy8yMy4xIiwKbG9hZE9yZGVyID0gNzksCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcGFja2FnaW5nLzIzLjEiLAp3ViA9ICIwMDAwMDAwMjMuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbInB5LXBhbmRhcyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEu', 'SP_INC4': 'include_4', 'FPGA_VARS_DIR': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga', 'FHOUT': 3, '__LMOD_REF_COUNT_MODULEPATH': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles:1;/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0-a66eaip/g2tmpl/1.13.0/intel/2021.9.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/g2tmpl/1.13.0/intel/2021.9.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/intel-oneapi-mpi/2021.9.0-a66eaip/oneapi/2023.1.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core:1;/work2/noaa/global/mterry/global-workflow_forked/modulefiles:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core:1;/apps/other/modulefiles:1;/apps/containers/modulefiles:1;/apps/licensed/modulefiles:1;/apps/contrib/modulefiles:1', 'cycle': 't12z', '_ModuleTable043_': 'aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcmVwb2JzLzEuMS4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKcHJvZF91dGlsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3Byb2RfdXRpbC8yLjEuMS5sdWEiLApmdWxsTmFtZSA9ICJwcm9kX3V0aWwvMi4xLjEiLApsb2FkT3JkZXIgPSA0MCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcm9kX3V0aWwvMi4x', 'MPI_CC': 'mpiicc', 'PARTITION_DTN': '', 'sp_ver': '2.5.0', 'SHELLOPTS': 'braceexpand:errexit:hashall:interactive-comments:nounset:xtrace', 'DO_AWIPS': False, 'EDATE': datetime.datetime(2021, 3, 23, 12, 0), '__LMOD_REF_COUNT___INTEL_POST_CFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1'], 'I_MPI_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0', 'CRTM_FIX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2/fix', '_ModuleTable062_': 'MTYuMCIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMTYuKnpmaW5hbCIsCn0sClsicHkteGFycmF5Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHkteGFycmF5LzIwMjMuNy4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXhhcnJheS8yMDIzLjcuMCIsCmxvYWRPcmRlciA9IDgwLApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXhhcnJheS8yMDIzLjcuMCIsCndWID0gIjAwMDAwMjAyMy4wMDAwMDAwMDcuKnpm', 'proj_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez', 'FSYNC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file', '_ModuleTable005_': 'cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJjLWJsb3NjLzEuMjEuNSIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMjEuMDAwMDAwMDA1Lip6ZmluYWwiLAp9LApjZG8gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9jZG8vMi4yLjAubHVhIiwKZnVsbE5hbWUgPSAiY2RvLzIuMi4wIiwKbG9hZE9yZGVyID0gMjYsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY2RvLzIu', 'g2tmpl_ver': '1.13.0', 'FHOUT_GOES': 3, '_ModuleTable021_': 'MC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nc2wvMi43LjEubHVhIiwKZnVsbE5hbWUgPSAiZ3NsLzIuNy4xIiwKbG9hZE9yZGVyID0gMzYsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ3NsLzIuNy4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sCmhkZjUgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxl', 'EUPD_CYC': 'gdas', 'fms_ver': '2023.02.01', 'util_linux_uuid_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj', '_ModuleTable052_': 'Y2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbnVtZXhwci8yLjguNC5sdWEiLApmdWxsTmFtZSA9ICJweS1udW1leHByLzIuOC40IiwKbG9hZE9yZGVyID0gNjQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktbnVtZXhwci8yLjguNCIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDguMDAwMDAwMDA0Lip6ZmluYWwiLAp9LApbInB5LW51bXB5Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2kt', '_ModuleTable030_': 'cFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJsaWJ5YW1sLzAuMi41IiwKd1YgPSAiMDAwMDAwMDAwLjAwMDAwMDAwMi4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCm1ldCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL21ldC85LjEuMy5sdWEiLApmdWxsTmFtZSA9ICJtZXQvOS4xLjMiLApsb2FkT3JkZXIgPSA3NywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0', 'COM_ATMOS_IMAGERY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery', 'I_MPI_EXTRA_FILESYSTEM_LIST': 'lustre', 'COM_WAVE_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history', 'SLURM_CLUSTER_NAME': 'hercules', '__LMOD_REF_COUNT_ACLOCAL_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/aclocal:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/aclocal:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/aclocal:2', 'SERIAL_CC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icc', 'COM_OBSPROC_TMPL': '${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos', 'SLURM_JOB_END_TIME': 1753757358, '_ModuleTable075_': 'ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3dncmliMi8yLjAuOC5sdWEiLApmdWxsTmFtZSA9ICJ3Z3JpYjIvMi4wLjgiLApsb2FkT3JkZXIgPSA1NSwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJ3Z3JpYjIvMi4wLjgiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKemxpYiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJj', 'FHMAX_GFS': 120, 'SENDAWIP': False, 'G2_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64/libg2_d.a', 'spack_mod_path': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core', 'OCNRES': 500, 'COM_MED_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart', 'SLURM_CPUS_ON_NODE': 1, 'LMOD_FAMILY_METAMPI': 'stack-intel-oneapi-mpi', 'LEVS': 128, 'FIXgsi': '/work2/noaa/global/mterry/global-workflow_forked/fix/gsi', 'NFHRS_PER_GROUP': 3, 'COM_WAVE_PREP_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep', 'FI_PROVIDER_PATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib/prov:/usr/lib64/libfabric', 'python_ver': '3.11.6', 'IAU_OFFSET': 0, 'py_packaging_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6', 'DIAGUTIL_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/sys_check/sys_check.sh', 'python_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2', 'SCRIPTScfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/scripts', 'py_markupsafe_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7', 'G2C_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include', 'py_xlrd_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f', 'LMOD_G2TMPLVIRT_VERSION': '1.13.0', 'pid': 615344, 'OUTPUT_GRID': 'gaussian_grid', 'jobid': 'oceanice_products_ice_f024.615012', 'SLURM_JOB_CPUS_PER_NODE': 1, 'spack_env': 'gsi-addon-dev-fms-2024.01', 'INTEL_ONEAPI_MPI_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse', 'LMOD_FAMILY_METACOMPILER': 'stack-intel', 'SP_INCd': 'include_d', 'py_six_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o', 'COM_ATMOS_GOES_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim', 'MPICXX': 'mpiicpc', 'restart_interval_gdas': 3, 'fit2obs_ver': '1.1.7.1', 'LMOD_DIR': '/apps/other/lmod/lmod/libexec', '_ModuleTable068_': 'LjAvc3AvMi41LjAubHVhIiwKZnVsbE5hbWUgPSAic3AvMi41LjAiLApsb2FkT3JkZXIgPSA0MiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMiwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzcC8yLjUuMCIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCnNxbGl0ZSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9zcWxpdGUvMy40My4yLmx1YSIsCmZ1bGxOYW1lID0gInNxbGl0ZS8zLjQzLjIiLApsb2FkT3Jk', 'REPLAY_ICS': False, 'COM_ICE_NETCDF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf', 'G2_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64/libg2_4.a', '_ModuleTable058_': 'InB5LXB5dHovMjAyMy4zIiwKbG9hZE9yZGVyID0gNjcsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcHl0ei8yMDIzLjMiLAp3ViA9ICIwMDAwMDIwMjMuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LXB5eGxzYiJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXB5eGxzYi8xLjAuMTAubHVhIiwKZnVsbE5hbWUgPSAicHktcHl4bHNiLzEuMC4xMCIsCmxvYWRPcmRlciA9', 'py_setuptools_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr', 'FHMIN': 0, 'COM_ICE_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input', '_ModuleTable048_': 'cHktZjkwbm1sIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktZjkwbm1sLzEuNC4zLmx1YSIsCmZ1bGxOYW1lID0gInB5LWY5MG5tbC8xLjQuMyIsCmxvYWRPcmRlciA9IDU2LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWY5MG5tbC8xLjQuMyIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMDQuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LWppbmphMiJdID0gewpmbiA9ICIvd29yay9ub2FhL2Vw', 'intel_mkl_ver': '2023.1.0', 'PRTE_MCA_plm_slurm_args': '--external-launcher', 'PWD': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ice_f024.615012', 'SLURM_GTIDS': 0, 'LOGNAME': 'mterry', 'MAKE_ACFTBUFR': False, 'W3EMC_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_8', 'HOMEDIR': '/work2/noaa/global/mterry', 'IAU_DELTHRS': 6, 'W3EMC_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_4', 'XDG_SESSION_TYPE': 'unspecified', 'ESMFMKFILE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib/esmf.mk', 'FHMAX': 9, 'WRITE_NSFLIP': True, 'SLURM_JOB_PARTITION': 'hercules', 'YAML_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x', 'PARTITION_SERVICE': 'service', 'MODULESHOME': '/apps/other/lmod/lmod', '__LMOD_REF_COUNT_DYLD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:1', 'COM_WAVE_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart', 'BUFR_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_4.so', 'BUFR_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_8.so', 'HOMEcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1', 'COM_OCEAN_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon', 'DEBUG_POSTSCRIPT': False, 'MANPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/share/man:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/man:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/documentation/en/man/common:/apps/other/lmod/lmod/share/man:/usr/share/man:/apps/share/man:/apps/man:/opt/slurm/share/man::', 'libyaml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x', 'jasper_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt', 'py_xlsxwriter_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok', 'SLURM_TRES_PER_TASK': 'cpu=1', 'nco_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq', 'stack_intel_ver': '2021.9.0', 'SLURM_OOM_KILL_STEP': 0, 'BUFR_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m', 'bufr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m', 'CLUSTERS_DTN': '', 'grib_util_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35', 'nghttp2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky', 'DO_TEST_MODE': True, 'PACKAGEROOT': '/work2/noaa/global/role-global/nwpara', 'lobsdiag_forenkf': True, 'SLURM_JOB_NUM_NODES': 1, 'py_netcdf4_ver': '1.5.8', 'ICERES': 500, 'CXX': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icpc', 'GRIB2GRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grib2grib2', 'DMPDIR': '/work/noaa/rstprod/dump', 'LSOIL_INCR': 2, 'metplus_ver': '3.1.1', '_ModuleTable008_': 'LjAvY3J0bS1maXgvMi40LjAuMV9lbWMubHVhIiwKZnVsbE5hbWUgPSAiY3J0bS1maXgvMi40LjAuMV9lbWMiLApsb2FkT3JkZXIgPSA0NywKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJjcnRtLWZpeC8yLjQuMC4xX2VtYyIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDQuMDAwMDAwMDAwLjAwMDAwMDAwMS4qXy4qZW1jLip6ZmluYWwiLAp9LApjdXJsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIw', 'COM_OCEAN_LETKF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf', 'zlib_ver': '1.2.13', 'BUFR_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_d.so', 'ENKF_SPREAD': True, 'zlib_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j', '__LMOD_REF_COUNT_CMAKE_PREFIX_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r:2;/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/IntelDPCPP:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72:1', '_ModuleTable035_': 'c3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9uZXRjZGYtY3h4NC80LjMuMS5sdWEiLApmdWxsTmFtZSA9ICJuZXRjZGYtY3h4NC80LjMuMSIsCmxvYWRPcmRlciA9IDc2LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm5ldGNkZi1jeHg0LzQuMy4xIiwKd1YgPSAiMDAwMDAwMDA0LjAwMDAwMDAwMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsibmV0Y2RmLWZvcnRyYW4iXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFj', 'FIXgfs': '/work2/noaa/global/mterry/global-workflow_forked/fix', 'DO_COUPLED': True, 'SLURM_JOBID': 5951726, 'SERIAL_FC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'COM_ATMOS_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input', 'W3EMC_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_d', '_ModuleTable013_': 'IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nMi8zLjQuNS5sdWEiLApmdWxsTmFtZSA9ICJnMi8zLjQuNSIsCmxvYWRPcmRlciA9IDQxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImcyLzMuNC41IiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAwNC4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCmcyYyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJj', 'COM_ATMOS_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history', 'FIXam': '/work2/noaa/global/mterry/global-workflow_forked/fix/am', 'hdf5_ver': '1.14.0', 'restart_interval_enkfgdas': 3, 'COM_ATMOS_RADMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon', 'DO_CALC_INCREMENT': False, 'DO_STARTMEM_FROM_JEDIICE': False, 'FIXprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/fix', 'SLURM_JOB_QOS': 'batch', 'I_MPI_HYDRA_BOOTSTRAP_EXEC_EXTRA_ARGS': '--external-launcher', 'obsproc_run_ver': '1.2.0', 'COM_ATMOS_MINMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon', 'crtm_ver': '2.4.0.1', 'FIXorog': '/work2/noaa/global/mterry/global-workflow_forked/fix/orog', 'COM_OCEAN_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history', '__LMOD_REF_COUNT_PATH': '/apps/other/globus-cli-3.35.2/bin:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r/ush:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/bin:2;/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/bin:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/bin/intel64:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/bin:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/bin:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/bin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/bin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin:1;/usr/sbin:1;/usr/bin:1;/apps/sbin:1;/apps/bin:1;/opt/slurm/bin:1;/home/gfekete/sven/bin:1', '_ModuleTable018_': 'KnpmaW5hbCIsCn0sCmdtYWtlID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2dtYWtlLzQuMi4xLmx1YSIsCmZ1bGxOYW1lID0gImdtYWtlLzQuMi4xIiwKbG9hZE9yZGVyID0gNTQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ21ha2UvNC4yLjEiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDAyLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJncmliLXV0aWwiXSA9IHsKZm4gPSAiL3dv', 'COMIN_ICE_HISTORY': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/ice/history', 'COMROOT': '/work2/noaa/global/mterry/RUNTESTS/COMROOT', 'HOME': '/home/mterry', 'FI_PROVIDER': 'mlx', 'met_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b', 'COM_RTOFS_TMPL': '${DMPDIR}', '_ModuleTable015_': 'bnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2cydG1wbC8xLjEzLjAubHVhIiwKZnVsbE5hbWUgPSAiZzJ0bXBsLzEuMTMuMCIsCmxvYWRPcmRlciA9IDQ1LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImcydG1wbC8xLjEzLjAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDEzLip6ZmluYWwiLAp9LApnZXR0ZXh0ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2dldHRleHQvMC4yMS4xLmx1YSIsCmZ1', '_ModuleTable_Sz_': 81, 'LANG': 'C.UTF-8', '__LMOD_REF_COUNT_LIBRARY_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:1', 'DO_BUFRSND': False, 'memory': '96GB', '_ModuleTable064_': 'PSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHkteGxzeHdyaXRlci8zLjEuNy5sdWEiLApmdWxsTmFtZSA9ICJweS14bHN4d3JpdGVyLzMuMS43IiwKbG9hZE9yZGVyID0gNzAsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHkteGxzeHdyaXRlci8zLjEuNyIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDA3Lip6ZmluYWwiLAp9LApbInB5LXhsd3QiXSA9IHsKZm4gPSAi', 'DYLD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib', 'PARTITION_BATCH': 'hercules', '_ModuleTable023_': 'cy8yMDIzLjEuMCIsCmxvYWRPcmRlciA9IDIsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiaW50ZWwtb25lYXBpLWNvbXBpbGVycy8yMDIzLjEuMCIsCndWID0gIjAwMDAwMjAyMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsiaW50ZWwtb25lYXBpLW1rbCJdID0gewpmbiA9ICIvYXBwcy9zcGFjay1tYW5hZ2VkL21vZHVsZWZpbGVzL2xpbnV4LXJvY2t5OS14ODZfNjQvQ29yZS9pbnRlbC1vbmVhcGktbWtsLzIwMjMuMS4wLmx1YSIsCmZ1bGxOYW1lID0gImludGVsLW9uZWFwaS1ta2wvMjAyMy4xLjAiLApsb2FkT3JkZXIgPSA2LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVz', '_ModuleTable081_': 'Ii9hcHBzL2NvbnRyaWIvbW9kdWxlZmlsZXMiLAp9LApzeXN0ZW1CYXNlTVBBVEggPSAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L0NvcmU6L2FwcHMvb3RoZXIvbW9kdWxlZmlsZXM6L2FwcHMvY29udGFpbmVycy9tb2R1bGVmaWxlczovYXBwcy9saWNlbnNlZC9tb2R1bGVmaWxlcyIsCn0K', 'REDERR': '2>', 'PSLOT': 'C48_S2SW', '__LMOD_REF_COUNT_PKG_CONFIG_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib/pkgconfig:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/pkgconfig:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib/pkgconfig:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/pkgconfig:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib/pkgconfig:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/lib/pkgconfig:1', '_ModuleTable036_': 'ay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL25ldGNkZi1mb3J0cmFuLzQuNi4xLmx1YSIsCmZ1bGxOYW1lID0gIm5ldGNkZi1mb3J0cmFuLzQuNi4xIiwKbG9hZE9yZGVyID0gMzEsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAibmV0Y2RmLWZvcnRyYW4vNC42LjEiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDA2LjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKbmdodHRwMiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFj', 'COM_ATMOS_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}', 'USHcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/ush', 'DO_GEMPAK': False, 'SLURM_PROCID': 0, 'USHgfs': '/work2/noaa/global/mterry/global-workflow_forked/ush', 'DOIBP_WAV': False, 'prepobs_run_ver': '1.1.0', 'EXPDIR': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', 'job': 'oceanice_products', 'imp_physics': 8, 'LMOD_SHELL_PRGM': 'bash', 'IAU_FHROT': 0, 'MPIF90': 'mpiifort', 'IP_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_4.a', 'waveGRD': 'uglo_100km', 'COM_CHEM_BMAT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix', 'jasper_ver': '2.0.32', 'SCRATCH': '/scratch/hercules', 'IP_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_8.a', 'CLIENT_GLOBUS_UUID': '869912fe-f6de-46c0-af10-b22efd84a022', 'ACCOUNT': 'fv3-cpu', 'HOMEgfs': '/work2/noaa/global/mterry/global-workflow_forked', 'NHOUR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/nhour', 'binary_diag': False, 'COM_WAVE_GRID_RES_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}', 'SCRIPTSprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/scripts', 'MP_STDOUTMODE': 'ORDERED', 'sigio_ver': '2.3.2', '__LMOD_REF_COUNT_CPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/include:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/include:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/include:1', 'LMOD_SETTARG_FULL_SUPPORT': False, 'COMINsyn': '/work2/noaa/global/role-global/com/gfs/prod/syndat', 'OFFSET_START_HOUR': 0, 'REDOUT': '1>', 'PTMP': '/work2/noaa/stmp/mterry/HERCULES', '_ModuleTable076_': 'dWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC96bGliLzEuMi4xMy5sdWEiLApmdWxsTmFtZSA9ICJ6bGliLzEuMi4xMyIsCmxvYWRPcmRlciA9IDcsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEwLApzdGFja0RlcHRoID0gNSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInpsaWIvMS4yLjEzIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMi4wMDAwMDAwMTMuKnpmaW5hbCIsCn0sCnpzdGQgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2', 'g2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato', 'MODE': 'forecast-only', 'DATAROOT': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312', 'bacio_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj', 'APP': 'S2SW', '_ModuleTable029_': 'LzQuNC4zNSIsCmxvYWRPcmRlciA9IDEyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImxpYnhjcnlwdC80LjQuMzUiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDA0LjAwMDAwMDAzNS4qemZpbmFsIiwKfSwKbGlieWFtbCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9saWJ5YW1sLzAuMi41Lmx1YSIsCmZ1bGxOYW1lID0gImxpYnlhbWwvMC4yLjUiLApsb2FkT3JkZXIgPSA1OSwKcHJv', 'openblas_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f', 'stack_impi_ver': '2021.9.0', 'COM_CHEM_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem', 'TMPDIR': '/local/scratch/mterry/5951726', 'HDF5_PLUGIN_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/plugins', 'DO_TRACKER': True, 'CMAKE_PREFIX_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r:/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/IntelDPCPP:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72', 'crtm_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r', '_ModuleTable038_': 'YWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvb3BlbmJsYXMvMC4zLjI0Lmx1YSIsCmZ1bGxOYW1lID0gIm9wZW5ibGFzLzAuMy4yNCIsCmxvYWRPcmRlciA9IDUwLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm9wZW5ibGFzLzAuMy4yNCIsCndWID0gIjAwMDAwMDAwMC4wMDAwMDAwMDMuMDAwMDAwMDI0Lip6ZmluYWwiLAp9LApvcGVuanBlZyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5z', 'nco_ver': '5.0.6', '_ModuleTable046_': 'IiwKfSwKWyJweS1jZnRpbWUiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1jZnRpbWUvMS4wLjMuNC5sdWEiLApmdWxsTmFtZSA9ICJweS1jZnRpbWUvMS4wLjMuNCIsCmxvYWRPcmRlciA9IDU3LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWNmdGltZS8xLjAuMy40IiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMC4wMDAwMDAwMDMuMDAwMDAwMDA0Lip6ZmluYWwiLAp9', 'GRBINDEX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grbindex', 'libjpeg_turbo_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy', 'MPIF77': 'mpiifort', '_ModuleTable073_': 'LjAwMDAwMDAwMi4wMDAwMDAwMjguKnpmaW5hbCIsCn0sClsidXRpbC1saW51eC11dWlkIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvdXRpbC1saW51eC11dWlkLzIuMzguMS5sdWEiLApmdWxsTmFtZSA9ICJ1dGlsLWxpbnV4LXV1aWQvMi4zOC4xIiwKbG9hZE9yZGVyID0gMTQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAidXRpbC1saW51eC11dWlkLzIuMzguMSIsCndWID0gIjAw', '_ModuleTable010_': 'Mi4wIiwKbG9hZE9yZGVyID0gMjAsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZWNjb2Rlcy8yLjMyLjAiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDMyLip6ZmluYWwiLAp9LAplc21mID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAvZXNtZi84LjYuMC5sdWEiLApmdWxsTmFtZSA9ICJlc21mLzguNi4wIiwKbG9hZE9yZGVyID0gMzQsCnBy', 'NTHSTACK': 1024000000, 'SLURM_CPUS_PER_TASK': 1, 'FIXcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/fix', 'SLURM_NTASKS': 1, 'DO_ICE': True, 'NET': 'gfs', 'COM_ATMOS_GENESIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital', 'HOMEpost': '/work2/noaa/global/mterry/global-workflow_forked', '_ModuleTable061_': 'CnVzZXJOYW1lID0gInB5LXNldHVwdG9vbHMvNjMuNC4zIiwKd1YgPSAiMDAwMDAwMDYzLjAwMDAwMDAwNC4wMDAwMDAwMDMuKnpmaW5hbCIsCn0sClsicHktc2l4Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktc2l4LzEuMTYuMC5sdWEiLApmdWxsTmFtZSA9ICJweS1zaXgvMS4xNi4wIiwKbG9hZE9yZGVyID0gNzMsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktc2l4LzEu', 'bacio_ver': '2.4.1', 'cmake_ver': '3.23.1', 'UTILROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq', '_ModuleTable024_': 'ZXJOYW1lID0gImludGVsLW9uZWFwaS1ta2wvMjAyMy4xLjAiLAp3ViA9ICIwMDAwMDIwMjMuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbImludGVsLW9uZWFwaS1tcGkiXSA9IHsKZm4gPSAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L29uZWFwaS8yMDIzLjEuMC9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLmx1YSIsCmZ1bGxOYW1lID0gImludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAiLApsb2FkT3JkZXIgPSA0LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAiLAp3ViA9ICIwMDAwMDIwMjEuMDAwMDAwMDA5Lip6ZmluYWwi', 'tar_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths', 'ntasks': 1, 'max_tasks_per_node': 80, 'NCDUMP': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump', '_ModuleTable057_': 'LjIubHVhIiwKZnVsbE5hbWUgPSAicHktcHl0aG9uLWRhdGV1dGlsLzIuOC4yIiwKbG9hZE9yZGVyID0gNzQsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcHl0aG9uLWRhdGV1dGlsLzIuOC4yIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwOC4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sClsicHktcHl0eiJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXB5dHovMjAyMy4zLmx1YSIsCmZ1bGxOYW1lID0g', 'py_numpy_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne', 'QUILTING': True, 'IP_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_d.a', 'SLURM_TOPOLOGY_ADDR': 'hercules-07-11', 'libxcrypt_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr', 'CRTM_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include', 'LMOD_VERSION': '8.7.14', 'spack_stack_ver': '1.6.0', '_ModuleTable041_': 'LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9wYXJhbGxlbGlvLzIuNS4xMC5sdWEiLApmdWxsTmFtZSA9ICJwYXJhbGxlbGlvLzIuNS4xMCIsCmxvYWRPcmRlciA9IDMzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInBhcmFsbGVsaW8vMi41LjEwIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNS4wMDAwMDAwMTAuKnpmaW5hbCIsCn0sCnBpZ3ogPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0y', 'HOMEprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0', 'COMPONENT': 'ice', 'BASE_GIT': '/work2/noaa/global/role-global/git', 'FCST_SEGMENTS': [0, 120], 'libpng_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu', '_ModuleTable044_': 'LjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKcHJvaiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9wcm9qLzkuMi4xLmx1YSIsCmZ1bGxOYW1lID0gInByb2ovOS4yLjEiLApsb2FkT3JkZXIgPSAyNCwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcm9qLzkuMi4xIiwKd1YgPSAiMDAwMDAwMDA5LjAwMDAwMDAwMi4wMDAwMDAwMDEuKnpm', 'DO_GENESIS': True, 'BUFR_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_4', '__LMOD_REF_COUNT_DIAGUTIL_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/sys_check/sys_check.sh:1', 'COM_WAVE_GEMPAK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak', 'COM_ATMOS_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2', 'RUN': 'gfs', 'BUFR_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_8', 'COM_ICE_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon', 'py_pandas_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw', '__INTEL_POST_CFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64'], 'ARCDIR': '/work2/noaa/global/mterry/archive/C48_S2SW', '__LMOD_REF_COUNT_NLSPATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64/locale/%l_%t/%N:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin/locale/%l_%t/%N:1', '_ModuleTable069_': 'ZXIgPSAxMywKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMiwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzcWxpdGUvMy40My4yIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDA0My4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sClsic3RhY2staW50ZWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9Db3JlL3N0YWNrLWludGVsLzIwMjEuOS4wLmx1YSIsCmZ1bGxOYW1lID0gInN0YWNrLWludGVsLzIwMjEuOS4wIiwKbG9hZE9yZGVyID0gMywKcHJvcFQgPSB7fSwK', 'SENDECF': False, 'parallelio_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed', 'py_pytz_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s', 'ILPOST': 3, 'WGRIB2_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib/libwgrib2.a', 'parallel_netcdf_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3', 'CMPLR_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0', 'prod_util_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq', 'COMINukmet': '/work2/noaa/global/role-global/data/external_gempak/ukmet', 'libpng_ver': '1.6.37', 'COPYGB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/copygb2', 'NMV': '/bin/mv', 'NOSCRUB': '/work2/noaa/global/mterry', 'SCRATCH_hercules': '/scratch/hercules', 'W3EMC_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_d.a', 'SENDSDM': False, 'envir': 'prod', 'DO_METP': False, 'CRTM_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib/libcrtm.a', 'LOGSCRIPT': '', 'INTEL_ONEAPI_COMPILERS_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72', '_ModuleTable074_': 'MDAwMDAwMi4wMDAwMDAwMzguMDAwMDAwMDAxLip6ZmluYWwiLAp9LAp3M2VtYyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC93M2VtYy8yLjEwLjAubHVhIiwKZnVsbE5hbWUgPSAidzNlbWMvMi4xMC4wIiwKbG9hZE9yZGVyID0gMzksCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAidzNlbWMvMi4xMC4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKd2dyaWIy', 'py_jinja2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj', 'FPGA_VARS_ARGS': '', 'COMOUT_ICE_NETCDF': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ice/netcdf', 'INCVARS_ZERO_STRAT': ["'sphum_inc'", "'liq_wat_inc'", "'icmr_inc'", "'rwmr_inc'", "'snmr_inc'", "'grle_inc'"], 'COM_CONF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf', '_ModuleTable003_': 'YSIsCmZ1bGxOYW1lID0gImJhY2lvLzIuNC4xIiwKbG9hZE9yZGVyID0gMzgsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAzLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiYmFjaW8vMi40LjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDA0LjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKYnVmciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9idWZyLzExLjcuMC5sdWEiLApmdWxsTmFtZSA9ICJidWZyLzExLjcuMCIsCmxvYWRPcmRlciA9', 'HYDRA_BOOTSTRAP': 'slurm', '__LMOD_REF_COUNT___INTEL_POST_FFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1'], 'COM_CHEM_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon', 'MDATE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/mdate', 'WGRIB2': 'wgrib2', '_ModuleTable054_': 'LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktb3BlbnB5eGwvMy4xLjIubHVhIiwKZnVsbE5hbWUgPSAicHktb3BlbnB5eGwvMy4xLjIiLApsb2FkT3JkZXIgPSA2NiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1vcGVucHl4bC8zLjEuMiIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbInB5LXBhY2thZ2luZyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxs', 'EXP_WARM_START': False, 'MODULEPATH_ROOT': '/apps/other/modulefiles', 'COM_SNOW_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon', 'CHGRP_RSTPROD': True, 'py_pyxlsb_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge', 'BACIO_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/include_8', 'restart_interval_gfs': 12, 'DO_GOES': False, 'pgm': '', 'SLURM_TOPOLOGY_ADDR_PATTERN': 'node', 'DO_LAND_IAU': False, 'ntiles': 6, 'FHCYC': 24, 'ROTDIR': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW', 'FHOUT_AERO': 3, 'py_jinja2_ver': '3.1.2', 'BACIO_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/include_4', 'INCREMENTS_TO_ZERO': ["'liq_wat_inc'", "'icmr_inc'", "'rwmr_inc'", "'snmr_inc'", "'grle_inc'"], 'FHR_LIST': 24, 'USHprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/ush', 'BUFR_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_d', 'hdf5_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh', 'INTEL_ONEAPI_MKL_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci', 'SCRIPTSfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/scripts', '__LMOD_REF_COUNT_FI_PROVIDER_PATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib/prov:1;/usr/lib64/libfabric:1', 'FIXcice': '/work2/noaa/global/mterry/global-workflow_forked/fix/cice', 'IAUFHRS': [6], 'XDG_SESSION_CLASS': 'background', 'INTERVAL_GFS': 6, 'py_openpyxl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe', 'FHMAX_FITS': 120, 'COM_ATMOS_MASTER_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master', 'LMOD_PKG': '/apps/other/lmod/lmod', 'MPI_MEMMAP_OFF': 1, 'UUID_HERCULES_DTN': '869912fe-f6de-46c0-af10-b22efd84a022', '_ModuleTable079_': 'L3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAiCiwgIi9hcHBzL3NwYWNrLW1hbmFnZWQvbW9kdWxlZmlsZXMvbGludXgtcm9ja3k5LXg4Nl82NC9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLWE2NmVhaXAvb25lYXBpLzIwMjMuMS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wIgosICIvYXBwcy9zcGFjay1tYW5hZ2VkL21vZHVsZWZp', 'SDATE': datetime.datetime(2021, 3, 23, 12, 0), 'SLURM_SCRIPT_CONTEXT': 'prolog_task', 'MPI_CXX': 'mpiicpc', 'lwrite4danl': True, 'SLURM_MEM_PER_NODE': 98304, 'CASE_ENS': '{{ CASE_ENS }}', '_ModuleTable049_': 'aWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LWppbmphMi8zLjEuMi5sdWEiLApmdWxsTmFtZSA9ICJweS1qaW5qYTIvMy4xLjIiLApsb2FkT3JkZXIgPSA2MiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1qaW5qYTIvMy4xLjIiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDAxLjAwMDAwMDAwMi4qemZpbmFsIiwKfSwKWyJweS1tYXJrdXBzYWZlIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3Vs', 'YAML_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include', 'PYTHONPATH': '/apps/other/globus-cli-3.35.2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:/work2/noaa/global/mterry/global-workflow_forked/sorc/wxflow/src:/work2/noaa/global/mterry/global-workflow_forked/ush/python', '__LMOD_REF_COUNT_PYTHONPATH': '/apps/other/globus-cli-3.35.2:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:3;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:1', 'py_xarray_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4', 'nst_anl': True, 'FHOUT_GFS': 3, 'WORK': '/work/hercules', 'W3EMC_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_4.a', 'F77': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'BASE_DATA': '/work2/noaa/global/role-global/data', 'py_xlwt_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56', 'ACLOCAL_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/aclocal:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/aclocal:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/aclocal', 'QUEUE_SERVICE': 'batch', 'W3EMC_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_8.a', 'DATA': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ice_f024.615012', 'LESSOPEN': '||/usr/bin/lesspipe.sh %s', 'DO_VERFOZN': True, 'antlr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2', 'crtm_fix_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2', 'TOCGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib', 'USER': 'mterry', 'HPSS_PROJECT': 'emc-global', 'FHMAX_WAV_GFS': 120, 'FIXgdas': '/work2/noaa/global/mterry/global-workflow_forked/fix/gdas', 'launcher': 'srun -l --export=ALL --hint=nomultithread', 'NDATE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/ndate', 'LIBRARY_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib', 'gsl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z', 'SLURM_NODELIST': 'hercules-07-11', 'DO_FIT2OBS': True, 'IP_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_4', 'DOIAU_ENKF': True, 'COM_ATMOS_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos', 'py_xarray_ver': '2023.7.0', '__INTEL_POST_FFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64'], 'netcdf_diag': True, 'DOBNDPNT_WAVE': True, 'FETCHDIR': '/NCEPDEV/emc-global/1year/David.Grumm/test_data', 'ENVIRONMENT': 'BATCH', 'gsi_ncdiag_ver': '1.1.2', 'IP_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_8', 'py_bottleneck_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a', 'w3emc_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo', 'py_netcdf4_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i', 'wgrib2_ver': '2.0.8', 'TOCGRIB2SUPER': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib2super', 'COM_ICE_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2', 'KEEPDATA': False, 'udunits_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo', 'cdo_ver': '2.2.0', 'COM_OCEAN_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input', 'FNTSFA': ' ', 'LOADEDMODULES': 'contrib/0.1:intel-oneapi-compilers/2023.1.0:stack-intel/2021.9.0:intel-oneapi-mpi/2021.9.0:stack-intel-oneapi-mpi/2021.9.0:intel-oneapi-mkl/2023.1.0:zlib/1.2.13:pigz/2.7:zstd/1.5.2:tar/1.34:gettext/0.21.1:libxcrypt/4.4.35:sqlite/3.43.2:util-linux-uuid/2.38.1:python/3.11.6:libjpeg/2.1.0:jasper/2.0.32:libpng/1.6.37:openjpeg/2.3.1:eccodes/2.32.0:fftw/3.3.10:nghttp2/1.57.0:curl/8.4.0:proj/9.2.1:udunits/2.2.28:cdo/2.2.0:hdf5/1.14.0:snappy/1.1.10:c-blosc/1.21.5:netcdf-c/4.9.2:netcdf-fortran/4.6.1:parallel-netcdf/1.12.2:parallelio/2.5.10:esmf/8.6.0:antlr/2.7.7:gsl/2.7.1:nco/5.0.6:bacio/2.4.1:w3emc/2.10.0:prod_util/2.1.1:g2/3.4.5:sp/2.5.0:ip/4.3.0:grib-util/1.3.0:g2tmpl/1.13.0:gsi-ncdiag/1.1.2:crtm-fix/2.4.0.1_emc:git-lfs/3.1.2:crtm/2.4.0.1:openblas/0.3.24:py-setuptools/63.4.3:py-numpy/1.23.4:bufr/11.7.0:gmake/4.2.1:wgrib2/2.0.8:py-f90nml/1.4.3:py-cftime/1.0.3.4:py-netcdf4/1.5.8:libyaml/0.2.5:py-pyyaml/6.0:py-markupsafe/2.1.3:py-jinja2/3.1.2:py-bottleneck/1.3.7:py-numexpr/2.8.4:py-et-xmlfile/1.0.1:py-openpyxl/3.1.2:py-pytz/2023.3:py-pyxlsb/1.0.10:py-xlrd/2.0.1:py-xlsxwriter/3.1.7:py-xlwt/1.3.0:py-pandas/1.5.3:py-six/1.16.0:py-python-dateutil/2.8.2:g2c/1.8.0:netcdf-cxx4/4.3.1:met/9.1.3:metplus/3.1.1:py-packaging/23.1:py-xarray/2023.7.0:prepobs/1.1.0:fit2obs/1.1.7.1:globus-cli/3.35.2:module_base.hercules', 'SLURM_JOB_ACCOUNT': 'fv3-cpu', 'DO_OCN': True, 'SLURM_PRIO_PROCESS': 0, 'HOMEfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1', 'gmake_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq', 'FIXfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/fix', 'py_python_dateutil_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy', 'tasks_per_node': 1, 'SLURM_NPROCS': 1, 'COM_CHEM_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history', 'LMOD_ROOT': '/apps/other/lmod', 'DOHYBVAR': '{{ DOHYBVAR }}', 'GSL_ROOT_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z', 'SERIAL_F77': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'IP_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_d', 'SHLVL': 4, '_ModuleTable071_': 'e30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAic3RhY2staW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMCIsCndWID0gIjAwMDAwMjAyMS4wMDAwMDAwMDkuKnpmaW5hbCIsCn0sCnRhciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC90YXIvMS4zNC5sdWEiLApmdWxsTmFtZSA9ICJ0YXIvMS4zNCIsCmxvYWRPcmRlciA9IDEwLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIs', '_ModuleTable012_': 'c3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImZmdHcvMy4zLjEwIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAwMy4wMDAwMDAwMTAuKnpmaW5hbCIsCn0sCmZpdDJvYnMgPSB7CmZuID0gIi93b3JrMi9ub2FhL2dsb2JhbC9yb2xlLWdsb2JhbC9naXQvRml0Mk9icy92MS4xLjcuMS9tb2R1bGVmaWxlcy9maXQyb2JzLzEuMS43LjEubHVhIiwKZnVsbE5hbWUgPSAiZml0Mm9icy8xLjEuNy4xIiwKbG9hZE9yZGVyID0gODIsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZml0Mm9icy8xLjEuNy4xIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4wMDAwMDAwMDcuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApnMiA9', 'INCVARS_EFOLD': 5, 'SLURM_NNODES': 1, 'RESERVATION': '', 'FHMIN_GFS': 0, 'FHMAX_HF_WAV': 48, 'FIXugwd': '/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd', 'BASH_ENV': '/apps/other/lmod/lmod/init/bash', 'CDATE': datetime.datetime(2021, 3, 23, 12, 0), '_ModuleTable063_': 'aW5hbCIsCn0sClsicHkteGxyZCJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXhscmQvMi4wLjEubHVhIiwKZnVsbE5hbWUgPSAicHkteGxyZC8yLjAuMSIsCmxvYWRPcmRlciA9IDY5LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXhscmQvMi4wLjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJweS14bHN4d3JpdGVyIl0g', '_ModuleTable006_': 'Mi4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMi4qemZpbmFsIiwKfSwKY29udHJpYiA9IHsKZm4gPSAiL2FwcHMvb3RoZXIvbW9kdWxlZmlsZXMvY29udHJpYi8wLjEiLApmdWxsTmFtZSA9ICJjb250cmliLzAuMSIsCmxvYWRPcmRlciA9IDEsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAwLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY29udHJpYiIsCndWID0gIjAwMDAwMDAwMC4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sCmNydG0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50', 'FHOUT_OCN_GFS': 6, '_ModuleTable026_': 'cmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2phc3Blci8yLjAuMzIubHVhIiwKZnVsbE5hbWUgPSAiamFzcGVyLzIuMC4zMiIsCmxvYWRPcmRlciA9IDE3LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImphc3Blci8yLjAuMzIiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAzMi4qemZpbmFsIiwKfSwKbGlianBlZyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1z', 'DONST': True, '_ModuleTable051_': 'ay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL3B5LW5ldGNkZjQvMS41LjgubHVhIiwKZnVsbE5hbWUgPSAicHktbmV0Y2RmNC8xLjUuOCIsCmxvYWRPcmRlciA9IDU4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LW5ldGNkZjQvMS41LjgiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA1LjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKWyJweS1udW1leHByIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3Rh', 'LMOD_sys': 'Linux', '_ModuleTable033_': 'L3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL25jby81LjAuNi5sdWEiLApmdWxsTmFtZSA9ICJuY28vNS4wLjYiLApsb2FkT3JkZXIgPSAzNywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJuY28vNS4wLjYiLAp3ViA9ICIwMDAwMDAwMDUuMDAwMDAwMDAwLjAwMDAwMDAwNi4qemZpbmFsIiwKfSwKWyJuZXRjZGYtYyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1', 'FHMIN_WAV': 0, 'LMOD_FAMILY_METACOMPILER_VERSION': '2021.9.0', 'I_MPI_PMI_LIBRARY': '/opt/slurm/lib/libpmi2.so', 'machine': 'HERCULES', 'py_pyyaml_ver': 6.0, 'OCL_ICD_FILENAMES': 'libintelocl_emu.so:libalteracl.so:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64/libintelocl.so', '_ModuleTable017_': 'bG9hZE9yZGVyID0gNDgsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ2l0LWxmcy8zLjEuMiIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbImdsb2J1cy1jbGkiXSA9IHsKZm4gPSAiL2FwcHMvb3RoZXIvbW9kdWxlZmlsZXMvZ2xvYnVzLWNsaS8zLjM1LjIubHVhIiwKZnVsbE5hbWUgPSAiZ2xvYnVzLWNsaS8zLjM1LjIiLApsb2FkT3JkZXIgPSA4MywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJnbG9idXMtY2xpIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAzNS4wMDAwMDAwMDIu', 'DO_FETCH_HPSS': False, 'PROJ_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/proj', 'FHOUT_ICE': 3, 'COM_WAVE_STATION_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station', 'COMINecmwf': '/work2/noaa/global/role-global/data/external_gempak/ecmwf', 'COM_TOP_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}', 'WGRIB2_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include', 'mpmd_opt': '--multi-prog --output=mpmd.%j.%t.out', 'DO_PREP_SFC': False, 'LMOD_G2TMPLVIRT_NAME': 'g2tmpl', '_ModuleTable066_': 'bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5dGhvbi8zLjExLjYubHVhIiwKZnVsbE5hbWUgPSAicHl0aG9uLzMuMTEuNiIsCmxvYWRPcmRlciA9IDE1LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5dGhvbi8zLjExLjYiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDExLjAwMDAwMDAwNi4qemZpbmFsIiwKfSwKc25hcHB5ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAy', '_ModuleTable009_': 'MjEuOS4wL2N1cmwvOC40LjAubHVhIiwKZnVsbE5hbWUgPSAiY3VybC84LjQuMCIsCmxvYWRPcmRlciA9IDIzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImN1cmwvOC40LjAiLAp3ViA9ICIwMDAwMDAwMDguMDAwMDAwMDA0Lip6ZmluYWwiLAp9LAplY2NvZGVzID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2VjY29kZXMvMi4zMi4wLmx1YSIsCmZ1bGxOYW1lID0gImVjY29kZXMvMi4z', 'py_numexpr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq', '__LMOD_REF_COUNT_MANPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/share/man:2;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/man:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/documentation/en/man/common:1;/apps/other/lmod/lmod/share/man:1;/usr/share/man:1;/apps/share/man:1;/apps/man:1;/opt/slurm/share/man:1', 'DOLETKF_OCN': False, 'MPI_GROUP_MAX': 256, 'py_cftime_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i', 'XDG_SESSION_ID': 'c38', 'zstd_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4', 'CLUSTERS': '', '_ModuleTable034_': 'bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAvbmV0Y2RmLWMvNC45LjIubHVhIiwKZnVsbE5hbWUgPSAibmV0Y2RmLWMvNC45LjIiLApsb2FkT3JkZXIgPSAzMCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJuZXRjZGYtYy80LjkuMiIsCndWID0gIjAwMDAwMDAwNC4wMDAwMDAwMDkuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbIm5ldGNkZi1jeHg0Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2st', 'COM_OCEAN_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean', 'CLUSTERS_SERVICE': '', 'pgmout': 'OUTPUT.615344', 'EXECfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/exec', '_ModuleTable031_': 'aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJtZXQvOS4xLjMiLAp3ViA9ICIwMDAwMDAwMDkuMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKbWV0cGx1cyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL21ldHBsdXMvMy4xLjEubHVhIiwKZnVsbE5hbWUgPSAibWV0cGx1cy8zLjEuMSIsCmxvYWRPcmRlciA9IDc4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIs', 'COMOUT_ICE_GRIB': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ice/grib2', 'COM_ATMOS_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon', 'HOMEobsproc': '/work2/noaa/global/role-global/git/obsproc/v1.2.0', 'esmf_ver': '8.6.0', 'DO_ARCHCOM': False, 'SLURM_SUBMIT_HOST': 'hercules-login-1.hpc.msstate.edu', 'WRITE_DOPOST': True, 'EXECcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/exec', 'DUMP_SUFFIX': '', '_ModuleTable072_': 'CnVzZXJOYW1lID0gInRhci8xLjM0IiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAzNC4qemZpbmFsIiwKfSwKdWR1bml0cyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC91ZHVuaXRzLzIuMi4yOC5sdWEiLApmdWxsTmFtZSA9ICJ1ZHVuaXRzLzIuMi4yOCIsCmxvYWRPcmRlciA9IDI1LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInVkdW5pdHMvMi4yLjI4IiwKd1YgPSAiMDAwMDAwMDAy', 'CASE': 'C48', 'SENDDBN': False, 'gettext_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43', 'COM_ICE_BMATRIX_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice', 'git_lfs_ROOT': '/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l', '_ModuleTable001_': 'X01vZHVsZVRhYmxlXyA9IHsKTVR2ZXJzaW9uID0gMywKY19yZWJ1aWxkVGltZSA9IGZhbHNlLApjX3Nob3J0VGltZSA9IGZhbHNlLApkZXB0aFQgPSB7fSwKZmFtaWx5ID0gewpNZXRhQ29tcGlsZXIgPSAic3RhY2staW50ZWwiLApNZXRhTVBJID0gInN0YWNrLWludGVsLW9uZWFwaS1tcGkiLApjb21waWxlciA9ICJpbnRlbC1vbmVhcGktY29tcGlsZXJzIiwKZzJ0bXBsdmlydCA9ICJnMnRtcGwiLAptcGkgPSAiaW50ZWwtb25lYXBpLW1waSIsCn0sCm1UID0gewphbnRsciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9t', 'COM_ICE_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart', '_ModuleTable025_': 'LAp9LAppcCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9pcC80LjMuMC5sdWEiLApmdWxsTmFtZSA9ICJpcC80LjMuMCIsCmxvYWRPcmRlciA9IDQzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImlwLzQuMy4wIiwKd1YgPSAiMDAwMDAwMDA0LjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKamFzcGVyID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hl', 'COM_OCEAN_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart', 'ARCHCOM_TO': 'globus_hpss', 'COM_ATMOS_WMO_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo', 'ip_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf', 'CLASSPATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/mpi.jar', 'INTELFPGAOCLSDKROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga', 'LD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/host/linux64/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/lib:/usr/lib64:/usr/lib:/opt/slurm/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib', 'DO_CA': True, 'g2_ver': '3.4.5', 'LMOD_FAMILY_COMPILER': 'intel-oneapi-compilers', 'XDG_RUNTIME_DIR': '/run/user/9583', 'BASE_ENV': '/work2/noaa/global/mterry/global-workflow_forked/env', 'SLURM_JOB_ID': 5951726, '_ModuleTable014_': 'dWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nMmMvMS44LjAubHVhIiwKZnVsbE5hbWUgPSAiZzJjLzEuOC4wIiwKbG9hZE9yZGVyID0gNzUsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZzJjLzEuOC4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKZzJ0bXBsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9p', 'OCEANICEPRODUCTS_CONFIG': '/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml', 'NTHREADS_OCNICEPOST': 1, 'DO_JEDIOCNVAR': False, 'FIXreg2grb2': '/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2', 'SLURM_NODEID': 0, 'ip_ver': '4.3.0', 'KMP_AFFINITY': 'scatter', 'FHOUT_ICE_GFS': 6, '_ModuleTable065_': 'L3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS14bHd0LzEuMy4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXhsd3QvMS4zLjAiLApsb2FkT3JkZXIgPSA3MSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS14bHd0LzEuMy4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKcHl0aG9uID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1', 'DO_ATM': True, '_ModuleTable004_': 'IDUzLApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImJ1ZnIvMTEuNy4wIiwKd1YgPSAiMDAwMDAwMDExLjAwMDAwMDAwNy4qemZpbmFsIiwKfSwKWyJjLWJsb3NjIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvYy1ibG9zYy8xLjIxLjUubHVhIiwKZnVsbE5hbWUgPSAiYy1ibG9zYy8xLjIxLjUiLApsb2FkT3JkZXIgPSAyOSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1', 'esmf_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep', 'LMOD_FAMILY_MPI_VERSION': '2021.9.0', '_ModuleTable020_': 'c3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9nc2ktbmNkaWFnLzEuMS4yLmx1YSIsCmZ1bGxOYW1lID0gImdzaS1uY2RpYWcvMS4xLjIiLApsb2FkT3JkZXIgPSA0NiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJnc2ktbmNkaWFnLzEuMS4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCmdzbCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYu', 'DO_JEDIATMVAR': False, 'DO_AERO_ANL': False, 'NCLEN': '/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen', 'PS4': '+ $(basename ${BASH_SOURCE[0]:-${FUNCNAME[0]:-"Unknown"}})[${LINENO}]', 'COM_SNOW_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow', '_ModuleTable053_': 'YWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbnVtcHkvMS4yMy40Lmx1YSIsCmZ1bGxOYW1lID0gInB5LW51bXB5LzEuMjMuNCIsCmxvYWRPcmRlciA9IDUyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSA4LApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LW51bXB5LzEuMjMuNCIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMjMuMDAwMDAwMDA0Lip6ZmluYWwiLAp9LApbInB5LW9wZW5weXhsIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0', '__LMOD_REF_COUNT_ACL_BOARD_VENDOR_PATH': '/opt/Intel/OpenCLFPGA/oneAPI/Boards:1', 'LMOD_FAMILY_METAMPI_VERSION': '2021.9.0', '_ModuleTable059_': 'IDY4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXB5eGxzYi8xLjAuMTAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAwLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKWyJweS1weXlhbWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1weXlhbWwvNi4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXB5eWFtbC82LjAiLApsb2FkT3JkZXIgPSA2MCwKcHJvcFQgPSB7fSwKc3RhY2tE', 'DO_CALC_INCREMENT_ENKF_GFS': False, 'MKLROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0', 'DEBUGINFOD_URLS': 'https://debuginfod.centos.org/ ', 'sven_root_path': '/home/gfekete/sven', '_ModuleTable047_': 'LApbInB5LWV0LXhtbGZpbGUiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1ldC14bWxmaWxlLzEuMC4xLmx1YSIsCmZ1bGxOYW1lID0gInB5LWV0LXhtbGZpbGUvMS4wLjEiLApsb2FkT3JkZXIgPSA2NSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1ldC14bWxmaWxlLzEuMC4xIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMC4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsi', 'restart_interval_enkfgfs': 3, 'SENDDBN_NTC': False, 'FIXaer': '/work2/noaa/global/mterry/global-workflow_forked/fix/aer', 'FHMAX_HF_GFS': 48, 'COM_ICE_LETKF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf', 'LMOD_FAMILY_MPI': 'intel-oneapi-mpi', 'netcdf_fortran_ver': '4.6.1', 'FIXmom': '/work2/noaa/global/mterry/global-workflow_forked/fix/mom6', 'cyc': 12, 'BACIO_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib/libbacio_4.a', 'FC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'FHOUT_WAV_GFS': 3, 'SDATE_GFS': datetime.datetime(2021, 3, 23, 12, 0), 'BACIO_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib/libbacio_8.a', 'sp_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr', 'OPS_RES': 'C768', 'NMEM_ENS_GFS': 30, 'which_declare': 'declare -f', '_ModuleTable056_': 'OS4wL3B5LXBhbmRhcy8xLjUuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1wYW5kYXMvMS41LjMiLApsb2FkT3JkZXIgPSA3MiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1wYW5kYXMvMS41LjMiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA1LjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKWyJweS1weXRob24tZGF0ZXV0aWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1weXRob24tZGF0ZXV0aWwvMi44', 'COM_WAVE_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded', 'LMOD_FAMILY_G2TMPLVIRT': 'g2tmpl', 'QUEUE_DTN': 'batch', 'DO_GENESIS_FSU': False, 'prod_util_ver': '2.1.1', 'MPI_F90': 'mpiifort', 'LMOD_FAMILY_COMPILER_VERSION': '2023.1.0', '_ModuleTable042_': 'MDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcGlnei8yLjcubHVhIiwKZnVsbE5hbWUgPSAicGlnei8yLjciLApsb2FkT3JkZXIgPSA4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gNCwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInBpZ3ovMi43IiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4qemZpbmFsIiwKfSwKcHJlcG9icyA9IHsKZm4gPSAiL3dvcmsyL25vYWEvZ2xvYmFsL3JvbGUtZ2xvYmFsL2dpdC9wcmVwb2JzL3YxLjEuMC9tb2R1bGVmaWxlcy9wcmVwb2JzLzEuMS4wLmx1YSIsCmZ1bGxOYW1lID0gInByZXBvYnMvMS4xLjAiLApsb2FkT3JkZXIgPSA4MSwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0', 'openjpeg_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu', 'USHfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/ush', 'YAML_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib', 'DO_GSISOILDA': False, 'nemsio_ver': '2.5.4', 'DO_NPOESS': False, 'SLURM_MPI_TYPE': 'pmi2', 'OMP_STACKSIZE': 2048000, 'py_et_xmlfile_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq', 'FORECAST_HOUR': 24, 'ATARDIR': '/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW', 'EXECgfs': '/work2/noaa/global/mterry/global-workflow_forked/exec', 'py_python_dateutil_ver': '2.8.2', 'assim_freq': 6, 'DOIAU': True, 'FHOUT_WAV': 1, 'py_pyyaml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2', 'met_ver': '9.1.3', 'gsi_ncdiag_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe', 'DO_FETCH_LOCAL': False, 'NLSPATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64/locale/%l_%t/%N:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin/locale/%l_%t/%N', 'netcdf_fortran_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe', 'SP_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_4.a', 'COM_OCEAN_BMATRIX_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean', 'SLURM_CONF': '/var/spool/slurmd/conf-cache/slurm.conf', 'PATH': '/apps/other/globus-cli-3.35.2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r/ush:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/bin:/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/bin/intel64:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/bin:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/bin:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin:/usr/sbin:/usr/bin:/apps/sbin:/apps/bin:/opt/slurm/bin:/home/gfekete/sven/bin', 'SP_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_8.a', 'netcdf_c_ver': '4.9.2', 'SLURM_JOB_NAME': 'C48_S2SW_gfs_ice_prod_f024_12', 'COM_ICE_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history', 'NLN': '/bin/ln -sf', 'MODULEPATH': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles:/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0-a66eaip/g2tmpl/1.13.0/intel/2021.9.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/g2tmpl/1.13.0/intel/2021.9.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/intel-oneapi-mpi/2021.9.0-a66eaip/oneapi/2023.1.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core:/work2/noaa/global/mterry/global-workflow_forked/modulefiles:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core:/apps/other/modulefiles:/apps/containers/modulefiles:/apps/licensed/modulefiles:/apps/contrib/modulefiles', 'CC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icc', 'grib_util_ver': '1.3.0', 'py_f90nml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z', 'DO_MERGENSST': False, 'SLURM_NTASKS_PER_NODE': 1, '_LMFILES_': '/apps/other/modulefiles/contrib/0.1:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core/intel-oneapi-compilers/2023.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core/stack-intel/2021.9.0.lua:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0/intel-oneapi-mpi/2021.9.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/stack-intel-oneapi-mpi/2021.9.0.lua:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core/intel-oneapi-mkl/2023.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/zlib/1.2.13.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/pigz/2.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/zstd/1.5.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/tar/1.34.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gettext/0.21.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libxcrypt/4.4.35.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/sqlite/3.43.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/util-linux-uuid/2.38.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/python/3.11.6.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libjpeg/2.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/jasper/2.0.32.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libpng/1.6.37.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/openjpeg/2.3.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/eccodes/2.32.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/fftw/3.3.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/nghttp2/1.57.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/curl/8.4.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/proj/9.2.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/udunits/2.2.28.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/cdo/2.2.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/hdf5/1.14.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/snappy/1.1.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/c-blosc/1.21.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-c/4.9.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-fortran/4.6.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/parallel-netcdf/1.12.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/parallelio/2.5.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/esmf/8.6.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/antlr/2.7.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gsl/2.7.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/nco/5.0.6.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/bacio/2.4.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/w3emc/2.10.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/prod_util/2.1.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2/3.4.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/sp/2.5.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/ip/4.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/grib-util/1.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2tmpl/1.13.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/gsi-ncdiag/1.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/crtm-fix/2.4.0.1_emc.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/git-lfs/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/crtm/2.4.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/openblas/0.3.24.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-setuptools/63.4.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-numpy/1.23.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/bufr/11.7.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gmake/4.2.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/wgrib2/2.0.8.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-f90nml/1.4.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-cftime/1.0.3.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/py-netcdf4/1.5.8.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libyaml/0.2.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pyyaml/6.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-markupsafe/2.1.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-jinja2/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-bottleneck/1.3.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-numexpr/2.8.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-et-xmlfile/1.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-openpyxl/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pytz/2023.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pyxlsb/1.0.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlrd/2.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlsxwriter/3.1.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlwt/1.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pandas/1.5.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-six/1.16.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-python-dateutil/2.8.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2c/1.8.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-cxx4/4.3.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/met/9.1.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/metplus/3.1.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-packaging/23.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xarray/2023.7.0.lua:/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles/prepobs/1.1.0.lua:/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles/fit2obs/1.1.7.1.lua:/apps/other/modulefiles/globus-cli/3.35.2.lua:/work2/noaa/global/mterry/global-workflow_forked/modulefiles/module_base.hercules.lua', 'globus_cli_ver': 3.27, '_ModuleTable028_': 'L2xpYnBuZy8xLjYuMzcubHVhIiwKZnVsbE5hbWUgPSAibGlicG5nLzEuNi4zNyIsCmxvYWRPcmRlciA9IDE4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImxpYnBuZy8xLjYuMzciLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA2LjAwMDAwMDAzNy4qemZpbmFsIiwKfSwKbGlieGNyeXB0ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2xpYnhjcnlwdC80LjQuMzUubHVhIiwKZnVsbE5hbWUgPSAibGlieGNyeXB0', 'py_numpy_ver': '1.23.4', '__LMOD_REF_COUNT_HDF5_PLUGIN_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/plugins:1', 'bufr_ver': '11.7.0', 'COMINnam': '/work2/noaa/global/role-global/data/external_gempak/nam', 'DBUS_SESSION_BUS_ADDRESS': 'unix:path=/run/user/9583/bus', 'DO_AERO_FCST': False, 'py_pandas_ver': '1.5.3', '_ModuleTable039_': 'dGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9vcGVuanBlZy8yLjMuMS5sdWEiLApmdWxsTmFtZSA9ICJvcGVuanBlZy8yLjMuMSIsCmxvYWRPcmRlciA9IDE5LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm9wZW5qcGVnLzIuMy4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsicGFyYWxsZWwtbmV0Y2RmIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMv', 'COM_ICE_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice', 'UUID_ORION_DTN': '8a10dd4f-24ee-4794-a39d-9c313ab6a34b', 'MPI_F77': 'mpiifort', 'LMOD_CMD': '/apps/other/lmod/lmod/libexec/lmod', '_ModuleTable045_': 'aW5hbCIsCn0sClsicHktYm90dGxlbmVjayJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LWJvdHRsZW5lY2svMS4zLjcubHVhIiwKZnVsbE5hbWUgPSAicHktYm90dGxlbmVjay8xLjMuNyIsCmxvYWRPcmRlciA9IDYzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWJvdHRsZW5lY2svMS4zLjciLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAzLjAwMDAwMDAwNy4qemZpbmFs', 'LMOD_MPI_VERSION': '2021.9.0-a66eaip', 'LMOD_SYSTEM_NAME': 'hercules', 'RUN_ENVIR': 'emc', 'w3emc_ver': '2.10.0', '__LMOD_REF_COUNT_LD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/host/linux64/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/lib:1;/usr/lib64:1;/usr/lib:1;/opt/slurm/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:2', 'DOHYBVAR_OCN': False, 'NUM_SND_COLLECTIVES': 9, 'walltime': '00:15:00', 'netcdf_cxx4_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu', 'ens_tracker_ver': 'v1.2.0', 'LMOD_SYSTEM_DEFAULT_MODULES': 'contrib', 'OMPI_MCA_plm_slurm_args': '--external-launcher', '_ModuleTable078_': 'cyIKLCAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLWE2NmVhaXAvZzJ0bXBsLzEuMTMuMC9pbnRlbC8yMDIxLjkuMCIKLCAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9nMnRtcGwvMS4xMy4wL2ludGVsLzIwMjEuOS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVz', 'FIXcpl': '/work2/noaa/global/mterry/global-workflow_forked/fix/cpl', 'G2C_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64/libg2c.so', '_ModuleTable019_': 'cmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9ncmliLXV0aWwvMS4zLjAubHVhIiwKZnVsbE5hbWUgPSAiZ3JpYi11dGlsLzEuMy4wIiwKbG9hZE9yZGVyID0gNDQsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ3JpYi11dGlsLzEuMy4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKWyJnc2ktbmNkaWFnIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMv', 'SLURM_JOB_GID': 17000, 'SLURM_GET_USER_ENV': 1, 'IAUFHRS_ENKF': [3, 6, 9], 'CPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/include:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/include:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/include', 'G2TMPL_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include', 'DO_JEDISNOWDA': False, 'SP_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_d.a', 'pigz_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq', 'COM_ATMOS_BUFR_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr', 'COM_OCEAN_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2', 'CNVGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/cnvgrib', 'COM_ATMOS_TRACK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks', 'DO_VERFRAD': True, 'pgmerr': 'errfile', 'MPI_BUFS_PER_HOST': 2048, 'DO_PREP_OBS_AERO': False, 'LMOD_FAMILY_G2TMPLVIRT_VERSION': '1.13.0', 'PNG_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu', 'COM_ATMOS_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart', 'DEGRIB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/degrib2', 'DO_ANLSTAT': False, 'LMOD_MPI_NAME': 'intel-oneapi-mpi', 'py_f90nml_ver': '1.4.3', 'NMEM_ENS_GFS_OFFSET': 20, 'SMOOTH_ENKF': False, 'FHMAX_GOES': 120, 'OLDPWD': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', 'QUEUE': 'batch', 'COM_OBS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/obs', 'DO_VMINMON': True, 'metplus_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r', 'VERBOSE': True, 'SLURM_JOB_NODELIST': 'hercules-07-11', 'CLUSTER': 'hercules', 'DBNROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn', 'FHOUT_HF_WAV': 1, 'PARMgfs': '/work2/noaa/global/mterry/global-workflow_forked/parm', 'I_MPI_HYDRA_BOOTSTRAP': 'slurm', 'APRUN_OCNICEPOST': 'srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1', 'BASH_FUNC_ml%%': '() { eval "$($LMOD_DIR/ml_cmd "$@")"\n}', 'BASH_FUNC_which%%': '() { ( alias;\n eval ${which_declare} ) | /usr/bin/which --tty-only --read-alias --read-functions --show-tilde --show-dot $@\n}', 'BASH_FUNC_module%%': '() { if [ -z "${LMOD_SH_DBG_ON+x}" ]; then\n case "$-" in \n *v*x*)\n __lmod_sh_dbg=\'vx\'\n ;;\n *v*)\n __lmod_sh_dbg=\'v\'\n ;;\n *x*)\n __lmod_sh_dbg=\'x\'\n ;;\n esac;\n fi;\n if [ -n "${__lmod_sh_dbg:-}" ]; then\n set +$__lmod_sh_dbg;\n echo "Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for Lmod\'s output" 1>&2;\n fi;\n eval "$($LMOD_CMD $LMOD_SHELL_PRGM "$@")" && eval "$(${LMOD_SETTARG_CMD:-:} -s sh)";\n __lmod_my_status=$?;\n if [ -n "${__lmod_sh_dbg:-}" ]; then\n echo "Shell debugging restarted" 1>&2;\n set -$__lmod_sh_dbg;\n fi;\n unset __lmod_sh_dbg;\n return $__lmod_my_status\n}', 'BASH_FUNC_declare_from_tmpl%%': '() { if [[ ${DEBUG_WORKFLOW:-"NO"} == "NO" ]]; then\n set +x;\n fi;\n local opts="-g";\n local OPTIND=1;\n while getopts "rx" option; do\n opts="${opts}${option}";\n done;\n shift $((OPTIND-1));\n for input in "$@";\n do\n IFS=\':\' read -ra args <<< "${input}";\n local com_var="${args[0]}";\n local template;\n local value;\n if (( ${#args[@]} > 1 )); then\n template="${args[1]}";\n else\n template="${com_var}_TMPL";\n fi;\n if [[ ! -v "${template}" ]]; then\n echo "FATAL ERROR in declare_from_tmpl: Requested template ${template} not defined!";\n exit 2;\n fi;\n value=$(echo "${!template}" | envsubst);\n declare ${opts} "${com_var}"="${value}";\n echo "declare_from_tmpl :: ${com_var}=${value}";\n done;\n set_trace\n}', 'BASH_FUNC_err_exit%%': ['() { set +eux;\n msg1=${*:-Job ${jobid} failed};\n if [[ -n "${pgm}" ]]; then\n msg1+="', 'ERROR IN ${pgm}";\n fi;\n if [[ -n "${err}" ]]; then\n msg1+=" RETURN CODE ${err}";\n fi;\n msg2="\n -------------------------------------------------------------\n -- FATAL ERROR: ${msg1}\n -- ABNORMAL EXIT at $(date) on ${HOSTNAME}\n -------------------------------------------------------------\n ";\n echo "${msg2}" 1>&2;\n module list;\n echo "" 1>&2;\n echo "${msg1}" 1>&2;\n if [[ -n "${DATA}" ]]; then\n echo "${DATA}" 1>&2;\n ls -ltr "${DATA}" 1>&2;\n else\n echo "WARNING: DATA variable not defined" 1>&2;\n fi;\n if [[ -n "${pgmout}" ]]; then\n if [[ -s errfile ]]; then\n echo "----- contents of errfile -----" >> "${pgmout}";\n cat errfile >> "${pgmout}";\n fi;\n cat "${pgmout}" 1>&2;\n else\n if [[ -s errfile ]]; then\n cat errfile 1>&2;\n fi;\n fi;\n if [[ "${SENDECF}" == "YES" ]]; then\n timeout 30 ecflow_client --msg "${ECF_NAME}: ${msg1}";\n timeout 30 ssh "${ECF_HOST}" "echo \\"${msg}2\\" >> ${ECF_JOBOUT:?}";\n fi;\n if [[ "${SENDECF}" == "YES" ]]; then\n ecflow_client --kill="${ECF_NAME:?}";\n fi;\n if [[ -n "${PBS_JOBID}" ]]; then\n qdel "${PBS_JOBID}";\n else\n if [[ -n "${SLURM_JOB_ID}" ]]; then\n scancel "${SLURM_JOB_ID}";\n fi;\n fi\n}'], 'BASH_FUNC_wait_for_file%%': '() { set +x;\n local file_name=${1:?"wait_for_file() requires a file name"};\n local sleep_interval=${2:-60};\n local max_tries=${3:-100};\n for ((iter=0; iter' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.oceanice_products -+++ config.oceanice_products[5]echo 'BEGIN: config.oceanice_products' -BEGIN: config.oceanice_products -+++ config.oceanice_products[8]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources oceanice_products -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=oceanice_products -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[978]walltime=00:15:00 -++++ config.resources[979]ntasks=1 -++++ config.resources[980]tasks_per_node=1 -++++ config.resources[981]threads_per_task=1 -++++ config.resources[982]memory=96GB -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n '' ]] -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export memory -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.oceanice_products[11]export MAX_TASKS=25 -+++ config.oceanice_products[11]MAX_TASKS=25 -+++ config.oceanice_products[13]export OCEANICEPRODUCTS_CONFIG=/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml -+++ config.oceanice_products[13]OCEANICEPRODUCTS_CONFIG=/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml -+++ config.oceanice_products[16]export NFHRS_PER_GROUP=3 -+++ config.oceanice_products[16]NFHRS_PER_GROUP=3 -+++ config.oceanice_products[18]echo 'END: config.oceanice_products' -END: config.oceanice_products -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env oceanice_products -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=oceanice_products -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 100663296 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 1 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 1 ]] -+++ HERCULES.env[34]max_threads_per_task=80 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 80 ]] -+++ HERCULES.env[40][[ 1 -gt 80 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 1' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[286]export NTHREADS_OCNICEPOST=1 -+++ HERCULES.env[286]NTHREADS_OCNICEPOST=1 -+++ HERCULES.env[287]export 'APRUN_OCNICEPOST=srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1' -+++ HERCULES.env[287]APRUN_OCNICEPOST='srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1' -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_OCEANICE_PRODUCTS[11]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[11]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[11]declare_from_tmpl -rx COMIN_ICE_HISTORY:COM_ICE_HISTORY_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_ICE_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/ice/history -+ JGLOBAL_OCEANICE_PRODUCTS[12]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[12]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[12]declare_from_tmpl -rx COMOUT_ICE_GRIB:COM_ICE_GRIB_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_ICE_GRIB=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ice/grib2 -+ JGLOBAL_OCEANICE_PRODUCTS[13]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[13]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[13]declare_from_tmpl -rx COMOUT_ICE_NETCDF:COM_ICE_NETCDF_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_ICE_NETCDF=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ice/netcdf -+ JGLOBAL_OCEANICE_PRODUCTS[17]/work2/noaa/global/mterry/global-workflow_forked/scripts/exglobal_oceanice_products.py -2025-07-28 21:34:54,440 - INFO - root : BEGIN: __main__.main -2025-07-28 21:34:54,441 - DEBUG - root : ( ) -2025-07-28 21:34:54,448 - INFO - oceanice_products: BEGIN: OceanIceProducts.__init__ -2025-07-28 21:34:54,449 - DEBUG - oceanice_products: ( , {'SHELL': '/bin/bash', 'PDY': datetime.datetime(2021, 3, 23, 0, 0), 'DO_JEDIATMENS': False, 'sqlite_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4', 'COPYGB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/copygb', 'WORK_hercules': '/work/hercules', 'TOCGRIB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib2', 'G2_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/include_d', 'PIO_TYPENAME_VALID_VALUES': ['netcdf', 'netcdf4p', 'netcdf4c', 'pnetcdf'], 'WGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/wgrib', 'SERIAL_CXX': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icpc', 'FHMAX_WAV': 9, 'FIXlut': '/work2/noaa/global/mterry/global-workflow_forked/fix/lut', 'fftw_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj', 'snappy_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx', '_ModuleTable016_': 'bGxOYW1lID0gImdldHRleHQvMC4yMS4xIiwKbG9hZE9yZGVyID0gMTEsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ2V0dGV4dC8wLjIxLjEiLAp3ViA9ICIwMDAwMDAwMDAuMDAwMDAwMDIxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJnaXQtbGZzIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvZ2l0LWxmcy8zLjEuMi5sdWEiLApmdWxsTmFtZSA9ICJnaXQtbGZzLzMuMS4yIiwK', 'MAKE_NSSTBUFR': False, 'cdo_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e', 'curl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops', 'PIO_VERSION_MAJOR': 2, 'threads_per_task': 1, 'SLURM_JOB_USER': 'mterry', '_ModuleTable067_': 'NC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3NuYXBweS8xLjEuMTAubHVhIiwKZnVsbE5hbWUgPSAic25hcHB5LzEuMS4xMCIsCmxvYWRPcmRlciA9IDI4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInNuYXBweS8xLjEuMTAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAxLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKc3AgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45', 'COM_ICE_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}', 'COM_OCEAN_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}', '_ModuleTable022_': 'ZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9oZGY1LzEuMTQuMC5sdWEiLApmdWxsTmFtZSA9ICJoZGY1LzEuMTQuMCIsCmxvYWRPcmRlciA9IDI3LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImhkZjUvMS4xNC4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAxNC4qemZpbmFsIiwKfSwKWyJpbnRlbC1vbmVhcGktY29tcGlsZXJzIl0gPSB7CmZuID0gIi9hcHBzL3NwYWNrLW1hbmFnZWQvbW9kdWxlZmlsZXMvbGludXgtcm9ja3k5LXg4Nl82NC9Db3JlL2ludGVsLW9uZWFwaS1jb21waWxlcnMvMjAyMy4xLjAubHVhIiwKZnVsbE5hbWUgPSAiaW50ZWwtb25lYXBpLWNvbXBpbGVy', 'SLURM_TASKS_PER_NODE': 1, 'COM_ATMOS_OZNMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon', 'COM_ATMOS_GEMPAK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}', 'SLURM_JOB_UID': 9583, 'HISTCONTROL': 'ignoredups', 'SLURM_EXPORT_ENV': 'NONE', 'g2c_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp', 'COM_OCEAN_NETCDF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf', 'GRB2INDEX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grb2index', 'l4densvar': True, 'FHOUT_HF_GFS': 1, 'I_MPI_EXTRA_FILESYSTEM': 1, 'SLURM_TASK_PID': 358186, 'EXECprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/exec', 'g2tmpl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel', 'MPI_BUFS_PER_PROC': 2048, 'IAU_DELTHRS_ENKF': 6, 'COM_WAVE_WMO_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo', '_ModuleTable027_': 'LTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9saWJqcGVnLzIuMS4wLmx1YSIsCmZ1bGxOYW1lID0gImxpYmpwZWcvMi4xLjAiLApsb2FkT3JkZXIgPSAxNiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJsaWJqcGVnLzIuMS4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKbGlicG5nID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4w', 'NMEM_ENS': 0, '__LMOD_REF_COUNT_CLASSPATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/mpi.jar:1', '_ModuleTable050_': 'ZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbWFya3Vwc2FmZS8yLjEuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1tYXJrdXBzYWZlLzIuMS4zIiwKbG9hZE9yZGVyID0gNjEsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktbWFya3Vwc2FmZS8yLjEuMyIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDEuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LW5ldGNkZjQiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFj', 'err': 0, '_ModuleTable032_': 'CnVzZXJOYW1lID0gIm1ldHBsdXMvMy4xLjEiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJtb2R1bGVfYmFzZS5oZXJjdWxlcyJdID0gewpmbiA9ICIvd29yazIvbm9hYS9nbG9iYWwvbXRlcnJ5L2dsb2JhbC13b3JrZmxvd19mb3JrZWQvbW9kdWxlZmlsZXMvbW9kdWxlX2Jhc2UuaGVyY3VsZXMubHVhIiwKZnVsbE5hbWUgPSAibW9kdWxlX2Jhc2UuaGVyY3VsZXMiLApsb2FkT3JkZXIgPSA4NCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDAsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJtb2R1bGVfYmFzZS5oZXJjdWxlcyIsCndWID0gIk0uKnpmaW5hbCIsCn0sCm5jbyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGlj', 'CHGRP_CMD': 'chgrp rstprod', 'PKG_CONFIG_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib/pkgconfig:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/pkgconfig:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/lib/pkgconfig', 'SLURM_LOCALID': 0, 'MAX_TASKS': 25, '_ModuleTable040_': 'aW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9wYXJhbGxlbC1uZXRjZGYvMS4xMi4yLmx1YSIsCmZ1bGxOYW1lID0gInBhcmFsbGVsLW5ldGNkZi8xLjEyLjIiLApsb2FkT3JkZXIgPSAzMiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwYXJhbGxlbC1uZXRjZGYvMS4xMi4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAxMi4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCnBhcmFsbGVsaW8gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0', 'SLURM_SUBMIT_DIR': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', '_ModuleTable077_': 'LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvenN0ZC8xLjUuMi5sdWEiLApmdWxsTmFtZSA9ICJ6c3RkLzEuNS4yIiwKbG9hZE9yZGVyID0gOSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMywKc3RhY2tEZXB0aCA9IDQsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJ6c3RkLzEuNS4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwNS4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCn0sCm1wYXRoQSA9IHsKIi93b3JrMi9ub2FhL2dsb2JhbC9yb2xlLWdsb2JhbC9naXQvRml0Mk9icy92MS4xLjcuMS9tb2R1bGVmaWxlcyIKLCAiL3dvcmsyL25vYWEvZ2xvYmFsL3JvbGUtZ2xvYmFsL2dpdC9wcmVwb2JzL3YxLjEuMC9tb2R1bGVmaWxl', 'HOSTNAME': 'hercules-08-38', 'HISTSIZE': 1000, 'wgrib2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln', 'FRAC_GRID': True, 'G2_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/include_4', 'intel_oneapi_mpi_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse', 'SLURMD_NODENAME': 'hercules-08-38', '_ModuleTable080_': 'bGVzL2xpbnV4LXJvY2t5OS14ODZfNjQvb25lYXBpLzIwMjMuMS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL0NvcmUiCiwgIi93b3JrMi9ub2FhL2dsb2JhbC9tdGVycnkvZ2xvYmFsLXdvcmtmbG93X2ZvcmtlZC9tb2R1bGVmaWxlcyIKLCAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L0NvcmUiCiwgIi9hcHBzL290aGVyL21vZHVsZWZpbGVzIiwgIi9hcHBzL2NvbnRhaW5lcnMvbW9kdWxlZmlsZXMiCiwgIi9hcHBzL2xpY2Vuc2VkL21vZHVsZWZpbGVzIiwg', '__LMOD_REF_COUNT_OCL_ICD_FILENAMES': 'libintelocl_emu.so:1;libalteracl.so:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64/libintelocl.so:1', 'c_blosc_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn', 'FHOUT_OCN': 3, 'COM_OBS_JEDI': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi', 'eccodes_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj', '_ModuleTable037_': 'ay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9uZ2h0dHAyLzEuNTcuMC5sdWEiLApmdWxsTmFtZSA9ICJuZ2h0dHAyLzEuNTcuMCIsCmxvYWRPcmRlciA9IDIyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gNCwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm5naHR0cDIvMS41Ny4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDA1Ny4qemZpbmFsIiwKfSwKb3BlbmJsYXMgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2kt', 'netcdf_c_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy', '_ModuleTable070_': 'c3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzdGFjay1pbnRlbC8yMDIxLjkuMCIsCndWID0gIjAwMDAwMjAyMS4wMDAwMDAwMDkuKnpmaW5hbCIsCn0sClsic3RhY2staW50ZWwtb25lYXBpLW1waSJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3N0YWNrLWludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAubHVhIiwKZnVsbE5hbWUgPSAic3RhY2staW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMCIsCmxvYWRPcmRlciA9IDUsCnByb3BUID0g', 'SLURM_JOB_START_TIME': 1753756458, '_ModuleTable011_': 'b3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZXNtZi84LjYuMCIsCndWID0gIjAwMDAwMDAwOC4wMDAwMDAwMDYuKnpmaW5hbCIsCn0sCmZmdHcgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9mZnR3LzMuMy4xMC5sdWEiLApmdWxsTmFtZSA9ICJmZnR3LzMuMy4xMCIsCmxvYWRPcmRlciA9IDIxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwK', 'DO_NEST': False, 'STMP': '/work2/noaa/stmp/mterry/HERCULES', 'HYDRA_LAUNCHER_EXTRA_ARGS': '--external-launcher', 'ACL_BOARD_VENDOR_PATH': '/opt/Intel/OpenCLFPGA/oneAPI/Boards', '_ModuleTable060_': 'ZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1weXlhbWwvNi4wIiwKd1YgPSAiMDAwMDAwMDA2Lip6ZmluYWwiLAp9LApbInB5LXNldHVwdG9vbHMiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1zZXR1cHRvb2xzLzYzLjQuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1zZXR1cHRvb2xzLzYzLjQuMyIsCmxvYWRPcmRlciA9IDUxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIs', '_ModuleTable007_': 'ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9jcnRtLzIuNC4wLjEubHVhIiwKZnVsbE5hbWUgPSAiY3J0bS8yLjQuMC4xIiwKbG9hZE9yZGVyID0gNDksCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY3J0bS8yLjQuMC4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNC4wMDAwMDAwMDAuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbImNydG0tZml4Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45', 'G2TMPL_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib/libg2tmpl.a', 'DO_WAVE': True, 'SCRgfs': '/work2/noaa/global/mterry/global-workflow_forked/scripts', 'SP_INC8': 'include_8', '_ModuleTable002_': 'b2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9hbnRsci8yLjcuNy5sdWEiLApmdWxsTmFtZSA9ICJhbnRsci8yLjcuNyIsCmxvYWRPcmRlciA9IDM1LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImFudGxyLzIuNy43IiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4wMDAwMDAwMDcuKnpmaW5hbCIsCn0sCmJhY2lvID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2JhY2lvLzIuNC4xLmx1', 'MPICC': 'mpiicc', 'HDF5_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh', '_ModuleTable055_': 'L21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXBhY2thZ2luZy8yMy4xLmx1YSIsCmZ1bGxOYW1lID0gInB5LXBhY2thZ2luZy8yMy4xIiwKbG9hZE9yZGVyID0gNzksCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcGFja2FnaW5nLzIzLjEiLAp3ViA9ICIwMDAwMDAwMjMuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbInB5LXBhbmRhcyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEu', 'SP_INC4': 'include_4', 'FPGA_VARS_DIR': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga', 'FHOUT': 3, '__LMOD_REF_COUNT_MODULEPATH': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles:1;/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0-a66eaip/g2tmpl/1.13.0/intel/2021.9.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/g2tmpl/1.13.0/intel/2021.9.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/intel-oneapi-mpi/2021.9.0-a66eaip/oneapi/2023.1.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core:1;/work2/noaa/global/mterry/global-workflow_forked/modulefiles:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core:1;/apps/other/modulefiles:1;/apps/containers/modulefiles:1;/apps/licensed/modulefiles:1;/apps/contrib/modulefiles:1', 'cycle': 't12z', '_ModuleTable043_': 'aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcmVwb2JzLzEuMS4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKcHJvZF91dGlsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3Byb2RfdXRpbC8yLjEuMS5sdWEiLApmdWxsTmFtZSA9ICJwcm9kX3V0aWwvMi4xLjEiLApsb2FkT3JkZXIgPSA0MCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcm9kX3V0aWwvMi4x', 'MPI_CC': 'mpiicc', 'PARTITION_DTN': '', 'sp_ver': '2.5.0', 'SHELLOPTS': 'braceexpand:errexit:hashall:interactive-comments:nounset:xtrace', 'DO_AWIPS': False, 'EDATE': datetime.datetime(2021, 3, 23, 12, 0), '__LMOD_REF_COUNT___INTEL_POST_CFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1'], 'I_MPI_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0', 'CRTM_FIX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2/fix', '_ModuleTable062_': 'MTYuMCIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMTYuKnpmaW5hbCIsCn0sClsicHkteGFycmF5Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHkteGFycmF5LzIwMjMuNy4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXhhcnJheS8yMDIzLjcuMCIsCmxvYWRPcmRlciA9IDgwLApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXhhcnJheS8yMDIzLjcuMCIsCndWID0gIjAwMDAwMjAyMy4wMDAwMDAwMDcuKnpm', 'proj_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez', 'FSYNC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file', '_ModuleTable005_': 'cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJjLWJsb3NjLzEuMjEuNSIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMjEuMDAwMDAwMDA1Lip6ZmluYWwiLAp9LApjZG8gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9jZG8vMi4yLjAubHVhIiwKZnVsbE5hbWUgPSAiY2RvLzIuMi4wIiwKbG9hZE9yZGVyID0gMjYsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY2RvLzIu', 'g2tmpl_ver': '1.13.0', 'FHOUT_GOES': 3, '_ModuleTable021_': 'MC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nc2wvMi43LjEubHVhIiwKZnVsbE5hbWUgPSAiZ3NsLzIuNy4xIiwKbG9hZE9yZGVyID0gMzYsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ3NsLzIuNy4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sCmhkZjUgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxl', 'EUPD_CYC': 'gdas', 'fms_ver': '2023.02.01', 'util_linux_uuid_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj', '_ModuleTable052_': 'Y2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbnVtZXhwci8yLjguNC5sdWEiLApmdWxsTmFtZSA9ICJweS1udW1leHByLzIuOC40IiwKbG9hZE9yZGVyID0gNjQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktbnVtZXhwci8yLjguNCIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDguMDAwMDAwMDA0Lip6ZmluYWwiLAp9LApbInB5LW51bXB5Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2kt', '_ModuleTable030_': 'cFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJsaWJ5YW1sLzAuMi41IiwKd1YgPSAiMDAwMDAwMDAwLjAwMDAwMDAwMi4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCm1ldCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL21ldC85LjEuMy5sdWEiLApmdWxsTmFtZSA9ICJtZXQvOS4xLjMiLApsb2FkT3JkZXIgPSA3NywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0', 'COM_ATMOS_IMAGERY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery', 'I_MPI_EXTRA_FILESYSTEM_LIST': 'lustre', 'COM_WAVE_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history', 'SLURM_CLUSTER_NAME': 'hercules', '__LMOD_REF_COUNT_ACLOCAL_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/aclocal:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/aclocal:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/aclocal:2', 'SERIAL_CC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icc', 'COM_OBSPROC_TMPL': '${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos', 'SLURM_JOB_END_TIME': 1753757358, '_ModuleTable075_': 'ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3dncmliMi8yLjAuOC5sdWEiLApmdWxsTmFtZSA9ICJ3Z3JpYjIvMi4wLjgiLApsb2FkT3JkZXIgPSA1NSwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJ3Z3JpYjIvMi4wLjgiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKemxpYiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJj', 'FHMAX_GFS': 120, 'SENDAWIP': False, 'G2_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64/libg2_d.a', 'spack_mod_path': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core', 'OCNRES': 500, 'COM_MED_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart', 'SLURM_CPUS_ON_NODE': 1, 'LMOD_FAMILY_METAMPI': 'stack-intel-oneapi-mpi', 'LEVS': 128, 'FIXgsi': '/work2/noaa/global/mterry/global-workflow_forked/fix/gsi', 'NFHRS_PER_GROUP': 3, 'COM_WAVE_PREP_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep', 'FI_PROVIDER_PATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib/prov:/usr/lib64/libfabric', 'python_ver': '3.11.6', 'IAU_OFFSET': 0, 'py_packaging_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6', 'DIAGUTIL_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/sys_check/sys_check.sh', 'python_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2', 'SCRIPTScfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/scripts', 'py_markupsafe_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7', 'G2C_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include', 'py_xlrd_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f', 'LMOD_G2TMPLVIRT_VERSION': '1.13.0', 'pid': 358497, 'OUTPUT_GRID': 'gaussian_grid', 'jobid': 'oceanice_products_ice_f066.358191', 'SLURM_JOB_CPUS_PER_NODE': 1, 'spack_env': 'gsi-addon-dev-fms-2024.01', 'INTEL_ONEAPI_MPI_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse', 'LMOD_FAMILY_METACOMPILER': 'stack-intel', 'SP_INCd': 'include_d', 'py_six_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o', 'COM_ATMOS_GOES_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim', 'MPICXX': 'mpiicpc', 'restart_interval_gdas': 3, 'fit2obs_ver': '1.1.7.1', 'LMOD_DIR': '/apps/other/lmod/lmod/libexec', '_ModuleTable068_': 'LjAvc3AvMi41LjAubHVhIiwKZnVsbE5hbWUgPSAic3AvMi41LjAiLApsb2FkT3JkZXIgPSA0MiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMiwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzcC8yLjUuMCIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCnNxbGl0ZSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9zcWxpdGUvMy40My4yLmx1YSIsCmZ1bGxOYW1lID0gInNxbGl0ZS8zLjQzLjIiLApsb2FkT3Jk', 'REPLAY_ICS': False, 'COM_ICE_NETCDF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf', 'G2_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64/libg2_4.a', '_ModuleTable058_': 'InB5LXB5dHovMjAyMy4zIiwKbG9hZE9yZGVyID0gNjcsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcHl0ei8yMDIzLjMiLAp3ViA9ICIwMDAwMDIwMjMuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LXB5eGxzYiJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXB5eGxzYi8xLjAuMTAubHVhIiwKZnVsbE5hbWUgPSAicHktcHl4bHNiLzEuMC4xMCIsCmxvYWRPcmRlciA9', 'py_setuptools_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr', 'FHMIN': 0, 'COM_ICE_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input', '_ModuleTable048_': 'cHktZjkwbm1sIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktZjkwbm1sLzEuNC4zLmx1YSIsCmZ1bGxOYW1lID0gInB5LWY5MG5tbC8xLjQuMyIsCmxvYWRPcmRlciA9IDU2LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWY5MG5tbC8xLjQuMyIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMDQuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LWppbmphMiJdID0gewpmbiA9ICIvd29yay9ub2FhL2Vw', 'intel_mkl_ver': '2023.1.0', 'PRTE_MCA_plm_slurm_args': '--external-launcher', 'PWD': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ice_f066.358191', 'SLURM_GTIDS': 0, 'LOGNAME': 'mterry', 'MAKE_ACFTBUFR': False, 'W3EMC_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_8', 'HOMEDIR': '/work2/noaa/global/mterry', 'IAU_DELTHRS': 6, 'W3EMC_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_4', 'XDG_SESSION_TYPE': 'unspecified', 'ESMFMKFILE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib/esmf.mk', 'FHMAX': 9, 'WRITE_NSFLIP': True, 'SLURM_JOB_PARTITION': 'hercules', 'YAML_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x', 'PARTITION_SERVICE': 'service', 'MODULESHOME': '/apps/other/lmod/lmod', '__LMOD_REF_COUNT_DYLD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:1', 'COM_WAVE_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart', 'BUFR_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_4.so', 'BUFR_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_8.so', 'HOMEcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1', 'COM_OCEAN_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon', 'DEBUG_POSTSCRIPT': False, 'MANPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/share/man:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/man:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/documentation/en/man/common:/apps/other/lmod/lmod/share/man:/usr/share/man:/apps/share/man:/apps/man:/opt/slurm/share/man::', 'libyaml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x', 'jasper_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt', 'py_xlsxwriter_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok', 'SLURM_TRES_PER_TASK': 'cpu=1', 'nco_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq', 'stack_intel_ver': '2021.9.0', 'SLURM_OOM_KILL_STEP': 0, 'BUFR_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m', 'bufr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m', 'CLUSTERS_DTN': '', 'grib_util_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35', 'nghttp2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky', 'DO_TEST_MODE': True, 'PACKAGEROOT': '/work2/noaa/global/role-global/nwpara', 'lobsdiag_forenkf': True, 'SLURM_JOB_NUM_NODES': 1, 'py_netcdf4_ver': '1.5.8', 'ICERES': 500, 'CXX': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icpc', 'GRIB2GRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grib2grib2', 'DMPDIR': '/work/noaa/rstprod/dump', 'LSOIL_INCR': 2, 'metplus_ver': '3.1.1', '_ModuleTable008_': 'LjAvY3J0bS1maXgvMi40LjAuMV9lbWMubHVhIiwKZnVsbE5hbWUgPSAiY3J0bS1maXgvMi40LjAuMV9lbWMiLApsb2FkT3JkZXIgPSA0NywKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJjcnRtLWZpeC8yLjQuMC4xX2VtYyIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDQuMDAwMDAwMDAwLjAwMDAwMDAwMS4qXy4qZW1jLip6ZmluYWwiLAp9LApjdXJsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIw', 'COM_OCEAN_LETKF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf', 'zlib_ver': '1.2.13', 'BUFR_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_d.so', 'ENKF_SPREAD': True, 'zlib_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j', '__LMOD_REF_COUNT_CMAKE_PREFIX_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r:2;/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/IntelDPCPP:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72:1', '_ModuleTable035_': 'c3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9uZXRjZGYtY3h4NC80LjMuMS5sdWEiLApmdWxsTmFtZSA9ICJuZXRjZGYtY3h4NC80LjMuMSIsCmxvYWRPcmRlciA9IDc2LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm5ldGNkZi1jeHg0LzQuMy4xIiwKd1YgPSAiMDAwMDAwMDA0LjAwMDAwMDAwMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsibmV0Y2RmLWZvcnRyYW4iXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFj', 'FIXgfs': '/work2/noaa/global/mterry/global-workflow_forked/fix', 'DO_COUPLED': True, 'SLURM_JOBID': 5951733, 'SERIAL_FC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'COM_ATMOS_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input', 'W3EMC_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_d', '_ModuleTable013_': 'IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nMi8zLjQuNS5sdWEiLApmdWxsTmFtZSA9ICJnMi8zLjQuNSIsCmxvYWRPcmRlciA9IDQxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImcyLzMuNC41IiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAwNC4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCmcyYyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJj', 'COM_ATMOS_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history', 'FIXam': '/work2/noaa/global/mterry/global-workflow_forked/fix/am', 'hdf5_ver': '1.14.0', 'restart_interval_enkfgdas': 3, 'COM_ATMOS_RADMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon', 'DO_CALC_INCREMENT': False, 'DO_STARTMEM_FROM_JEDIICE': False, 'FIXprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/fix', 'SLURM_JOB_QOS': 'batch', 'I_MPI_HYDRA_BOOTSTRAP_EXEC_EXTRA_ARGS': '--external-launcher', 'obsproc_run_ver': '1.2.0', 'COM_ATMOS_MINMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon', 'crtm_ver': '2.4.0.1', 'FIXorog': '/work2/noaa/global/mterry/global-workflow_forked/fix/orog', 'COM_OCEAN_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history', '__LMOD_REF_COUNT_PATH': '/apps/other/globus-cli-3.35.2/bin:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r/ush:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/bin:2;/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/bin:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/bin/intel64:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/bin:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/bin:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/bin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/bin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin:1;/usr/sbin:1;/usr/bin:1;/apps/sbin:1;/apps/bin:1;/opt/slurm/bin:1;/home/gfekete/sven/bin:1', '_ModuleTable018_': 'KnpmaW5hbCIsCn0sCmdtYWtlID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2dtYWtlLzQuMi4xLmx1YSIsCmZ1bGxOYW1lID0gImdtYWtlLzQuMi4xIiwKbG9hZE9yZGVyID0gNTQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ21ha2UvNC4yLjEiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDAyLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJncmliLXV0aWwiXSA9IHsKZm4gPSAiL3dv', 'COMIN_ICE_HISTORY': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/ice/history', 'COMROOT': '/work2/noaa/global/mterry/RUNTESTS/COMROOT', 'HOME': '/home/mterry', 'FI_PROVIDER': 'mlx', 'met_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b', 'COM_RTOFS_TMPL': '${DMPDIR}', '_ModuleTable015_': 'bnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2cydG1wbC8xLjEzLjAubHVhIiwKZnVsbE5hbWUgPSAiZzJ0bXBsLzEuMTMuMCIsCmxvYWRPcmRlciA9IDQ1LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImcydG1wbC8xLjEzLjAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDEzLip6ZmluYWwiLAp9LApnZXR0ZXh0ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2dldHRleHQvMC4yMS4xLmx1YSIsCmZ1', '_ModuleTable_Sz_': 81, 'LANG': 'C.UTF-8', '__LMOD_REF_COUNT_LIBRARY_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:1', 'DO_BUFRSND': False, 'memory': '96GB', '_ModuleTable064_': 'PSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHkteGxzeHdyaXRlci8zLjEuNy5sdWEiLApmdWxsTmFtZSA9ICJweS14bHN4d3JpdGVyLzMuMS43IiwKbG9hZE9yZGVyID0gNzAsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHkteGxzeHdyaXRlci8zLjEuNyIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDA3Lip6ZmluYWwiLAp9LApbInB5LXhsd3QiXSA9IHsKZm4gPSAi', 'DYLD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib', 'PARTITION_BATCH': 'hercules', '_ModuleTable023_': 'cy8yMDIzLjEuMCIsCmxvYWRPcmRlciA9IDIsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiaW50ZWwtb25lYXBpLWNvbXBpbGVycy8yMDIzLjEuMCIsCndWID0gIjAwMDAwMjAyMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsiaW50ZWwtb25lYXBpLW1rbCJdID0gewpmbiA9ICIvYXBwcy9zcGFjay1tYW5hZ2VkL21vZHVsZWZpbGVzL2xpbnV4LXJvY2t5OS14ODZfNjQvQ29yZS9pbnRlbC1vbmVhcGktbWtsLzIwMjMuMS4wLmx1YSIsCmZ1bGxOYW1lID0gImludGVsLW9uZWFwaS1ta2wvMjAyMy4xLjAiLApsb2FkT3JkZXIgPSA2LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVz', '_ModuleTable081_': 'Ii9hcHBzL2NvbnRyaWIvbW9kdWxlZmlsZXMiLAp9LApzeXN0ZW1CYXNlTVBBVEggPSAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L0NvcmU6L2FwcHMvb3RoZXIvbW9kdWxlZmlsZXM6L2FwcHMvY29udGFpbmVycy9tb2R1bGVmaWxlczovYXBwcy9saWNlbnNlZC9tb2R1bGVmaWxlcyIsCn0K', 'REDERR': '2>', 'PSLOT': 'C48_S2SW', '__LMOD_REF_COUNT_PKG_CONFIG_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib/pkgconfig:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/pkgconfig:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib/pkgconfig:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/pkgconfig:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib/pkgconfig:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/lib/pkgconfig:1', '_ModuleTable036_': 'ay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL25ldGNkZi1mb3J0cmFuLzQuNi4xLmx1YSIsCmZ1bGxOYW1lID0gIm5ldGNkZi1mb3J0cmFuLzQuNi4xIiwKbG9hZE9yZGVyID0gMzEsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAibmV0Y2RmLWZvcnRyYW4vNC42LjEiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDA2LjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKbmdodHRwMiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFj', 'COM_ATMOS_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}', 'USHcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/ush', 'DO_GEMPAK': False, 'SLURM_PROCID': 0, 'USHgfs': '/work2/noaa/global/mterry/global-workflow_forked/ush', 'DOIBP_WAV': False, 'prepobs_run_ver': '1.1.0', 'EXPDIR': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', 'job': 'oceanice_products', 'imp_physics': 8, 'LMOD_SHELL_PRGM': 'bash', 'IAU_FHROT': 0, 'MPIF90': 'mpiifort', 'IP_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_4.a', 'waveGRD': 'uglo_100km', 'COM_CHEM_BMAT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix', 'jasper_ver': '2.0.32', 'SCRATCH': '/scratch/hercules', 'IP_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_8.a', 'CLIENT_GLOBUS_UUID': '869912fe-f6de-46c0-af10-b22efd84a022', 'ACCOUNT': 'fv3-cpu', 'HOMEgfs': '/work2/noaa/global/mterry/global-workflow_forked', 'NHOUR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/nhour', 'binary_diag': False, 'COM_WAVE_GRID_RES_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}', 'SCRIPTSprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/scripts', 'MP_STDOUTMODE': 'ORDERED', 'sigio_ver': '2.3.2', '__LMOD_REF_COUNT_CPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/include:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/include:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/include:1', 'LMOD_SETTARG_FULL_SUPPORT': False, 'COMINsyn': '/work2/noaa/global/role-global/com/gfs/prod/syndat', 'OFFSET_START_HOUR': 0, 'REDOUT': '1>', 'PTMP': '/work2/noaa/stmp/mterry/HERCULES', '_ModuleTable076_': 'dWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC96bGliLzEuMi4xMy5sdWEiLApmdWxsTmFtZSA9ICJ6bGliLzEuMi4xMyIsCmxvYWRPcmRlciA9IDcsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEwLApzdGFja0RlcHRoID0gNSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInpsaWIvMS4yLjEzIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMi4wMDAwMDAwMTMuKnpmaW5hbCIsCn0sCnpzdGQgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2', 'g2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato', 'MODE': 'forecast-only', 'DATAROOT': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312', 'bacio_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj', 'APP': 'S2SW', '_ModuleTable029_': 'LzQuNC4zNSIsCmxvYWRPcmRlciA9IDEyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImxpYnhjcnlwdC80LjQuMzUiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDA0LjAwMDAwMDAzNS4qemZpbmFsIiwKfSwKbGlieWFtbCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9saWJ5YW1sLzAuMi41Lmx1YSIsCmZ1bGxOYW1lID0gImxpYnlhbWwvMC4yLjUiLApsb2FkT3JkZXIgPSA1OSwKcHJv', 'openblas_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f', 'stack_impi_ver': '2021.9.0', 'COM_CHEM_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem', 'TMPDIR': '/local/scratch/mterry/5951733', 'HDF5_PLUGIN_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/plugins', 'DO_TRACKER': True, 'CMAKE_PREFIX_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r:/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/IntelDPCPP:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72', 'crtm_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r', '_ModuleTable038_': 'YWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvb3BlbmJsYXMvMC4zLjI0Lmx1YSIsCmZ1bGxOYW1lID0gIm9wZW5ibGFzLzAuMy4yNCIsCmxvYWRPcmRlciA9IDUwLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm9wZW5ibGFzLzAuMy4yNCIsCndWID0gIjAwMDAwMDAwMC4wMDAwMDAwMDMuMDAwMDAwMDI0Lip6ZmluYWwiLAp9LApvcGVuanBlZyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5z', 'nco_ver': '5.0.6', '_ModuleTable046_': 'IiwKfSwKWyJweS1jZnRpbWUiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1jZnRpbWUvMS4wLjMuNC5sdWEiLApmdWxsTmFtZSA9ICJweS1jZnRpbWUvMS4wLjMuNCIsCmxvYWRPcmRlciA9IDU3LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWNmdGltZS8xLjAuMy40IiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMC4wMDAwMDAwMDMuMDAwMDAwMDA0Lip6ZmluYWwiLAp9', 'GRBINDEX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grbindex', 'libjpeg_turbo_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy', 'MPIF77': 'mpiifort', '_ModuleTable073_': 'LjAwMDAwMDAwMi4wMDAwMDAwMjguKnpmaW5hbCIsCn0sClsidXRpbC1saW51eC11dWlkIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvdXRpbC1saW51eC11dWlkLzIuMzguMS5sdWEiLApmdWxsTmFtZSA9ICJ1dGlsLWxpbnV4LXV1aWQvMi4zOC4xIiwKbG9hZE9yZGVyID0gMTQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAidXRpbC1saW51eC11dWlkLzIuMzguMSIsCndWID0gIjAw', '_ModuleTable010_': 'Mi4wIiwKbG9hZE9yZGVyID0gMjAsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZWNjb2Rlcy8yLjMyLjAiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDMyLip6ZmluYWwiLAp9LAplc21mID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAvZXNtZi84LjYuMC5sdWEiLApmdWxsTmFtZSA9ICJlc21mLzguNi4wIiwKbG9hZE9yZGVyID0gMzQsCnBy', 'NTHSTACK': 1024000000, 'SLURM_CPUS_PER_TASK': 1, 'FIXcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/fix', 'SLURM_NTASKS': 1, 'DO_ICE': True, 'NET': 'gfs', 'COM_ATMOS_GENESIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital', 'HOMEpost': '/work2/noaa/global/mterry/global-workflow_forked', '_ModuleTable061_': 'CnVzZXJOYW1lID0gInB5LXNldHVwdG9vbHMvNjMuNC4zIiwKd1YgPSAiMDAwMDAwMDYzLjAwMDAwMDAwNC4wMDAwMDAwMDMuKnpmaW5hbCIsCn0sClsicHktc2l4Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktc2l4LzEuMTYuMC5sdWEiLApmdWxsTmFtZSA9ICJweS1zaXgvMS4xNi4wIiwKbG9hZE9yZGVyID0gNzMsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktc2l4LzEu', 'bacio_ver': '2.4.1', 'cmake_ver': '3.23.1', 'UTILROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq', '_ModuleTable024_': 'ZXJOYW1lID0gImludGVsLW9uZWFwaS1ta2wvMjAyMy4xLjAiLAp3ViA9ICIwMDAwMDIwMjMuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbImludGVsLW9uZWFwaS1tcGkiXSA9IHsKZm4gPSAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L29uZWFwaS8yMDIzLjEuMC9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLmx1YSIsCmZ1bGxOYW1lID0gImludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAiLApsb2FkT3JkZXIgPSA0LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAiLAp3ViA9ICIwMDAwMDIwMjEuMDAwMDAwMDA5Lip6ZmluYWwi', 'tar_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths', 'ntasks': 1, 'max_tasks_per_node': 80, 'NCDUMP': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump', '_ModuleTable057_': 'LjIubHVhIiwKZnVsbE5hbWUgPSAicHktcHl0aG9uLWRhdGV1dGlsLzIuOC4yIiwKbG9hZE9yZGVyID0gNzQsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcHl0aG9uLWRhdGV1dGlsLzIuOC4yIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwOC4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sClsicHktcHl0eiJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXB5dHovMjAyMy4zLmx1YSIsCmZ1bGxOYW1lID0g', 'py_numpy_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne', 'QUILTING': True, 'IP_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_d.a', 'SLURM_TOPOLOGY_ADDR': 'hercules-08-38', 'libxcrypt_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr', 'CRTM_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include', 'LMOD_VERSION': '8.7.14', 'spack_stack_ver': '1.6.0', '_ModuleTable041_': 'LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9wYXJhbGxlbGlvLzIuNS4xMC5sdWEiLApmdWxsTmFtZSA9ICJwYXJhbGxlbGlvLzIuNS4xMCIsCmxvYWRPcmRlciA9IDMzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInBhcmFsbGVsaW8vMi41LjEwIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNS4wMDAwMDAwMTAuKnpmaW5hbCIsCn0sCnBpZ3ogPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0y', 'HOMEprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0', 'COMPONENT': 'ice', 'BASE_GIT': '/work2/noaa/global/role-global/git', 'FCST_SEGMENTS': [0, 120], 'libpng_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu', '_ModuleTable044_': 'LjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKcHJvaiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9wcm9qLzkuMi4xLmx1YSIsCmZ1bGxOYW1lID0gInByb2ovOS4yLjEiLApsb2FkT3JkZXIgPSAyNCwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcm9qLzkuMi4xIiwKd1YgPSAiMDAwMDAwMDA5LjAwMDAwMDAwMi4wMDAwMDAwMDEuKnpm', 'DO_GENESIS': True, 'BUFR_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_4', '__LMOD_REF_COUNT_DIAGUTIL_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/sys_check/sys_check.sh:1', 'COM_WAVE_GEMPAK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak', 'COM_ATMOS_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2', 'RUN': 'gfs', 'BUFR_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_8', 'COM_ICE_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon', 'py_pandas_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw', '__INTEL_POST_CFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64'], 'ARCDIR': '/work2/noaa/global/mterry/archive/C48_S2SW', '__LMOD_REF_COUNT_NLSPATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64/locale/%l_%t/%N:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin/locale/%l_%t/%N:1', '_ModuleTable069_': 'ZXIgPSAxMywKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMiwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzcWxpdGUvMy40My4yIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDA0My4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sClsic3RhY2staW50ZWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9Db3JlL3N0YWNrLWludGVsLzIwMjEuOS4wLmx1YSIsCmZ1bGxOYW1lID0gInN0YWNrLWludGVsLzIwMjEuOS4wIiwKbG9hZE9yZGVyID0gMywKcHJvcFQgPSB7fSwK', 'SENDECF': False, 'parallelio_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed', 'py_pytz_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s', 'ILPOST': 3, 'WGRIB2_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib/libwgrib2.a', 'parallel_netcdf_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3', 'CMPLR_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0', 'prod_util_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq', 'COMINukmet': '/work2/noaa/global/role-global/data/external_gempak/ukmet', 'libpng_ver': '1.6.37', 'COPYGB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/copygb2', 'NMV': '/bin/mv', 'NOSCRUB': '/work2/noaa/global/mterry', 'SCRATCH_hercules': '/scratch/hercules', 'W3EMC_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_d.a', 'SENDSDM': False, 'envir': 'prod', 'DO_METP': False, 'CRTM_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib/libcrtm.a', 'LOGSCRIPT': '', 'INTEL_ONEAPI_COMPILERS_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72', '_ModuleTable074_': 'MDAwMDAwMi4wMDAwMDAwMzguMDAwMDAwMDAxLip6ZmluYWwiLAp9LAp3M2VtYyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC93M2VtYy8yLjEwLjAubHVhIiwKZnVsbE5hbWUgPSAidzNlbWMvMi4xMC4wIiwKbG9hZE9yZGVyID0gMzksCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAidzNlbWMvMi4xMC4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKd2dyaWIy', 'py_jinja2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj', 'FPGA_VARS_ARGS': '', 'COMOUT_ICE_NETCDF': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ice/netcdf', 'INCVARS_ZERO_STRAT': ["'sphum_inc'", "'liq_wat_inc'", "'icmr_inc'", "'rwmr_inc'", "'snmr_inc'", "'grle_inc'"], 'COM_CONF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf', '_ModuleTable003_': 'YSIsCmZ1bGxOYW1lID0gImJhY2lvLzIuNC4xIiwKbG9hZE9yZGVyID0gMzgsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAzLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiYmFjaW8vMi40LjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDA0LjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKYnVmciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9idWZyLzExLjcuMC5sdWEiLApmdWxsTmFtZSA9ICJidWZyLzExLjcuMCIsCmxvYWRPcmRlciA9', 'HYDRA_BOOTSTRAP': 'slurm', '__LMOD_REF_COUNT___INTEL_POST_FFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1'], 'COM_CHEM_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon', 'MDATE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/mdate', 'WGRIB2': 'wgrib2', '_ModuleTable054_': 'LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktb3BlbnB5eGwvMy4xLjIubHVhIiwKZnVsbE5hbWUgPSAicHktb3BlbnB5eGwvMy4xLjIiLApsb2FkT3JkZXIgPSA2NiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1vcGVucHl4bC8zLjEuMiIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbInB5LXBhY2thZ2luZyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxs', 'EXP_WARM_START': False, 'MODULEPATH_ROOT': '/apps/other/modulefiles', 'COM_SNOW_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon', 'CHGRP_RSTPROD': True, 'py_pyxlsb_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge', 'BACIO_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/include_8', 'restart_interval_gfs': 12, 'DO_GOES': False, 'pgm': '', 'SLURM_TOPOLOGY_ADDR_PATTERN': 'node', 'DO_LAND_IAU': False, 'ntiles': 6, 'FHCYC': 24, 'ROTDIR': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW', 'FHOUT_AERO': 3, 'py_jinja2_ver': '3.1.2', 'BACIO_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/include_4', 'INCREMENTS_TO_ZERO': ["'liq_wat_inc'", "'icmr_inc'", "'rwmr_inc'", "'snmr_inc'", "'grle_inc'"], 'FHR_LIST': 66, 'USHprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/ush', 'BUFR_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_d', 'hdf5_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh', 'INTEL_ONEAPI_MKL_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci', 'SCRIPTSfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/scripts', '__LMOD_REF_COUNT_FI_PROVIDER_PATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib/prov:1;/usr/lib64/libfabric:1', 'FIXcice': '/work2/noaa/global/mterry/global-workflow_forked/fix/cice', 'IAUFHRS': [6], 'XDG_SESSION_CLASS': 'background', 'INTERVAL_GFS': 6, 'py_openpyxl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe', 'FHMAX_FITS': 120, 'COM_ATMOS_MASTER_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master', 'LMOD_PKG': '/apps/other/lmod/lmod', 'MPI_MEMMAP_OFF': 1, 'UUID_HERCULES_DTN': '869912fe-f6de-46c0-af10-b22efd84a022', '_ModuleTable079_': 'L3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAiCiwgIi9hcHBzL3NwYWNrLW1hbmFnZWQvbW9kdWxlZmlsZXMvbGludXgtcm9ja3k5LXg4Nl82NC9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLWE2NmVhaXAvb25lYXBpLzIwMjMuMS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wIgosICIvYXBwcy9zcGFjay1tYW5hZ2VkL21vZHVsZWZp', 'SDATE': datetime.datetime(2021, 3, 23, 12, 0), 'SLURM_SCRIPT_CONTEXT': 'prolog_task', 'MPI_CXX': 'mpiicpc', 'lwrite4danl': True, 'SLURM_MEM_PER_NODE': 98304, 'CASE_ENS': '{{ CASE_ENS }}', '_ModuleTable049_': 'aWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LWppbmphMi8zLjEuMi5sdWEiLApmdWxsTmFtZSA9ICJweS1qaW5qYTIvMy4xLjIiLApsb2FkT3JkZXIgPSA2MiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1qaW5qYTIvMy4xLjIiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDAxLjAwMDAwMDAwMi4qemZpbmFsIiwKfSwKWyJweS1tYXJrdXBzYWZlIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3Vs', 'YAML_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include', 'PYTHONPATH': '/apps/other/globus-cli-3.35.2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:/work2/noaa/global/mterry/global-workflow_forked/sorc/wxflow/src:/work2/noaa/global/mterry/global-workflow_forked/ush/python', '__LMOD_REF_COUNT_PYTHONPATH': '/apps/other/globus-cli-3.35.2:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:3;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:1', 'py_xarray_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4', 'nst_anl': True, 'FHOUT_GFS': 3, 'WORK': '/work/hercules', 'W3EMC_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_4.a', 'F77': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'BASE_DATA': '/work2/noaa/global/role-global/data', 'py_xlwt_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56', 'ACLOCAL_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/aclocal:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/aclocal:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/aclocal', 'QUEUE_SERVICE': 'batch', 'W3EMC_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_8.a', 'DATA': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ice_f066.358191', 'LESSOPEN': '||/usr/bin/lesspipe.sh %s', 'DO_VERFOZN': True, 'antlr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2', 'crtm_fix_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2', 'TOCGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib', 'USER': 'mterry', 'HPSS_PROJECT': 'emc-global', 'FHMAX_WAV_GFS': 120, 'FIXgdas': '/work2/noaa/global/mterry/global-workflow_forked/fix/gdas', 'launcher': 'srun -l --export=ALL --hint=nomultithread', 'NDATE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/ndate', 'LIBRARY_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib', 'gsl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z', 'SLURM_NODELIST': 'hercules-08-38', 'DO_FIT2OBS': True, 'IP_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_4', 'DOIAU_ENKF': True, 'COM_ATMOS_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos', 'py_xarray_ver': '2023.7.0', '__INTEL_POST_FFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64'], 'netcdf_diag': True, 'DOBNDPNT_WAVE': True, 'FETCHDIR': '/NCEPDEV/emc-global/1year/David.Grumm/test_data', 'ENVIRONMENT': 'BATCH', 'gsi_ncdiag_ver': '1.1.2', 'IP_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_8', 'py_bottleneck_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a', 'w3emc_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo', 'py_netcdf4_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i', 'wgrib2_ver': '2.0.8', 'TOCGRIB2SUPER': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib2super', 'COM_ICE_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2', 'KEEPDATA': False, 'udunits_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo', 'cdo_ver': '2.2.0', 'COM_OCEAN_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input', 'FNTSFA': ' ', 'LOADEDMODULES': 'contrib/0.1:intel-oneapi-compilers/2023.1.0:stack-intel/2021.9.0:intel-oneapi-mpi/2021.9.0:stack-intel-oneapi-mpi/2021.9.0:intel-oneapi-mkl/2023.1.0:zlib/1.2.13:pigz/2.7:zstd/1.5.2:tar/1.34:gettext/0.21.1:libxcrypt/4.4.35:sqlite/3.43.2:util-linux-uuid/2.38.1:python/3.11.6:libjpeg/2.1.0:jasper/2.0.32:libpng/1.6.37:openjpeg/2.3.1:eccodes/2.32.0:fftw/3.3.10:nghttp2/1.57.0:curl/8.4.0:proj/9.2.1:udunits/2.2.28:cdo/2.2.0:hdf5/1.14.0:snappy/1.1.10:c-blosc/1.21.5:netcdf-c/4.9.2:netcdf-fortran/4.6.1:parallel-netcdf/1.12.2:parallelio/2.5.10:esmf/8.6.0:antlr/2.7.7:gsl/2.7.1:nco/5.0.6:bacio/2.4.1:w3emc/2.10.0:prod_util/2.1.1:g2/3.4.5:sp/2.5.0:ip/4.3.0:grib-util/1.3.0:g2tmpl/1.13.0:gsi-ncdiag/1.1.2:crtm-fix/2.4.0.1_emc:git-lfs/3.1.2:crtm/2.4.0.1:openblas/0.3.24:py-setuptools/63.4.3:py-numpy/1.23.4:bufr/11.7.0:gmake/4.2.1:wgrib2/2.0.8:py-f90nml/1.4.3:py-cftime/1.0.3.4:py-netcdf4/1.5.8:libyaml/0.2.5:py-pyyaml/6.0:py-markupsafe/2.1.3:py-jinja2/3.1.2:py-bottleneck/1.3.7:py-numexpr/2.8.4:py-et-xmlfile/1.0.1:py-openpyxl/3.1.2:py-pytz/2023.3:py-pyxlsb/1.0.10:py-xlrd/2.0.1:py-xlsxwriter/3.1.7:py-xlwt/1.3.0:py-pandas/1.5.3:py-six/1.16.0:py-python-dateutil/2.8.2:g2c/1.8.0:netcdf-cxx4/4.3.1:met/9.1.3:metplus/3.1.1:py-packaging/23.1:py-xarray/2023.7.0:prepobs/1.1.0:fit2obs/1.1.7.1:globus-cli/3.35.2:module_base.hercules', 'SLURM_JOB_ACCOUNT': 'fv3-cpu', 'DO_OCN': True, 'SLURM_PRIO_PROCESS': 0, 'HOMEfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1', 'gmake_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq', 'FIXfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/fix', 'py_python_dateutil_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy', 'tasks_per_node': 1, 'SLURM_NPROCS': 1, 'COM_CHEM_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history', 'LMOD_ROOT': '/apps/other/lmod', 'DOHYBVAR': '{{ DOHYBVAR }}', 'GSL_ROOT_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z', 'SERIAL_F77': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'IP_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_d', 'SHLVL': 4, '_ModuleTable071_': 'e30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAic3RhY2staW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMCIsCndWID0gIjAwMDAwMjAyMS4wMDAwMDAwMDkuKnpmaW5hbCIsCn0sCnRhciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC90YXIvMS4zNC5sdWEiLApmdWxsTmFtZSA9ICJ0YXIvMS4zNCIsCmxvYWRPcmRlciA9IDEwLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIs', '_ModuleTable012_': 'c3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImZmdHcvMy4zLjEwIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAwMy4wMDAwMDAwMTAuKnpmaW5hbCIsCn0sCmZpdDJvYnMgPSB7CmZuID0gIi93b3JrMi9ub2FhL2dsb2JhbC9yb2xlLWdsb2JhbC9naXQvRml0Mk9icy92MS4xLjcuMS9tb2R1bGVmaWxlcy9maXQyb2JzLzEuMS43LjEubHVhIiwKZnVsbE5hbWUgPSAiZml0Mm9icy8xLjEuNy4xIiwKbG9hZE9yZGVyID0gODIsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZml0Mm9icy8xLjEuNy4xIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4wMDAwMDAwMDcuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApnMiA9', 'INCVARS_EFOLD': 5, 'SLURM_NNODES': 1, 'RESERVATION': '', 'FHMIN_GFS': 0, 'FHMAX_HF_WAV': 48, 'FIXugwd': '/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd', 'BASH_ENV': '/apps/other/lmod/lmod/init/bash', 'CDATE': datetime.datetime(2021, 3, 23, 12, 0), '_ModuleTable063_': 'aW5hbCIsCn0sClsicHkteGxyZCJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXhscmQvMi4wLjEubHVhIiwKZnVsbE5hbWUgPSAicHkteGxyZC8yLjAuMSIsCmxvYWRPcmRlciA9IDY5LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXhscmQvMi4wLjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJweS14bHN4d3JpdGVyIl0g', '_ModuleTable006_': 'Mi4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMi4qemZpbmFsIiwKfSwKY29udHJpYiA9IHsKZm4gPSAiL2FwcHMvb3RoZXIvbW9kdWxlZmlsZXMvY29udHJpYi8wLjEiLApmdWxsTmFtZSA9ICJjb250cmliLzAuMSIsCmxvYWRPcmRlciA9IDEsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAwLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY29udHJpYiIsCndWID0gIjAwMDAwMDAwMC4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sCmNydG0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50', 'FHOUT_OCN_GFS': 6, '_ModuleTable026_': 'cmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2phc3Blci8yLjAuMzIubHVhIiwKZnVsbE5hbWUgPSAiamFzcGVyLzIuMC4zMiIsCmxvYWRPcmRlciA9IDE3LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImphc3Blci8yLjAuMzIiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAzMi4qemZpbmFsIiwKfSwKbGlianBlZyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1z', 'DONST': True, '_ModuleTable051_': 'ay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL3B5LW5ldGNkZjQvMS41LjgubHVhIiwKZnVsbE5hbWUgPSAicHktbmV0Y2RmNC8xLjUuOCIsCmxvYWRPcmRlciA9IDU4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LW5ldGNkZjQvMS41LjgiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA1LjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKWyJweS1udW1leHByIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3Rh', 'LMOD_sys': 'Linux', '_ModuleTable033_': 'L3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL25jby81LjAuNi5sdWEiLApmdWxsTmFtZSA9ICJuY28vNS4wLjYiLApsb2FkT3JkZXIgPSAzNywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJuY28vNS4wLjYiLAp3ViA9ICIwMDAwMDAwMDUuMDAwMDAwMDAwLjAwMDAwMDAwNi4qemZpbmFsIiwKfSwKWyJuZXRjZGYtYyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1', 'FHMIN_WAV': 0, 'LMOD_FAMILY_METACOMPILER_VERSION': '2021.9.0', 'I_MPI_PMI_LIBRARY': '/opt/slurm/lib/libpmi2.so', 'machine': 'HERCULES', 'py_pyyaml_ver': 6.0, 'OCL_ICD_FILENAMES': 'libintelocl_emu.so:libalteracl.so:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64/libintelocl.so', '_ModuleTable017_': 'bG9hZE9yZGVyID0gNDgsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ2l0LWxmcy8zLjEuMiIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbImdsb2J1cy1jbGkiXSA9IHsKZm4gPSAiL2FwcHMvb3RoZXIvbW9kdWxlZmlsZXMvZ2xvYnVzLWNsaS8zLjM1LjIubHVhIiwKZnVsbE5hbWUgPSAiZ2xvYnVzLWNsaS8zLjM1LjIiLApsb2FkT3JkZXIgPSA4MywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJnbG9idXMtY2xpIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAzNS4wMDAwMDAwMDIu', 'DO_FETCH_HPSS': False, 'PROJ_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/proj', 'FHOUT_ICE': 3, 'COM_WAVE_STATION_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station', 'COMINecmwf': '/work2/noaa/global/role-global/data/external_gempak/ecmwf', 'COM_TOP_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}', 'WGRIB2_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include', 'mpmd_opt': '--multi-prog --output=mpmd.%j.%t.out', 'DO_PREP_SFC': False, 'LMOD_G2TMPLVIRT_NAME': 'g2tmpl', '_ModuleTable066_': 'bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5dGhvbi8zLjExLjYubHVhIiwKZnVsbE5hbWUgPSAicHl0aG9uLzMuMTEuNiIsCmxvYWRPcmRlciA9IDE1LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5dGhvbi8zLjExLjYiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDExLjAwMDAwMDAwNi4qemZpbmFsIiwKfSwKc25hcHB5ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAy', '_ModuleTable009_': 'MjEuOS4wL2N1cmwvOC40LjAubHVhIiwKZnVsbE5hbWUgPSAiY3VybC84LjQuMCIsCmxvYWRPcmRlciA9IDIzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImN1cmwvOC40LjAiLAp3ViA9ICIwMDAwMDAwMDguMDAwMDAwMDA0Lip6ZmluYWwiLAp9LAplY2NvZGVzID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2VjY29kZXMvMi4zMi4wLmx1YSIsCmZ1bGxOYW1lID0gImVjY29kZXMvMi4z', 'py_numexpr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq', '__LMOD_REF_COUNT_MANPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/share/man:2;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/man:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/documentation/en/man/common:1;/apps/other/lmod/lmod/share/man:1;/usr/share/man:1;/apps/share/man:1;/apps/man:1;/opt/slurm/share/man:1', 'DOLETKF_OCN': False, 'MPI_GROUP_MAX': 256, 'py_cftime_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i', 'XDG_SESSION_ID': 'c22', 'zstd_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4', 'CLUSTERS': '', '_ModuleTable034_': 'bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAvbmV0Y2RmLWMvNC45LjIubHVhIiwKZnVsbE5hbWUgPSAibmV0Y2RmLWMvNC45LjIiLApsb2FkT3JkZXIgPSAzMCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJuZXRjZGYtYy80LjkuMiIsCndWID0gIjAwMDAwMDAwNC4wMDAwMDAwMDkuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbIm5ldGNkZi1jeHg0Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2st', 'COM_OCEAN_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean', 'CLUSTERS_SERVICE': '', 'pgmout': 'OUTPUT.358497', 'EXECfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/exec', '_ModuleTable031_': 'aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJtZXQvOS4xLjMiLAp3ViA9ICIwMDAwMDAwMDkuMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKbWV0cGx1cyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL21ldHBsdXMvMy4xLjEubHVhIiwKZnVsbE5hbWUgPSAibWV0cGx1cy8zLjEuMSIsCmxvYWRPcmRlciA9IDc4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIs', 'COMOUT_ICE_GRIB': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ice/grib2', 'COM_ATMOS_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon', 'HOMEobsproc': '/work2/noaa/global/role-global/git/obsproc/v1.2.0', 'esmf_ver': '8.6.0', 'DO_ARCHCOM': False, 'SLURM_SUBMIT_HOST': 'hercules-login-1.hpc.msstate.edu', 'WRITE_DOPOST': True, 'EXECcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/exec', 'DUMP_SUFFIX': '', '_ModuleTable072_': 'CnVzZXJOYW1lID0gInRhci8xLjM0IiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAzNC4qemZpbmFsIiwKfSwKdWR1bml0cyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC91ZHVuaXRzLzIuMi4yOC5sdWEiLApmdWxsTmFtZSA9ICJ1ZHVuaXRzLzIuMi4yOCIsCmxvYWRPcmRlciA9IDI1LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInVkdW5pdHMvMi4yLjI4IiwKd1YgPSAiMDAwMDAwMDAy', 'CASE': 'C48', 'SENDDBN': False, 'gettext_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43', 'COM_ICE_BMATRIX_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice', 'git_lfs_ROOT': '/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l', '_ModuleTable001_': 'X01vZHVsZVRhYmxlXyA9IHsKTVR2ZXJzaW9uID0gMywKY19yZWJ1aWxkVGltZSA9IGZhbHNlLApjX3Nob3J0VGltZSA9IGZhbHNlLApkZXB0aFQgPSB7fSwKZmFtaWx5ID0gewpNZXRhQ29tcGlsZXIgPSAic3RhY2staW50ZWwiLApNZXRhTVBJID0gInN0YWNrLWludGVsLW9uZWFwaS1tcGkiLApjb21waWxlciA9ICJpbnRlbC1vbmVhcGktY29tcGlsZXJzIiwKZzJ0bXBsdmlydCA9ICJnMnRtcGwiLAptcGkgPSAiaW50ZWwtb25lYXBpLW1waSIsCn0sCm1UID0gewphbnRsciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9t', 'COM_ICE_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart', '_ModuleTable025_': 'LAp9LAppcCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9pcC80LjMuMC5sdWEiLApmdWxsTmFtZSA9ICJpcC80LjMuMCIsCmxvYWRPcmRlciA9IDQzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImlwLzQuMy4wIiwKd1YgPSAiMDAwMDAwMDA0LjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKamFzcGVyID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hl', 'COM_OCEAN_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart', 'ARCHCOM_TO': 'globus_hpss', 'COM_ATMOS_WMO_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo', 'ip_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf', 'CLASSPATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/mpi.jar', 'INTELFPGAOCLSDKROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga', 'LD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/host/linux64/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/lib:/usr/lib64:/usr/lib:/opt/slurm/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib', 'DO_CA': True, 'g2_ver': '3.4.5', 'LMOD_FAMILY_COMPILER': 'intel-oneapi-compilers', 'XDG_RUNTIME_DIR': '/run/user/9583', 'BASE_ENV': '/work2/noaa/global/mterry/global-workflow_forked/env', 'SLURM_JOB_ID': 5951733, '_ModuleTable014_': 'dWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nMmMvMS44LjAubHVhIiwKZnVsbE5hbWUgPSAiZzJjLzEuOC4wIiwKbG9hZE9yZGVyID0gNzUsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZzJjLzEuOC4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKZzJ0bXBsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9p', 'OCEANICEPRODUCTS_CONFIG': '/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml', 'NTHREADS_OCNICEPOST': 1, 'DO_JEDIOCNVAR': False, 'FIXreg2grb2': '/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2', 'SLURM_NODEID': 0, 'ip_ver': '4.3.0', 'KMP_AFFINITY': 'scatter', 'FHOUT_ICE_GFS': 6, '_ModuleTable065_': 'L3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS14bHd0LzEuMy4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXhsd3QvMS4zLjAiLApsb2FkT3JkZXIgPSA3MSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS14bHd0LzEuMy4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKcHl0aG9uID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1', 'DO_ATM': True, '_ModuleTable004_': 'IDUzLApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImJ1ZnIvMTEuNy4wIiwKd1YgPSAiMDAwMDAwMDExLjAwMDAwMDAwNy4qemZpbmFsIiwKfSwKWyJjLWJsb3NjIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvYy1ibG9zYy8xLjIxLjUubHVhIiwKZnVsbE5hbWUgPSAiYy1ibG9zYy8xLjIxLjUiLApsb2FkT3JkZXIgPSAyOSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1', 'esmf_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep', 'LMOD_FAMILY_MPI_VERSION': '2021.9.0', '_ModuleTable020_': 'c3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9nc2ktbmNkaWFnLzEuMS4yLmx1YSIsCmZ1bGxOYW1lID0gImdzaS1uY2RpYWcvMS4xLjIiLApsb2FkT3JkZXIgPSA0NiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJnc2ktbmNkaWFnLzEuMS4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCmdzbCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYu', 'DO_JEDIATMVAR': False, 'DO_AERO_ANL': False, 'NCLEN': '/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen', 'PS4': '+ $(basename ${BASH_SOURCE[0]:-${FUNCNAME[0]:-"Unknown"}})[${LINENO}]', 'COM_SNOW_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow', '_ModuleTable053_': 'YWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbnVtcHkvMS4yMy40Lmx1YSIsCmZ1bGxOYW1lID0gInB5LW51bXB5LzEuMjMuNCIsCmxvYWRPcmRlciA9IDUyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSA4LApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LW51bXB5LzEuMjMuNCIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMjMuMDAwMDAwMDA0Lip6ZmluYWwiLAp9LApbInB5LW9wZW5weXhsIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0', '__LMOD_REF_COUNT_ACL_BOARD_VENDOR_PATH': '/opt/Intel/OpenCLFPGA/oneAPI/Boards:1', 'LMOD_FAMILY_METAMPI_VERSION': '2021.9.0', '_ModuleTable059_': 'IDY4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXB5eGxzYi8xLjAuMTAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAwLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKWyJweS1weXlhbWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1weXlhbWwvNi4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXB5eWFtbC82LjAiLApsb2FkT3JkZXIgPSA2MCwKcHJvcFQgPSB7fSwKc3RhY2tE', 'DO_CALC_INCREMENT_ENKF_GFS': False, 'MKLROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0', 'DEBUGINFOD_URLS': 'https://debuginfod.centos.org/ ', 'sven_root_path': '/home/gfekete/sven', '_ModuleTable047_': 'LApbInB5LWV0LXhtbGZpbGUiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1ldC14bWxmaWxlLzEuMC4xLmx1YSIsCmZ1bGxOYW1lID0gInB5LWV0LXhtbGZpbGUvMS4wLjEiLApsb2FkT3JkZXIgPSA2NSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1ldC14bWxmaWxlLzEuMC4xIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMC4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsi', 'restart_interval_enkfgfs': 3, 'SENDDBN_NTC': False, 'FIXaer': '/work2/noaa/global/mterry/global-workflow_forked/fix/aer', 'FHMAX_HF_GFS': 48, 'COM_ICE_LETKF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf', 'LMOD_FAMILY_MPI': 'intel-oneapi-mpi', 'netcdf_fortran_ver': '4.6.1', 'FIXmom': '/work2/noaa/global/mterry/global-workflow_forked/fix/mom6', 'cyc': 12, 'BACIO_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib/libbacio_4.a', 'FC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'FHOUT_WAV_GFS': 3, 'SDATE_GFS': datetime.datetime(2021, 3, 23, 12, 0), 'BACIO_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib/libbacio_8.a', 'sp_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr', 'OPS_RES': 'C768', 'NMEM_ENS_GFS': 30, 'which_declare': 'declare -f', '_ModuleTable056_': 'OS4wL3B5LXBhbmRhcy8xLjUuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1wYW5kYXMvMS41LjMiLApsb2FkT3JkZXIgPSA3MiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1wYW5kYXMvMS41LjMiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA1LjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKWyJweS1weXRob24tZGF0ZXV0aWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1weXRob24tZGF0ZXV0aWwvMi44', 'COM_WAVE_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded', 'LMOD_FAMILY_G2TMPLVIRT': 'g2tmpl', 'QUEUE_DTN': 'batch', 'DO_GENESIS_FSU': False, 'prod_util_ver': '2.1.1', 'MPI_F90': 'mpiifort', 'LMOD_FAMILY_COMPILER_VERSION': '2023.1.0', '_ModuleTable042_': 'MDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcGlnei8yLjcubHVhIiwKZnVsbE5hbWUgPSAicGlnei8yLjciLApsb2FkT3JkZXIgPSA4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gNCwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInBpZ3ovMi43IiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4qemZpbmFsIiwKfSwKcHJlcG9icyA9IHsKZm4gPSAiL3dvcmsyL25vYWEvZ2xvYmFsL3JvbGUtZ2xvYmFsL2dpdC9wcmVwb2JzL3YxLjEuMC9tb2R1bGVmaWxlcy9wcmVwb2JzLzEuMS4wLmx1YSIsCmZ1bGxOYW1lID0gInByZXBvYnMvMS4xLjAiLApsb2FkT3JkZXIgPSA4MSwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0', 'openjpeg_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu', 'USHfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/ush', 'YAML_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib', 'DO_GSISOILDA': False, 'nemsio_ver': '2.5.4', 'DO_NPOESS': False, 'SLURM_MPI_TYPE': 'pmi2', 'OMP_STACKSIZE': 2048000, 'py_et_xmlfile_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq', 'FORECAST_HOUR': 66, 'ATARDIR': '/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW', 'EXECgfs': '/work2/noaa/global/mterry/global-workflow_forked/exec', 'py_python_dateutil_ver': '2.8.2', 'assim_freq': 6, 'DOIAU': True, 'FHOUT_WAV': 1, 'py_pyyaml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2', 'met_ver': '9.1.3', 'gsi_ncdiag_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe', 'DO_FETCH_LOCAL': False, 'NLSPATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64/locale/%l_%t/%N:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin/locale/%l_%t/%N', 'netcdf_fortran_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe', 'SP_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_4.a', 'COM_OCEAN_BMATRIX_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean', 'SLURM_CONF': '/var/spool/slurmd/conf-cache/slurm.conf', 'PATH': '/apps/other/globus-cli-3.35.2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r/ush:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/bin:/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/bin/intel64:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/bin:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/bin:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin:/usr/sbin:/usr/bin:/apps/sbin:/apps/bin:/opt/slurm/bin:/home/gfekete/sven/bin', 'SP_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_8.a', 'netcdf_c_ver': '4.9.2', 'SLURM_JOB_NAME': 'C48_S2SW_gfs_ice_prod_f066_12', 'COM_ICE_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history', 'NLN': '/bin/ln -sf', 'MODULEPATH': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles:/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0-a66eaip/g2tmpl/1.13.0/intel/2021.9.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/g2tmpl/1.13.0/intel/2021.9.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/intel-oneapi-mpi/2021.9.0-a66eaip/oneapi/2023.1.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core:/work2/noaa/global/mterry/global-workflow_forked/modulefiles:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core:/apps/other/modulefiles:/apps/containers/modulefiles:/apps/licensed/modulefiles:/apps/contrib/modulefiles', 'CC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icc', 'grib_util_ver': '1.3.0', 'py_f90nml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z', 'DO_MERGENSST': False, 'SLURM_NTASKS_PER_NODE': 1, '_LMFILES_': '/apps/other/modulefiles/contrib/0.1:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core/intel-oneapi-compilers/2023.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core/stack-intel/2021.9.0.lua:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0/intel-oneapi-mpi/2021.9.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/stack-intel-oneapi-mpi/2021.9.0.lua:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core/intel-oneapi-mkl/2023.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/zlib/1.2.13.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/pigz/2.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/zstd/1.5.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/tar/1.34.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gettext/0.21.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libxcrypt/4.4.35.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/sqlite/3.43.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/util-linux-uuid/2.38.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/python/3.11.6.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libjpeg/2.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/jasper/2.0.32.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libpng/1.6.37.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/openjpeg/2.3.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/eccodes/2.32.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/fftw/3.3.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/nghttp2/1.57.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/curl/8.4.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/proj/9.2.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/udunits/2.2.28.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/cdo/2.2.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/hdf5/1.14.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/snappy/1.1.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/c-blosc/1.21.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-c/4.9.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-fortran/4.6.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/parallel-netcdf/1.12.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/parallelio/2.5.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/esmf/8.6.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/antlr/2.7.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gsl/2.7.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/nco/5.0.6.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/bacio/2.4.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/w3emc/2.10.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/prod_util/2.1.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2/3.4.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/sp/2.5.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/ip/4.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/grib-util/1.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2tmpl/1.13.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/gsi-ncdiag/1.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/crtm-fix/2.4.0.1_emc.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/git-lfs/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/crtm/2.4.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/openblas/0.3.24.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-setuptools/63.4.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-numpy/1.23.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/bufr/11.7.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gmake/4.2.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/wgrib2/2.0.8.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-f90nml/1.4.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-cftime/1.0.3.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/py-netcdf4/1.5.8.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libyaml/0.2.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pyyaml/6.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-markupsafe/2.1.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-jinja2/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-bottleneck/1.3.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-numexpr/2.8.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-et-xmlfile/1.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-openpyxl/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pytz/2023.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pyxlsb/1.0.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlrd/2.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlsxwriter/3.1.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlwt/1.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pandas/1.5.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-six/1.16.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-python-dateutil/2.8.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2c/1.8.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-cxx4/4.3.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/met/9.1.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/metplus/3.1.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-packaging/23.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xarray/2023.7.0.lua:/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles/prepobs/1.1.0.lua:/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles/fit2obs/1.1.7.1.lua:/apps/other/modulefiles/globus-cli/3.35.2.lua:/work2/noaa/global/mterry/global-workflow_forked/modulefiles/module_base.hercules.lua', 'globus_cli_ver': 3.27, '_ModuleTable028_': 'L2xpYnBuZy8xLjYuMzcubHVhIiwKZnVsbE5hbWUgPSAibGlicG5nLzEuNi4zNyIsCmxvYWRPcmRlciA9IDE4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImxpYnBuZy8xLjYuMzciLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA2LjAwMDAwMDAzNy4qemZpbmFsIiwKfSwKbGlieGNyeXB0ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2xpYnhjcnlwdC80LjQuMzUubHVhIiwKZnVsbE5hbWUgPSAibGlieGNyeXB0', 'py_numpy_ver': '1.23.4', '__LMOD_REF_COUNT_HDF5_PLUGIN_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/plugins:1', 'bufr_ver': '11.7.0', 'COMINnam': '/work2/noaa/global/role-global/data/external_gempak/nam', 'DBUS_SESSION_BUS_ADDRESS': 'unix:path=/run/user/9583/bus', 'DO_AERO_FCST': False, 'py_pandas_ver': '1.5.3', '_ModuleTable039_': 'dGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9vcGVuanBlZy8yLjMuMS5sdWEiLApmdWxsTmFtZSA9ICJvcGVuanBlZy8yLjMuMSIsCmxvYWRPcmRlciA9IDE5LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm9wZW5qcGVnLzIuMy4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsicGFyYWxsZWwtbmV0Y2RmIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMv', 'COM_ICE_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice', 'UUID_ORION_DTN': '8a10dd4f-24ee-4794-a39d-9c313ab6a34b', 'MPI_F77': 'mpiifort', 'LMOD_CMD': '/apps/other/lmod/lmod/libexec/lmod', '_ModuleTable045_': 'aW5hbCIsCn0sClsicHktYm90dGxlbmVjayJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LWJvdHRsZW5lY2svMS4zLjcubHVhIiwKZnVsbE5hbWUgPSAicHktYm90dGxlbmVjay8xLjMuNyIsCmxvYWRPcmRlciA9IDYzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWJvdHRsZW5lY2svMS4zLjciLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAzLjAwMDAwMDAwNy4qemZpbmFs', 'LMOD_MPI_VERSION': '2021.9.0-a66eaip', 'LMOD_SYSTEM_NAME': 'hercules', 'RUN_ENVIR': 'emc', 'w3emc_ver': '2.10.0', '__LMOD_REF_COUNT_LD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/host/linux64/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/lib:1;/usr/lib64:1;/usr/lib:1;/opt/slurm/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:2', 'DOHYBVAR_OCN': False, 'NUM_SND_COLLECTIVES': 9, 'walltime': '00:15:00', 'netcdf_cxx4_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu', 'ens_tracker_ver': 'v1.2.0', 'LMOD_SYSTEM_DEFAULT_MODULES': 'contrib', 'OMPI_MCA_plm_slurm_args': '--external-launcher', '_ModuleTable078_': 'cyIKLCAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLWE2NmVhaXAvZzJ0bXBsLzEuMTMuMC9pbnRlbC8yMDIxLjkuMCIKLCAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9nMnRtcGwvMS4xMy4wL2ludGVsLzIwMjEuOS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVz', 'FIXcpl': '/work2/noaa/global/mterry/global-workflow_forked/fix/cpl', 'G2C_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64/libg2c.so', '_ModuleTable019_': 'cmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9ncmliLXV0aWwvMS4zLjAubHVhIiwKZnVsbE5hbWUgPSAiZ3JpYi11dGlsLzEuMy4wIiwKbG9hZE9yZGVyID0gNDQsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ3JpYi11dGlsLzEuMy4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKWyJnc2ktbmNkaWFnIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMv', 'SLURM_JOB_GID': 17000, 'SLURM_GET_USER_ENV': 1, 'IAUFHRS_ENKF': [3, 6, 9], 'CPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/include:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/include:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/include', 'G2TMPL_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include', 'DO_JEDISNOWDA': False, 'SP_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_d.a', 'pigz_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq', 'COM_ATMOS_BUFR_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr', 'COM_OCEAN_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2', 'CNVGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/cnvgrib', 'COM_ATMOS_TRACK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks', 'DO_VERFRAD': True, 'pgmerr': 'errfile', 'MPI_BUFS_PER_HOST': 2048, 'DO_PREP_OBS_AERO': False, 'LMOD_FAMILY_G2TMPLVIRT_VERSION': '1.13.0', 'PNG_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu', 'COM_ATMOS_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart', 'DEGRIB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/degrib2', 'DO_ANLSTAT': False, 'LMOD_MPI_NAME': 'intel-oneapi-mpi', 'py_f90nml_ver': '1.4.3', 'NMEM_ENS_GFS_OFFSET': 20, 'SMOOTH_ENKF': False, 'FHMAX_GOES': 120, 'OLDPWD': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', 'QUEUE': 'batch', 'COM_OBS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/obs', 'DO_VMINMON': True, 'metplus_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r', 'VERBOSE': True, 'SLURM_JOB_NODELIST': 'hercules-08-38', 'CLUSTER': 'hercules', 'DBNROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn', 'FHOUT_HF_WAV': 1, 'PARMgfs': '/work2/noaa/global/mterry/global-workflow_forked/parm', 'I_MPI_HYDRA_BOOTSTRAP': 'slurm', 'APRUN_OCNICEPOST': 'srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1', 'BASH_FUNC_ml%%': '() { eval "$($LMOD_DIR/ml_cmd "$@")"\n}', 'BASH_FUNC_which%%': '() { ( alias;\n eval ${which_declare} ) | /usr/bin/which --tty-only --read-alias --read-functions --show-tilde --show-dot $@\n}', 'BASH_FUNC_module%%': '() { if [ -z "${LMOD_SH_DBG_ON+x}" ]; then\n case "$-" in \n *v*x*)\n __lmod_sh_dbg=\'vx\'\n ;;\n *v*)\n __lmod_sh_dbg=\'v\'\n ;;\n *x*)\n __lmod_sh_dbg=\'x\'\n ;;\n esac;\n fi;\n if [ -n "${__lmod_sh_dbg:-}" ]; then\n set +$__lmod_sh_dbg;\n echo "Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for Lmod\'s output" 1>&2;\n fi;\n eval "$($LMOD_CMD $LMOD_SHELL_PRGM "$@")" && eval "$(${LMOD_SETTARG_CMD:-:} -s sh)";\n __lmod_my_status=$?;\n if [ -n "${__lmod_sh_dbg:-}" ]; then\n echo "Shell debugging restarted" 1>&2;\n set -$__lmod_sh_dbg;\n fi;\n unset __lmod_sh_dbg;\n return $__lmod_my_status\n}', 'BASH_FUNC_declare_from_tmpl%%': '() { if [[ ${DEBUG_WORKFLOW:-"NO"} == "NO" ]]; then\n set +x;\n fi;\n local opts="-g";\n local OPTIND=1;\n while getopts "rx" option; do\n opts="${opts}${option}";\n done;\n shift $((OPTIND-1));\n for input in "$@";\n do\n IFS=\':\' read -ra args <<< "${input}";\n local com_var="${args[0]}";\n local template;\n local value;\n if (( ${#args[@]} > 1 )); then\n template="${args[1]}";\n else\n template="${com_var}_TMPL";\n fi;\n if [[ ! -v "${template}" ]]; then\n echo "FATAL ERROR in declare_from_tmpl: Requested template ${template} not defined!";\n exit 2;\n fi;\n value=$(echo "${!template}" | envsubst);\n declare ${opts} "${com_var}"="${value}";\n echo "declare_from_tmpl :: ${com_var}=${value}";\n done;\n set_trace\n}', 'BASH_FUNC_err_exit%%': ['() { set +eux;\n msg1=${*:-Job ${jobid} failed};\n if [[ -n "${pgm}" ]]; then\n msg1+="', 'ERROR IN ${pgm}";\n fi;\n if [[ -n "${err}" ]]; then\n msg1+=" RETURN CODE ${err}";\n fi;\n msg2="\n -------------------------------------------------------------\n -- FATAL ERROR: ${msg1}\n -- ABNORMAL EXIT at $(date) on ${HOSTNAME}\n -------------------------------------------------------------\n ";\n echo "${msg2}" 1>&2;\n module list;\n echo "" 1>&2;\n echo "${msg1}" 1>&2;\n if [[ -n "${DATA}" ]]; then\n echo "${DATA}" 1>&2;\n ls -ltr "${DATA}" 1>&2;\n else\n echo "WARNING: DATA variable not defined" 1>&2;\n fi;\n if [[ -n "${pgmout}" ]]; then\n if [[ -s errfile ]]; then\n echo "----- contents of errfile -----" >> "${pgmout}";\n cat errfile >> "${pgmout}";\n fi;\n cat "${pgmout}" 1>&2;\n else\n if [[ -s errfile ]]; then\n cat errfile 1>&2;\n fi;\n fi;\n if [[ "${SENDECF}" == "YES" ]]; then\n timeout 30 ecflow_client --msg "${ECF_NAME}: ${msg1}";\n timeout 30 ssh "${ECF_HOST}" "echo \\"${msg}2\\" >> ${ECF_JOBOUT:?}";\n fi;\n if [[ "${SENDECF}" == "YES" ]]; then\n ecflow_client --kill="${ECF_NAME:?}";\n fi;\n if [[ -n "${PBS_JOBID}" ]]; then\n qdel "${PBS_JOBID}";\n else\n if [[ -n "${SLURM_JOB_ID}" ]]; then\n scancel "${SLURM_JOB_ID}";\n fi;\n fi\n}'], 'BASH_FUNC_wait_for_file%%': '() { set +x;\n local file_name=${1:?"wait_for_file() requires a file name"};\n local sleep_interval=${2:-60};\n local max_tries=${3:-100};\n for ((iter=0; iter' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.oceanice_products -+++ config.oceanice_products[5]echo 'BEGIN: config.oceanice_products' -BEGIN: config.oceanice_products -+++ config.oceanice_products[8]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources oceanice_products -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=oceanice_products -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[978]walltime=00:15:00 -++++ config.resources[979]ntasks=1 -++++ config.resources[980]tasks_per_node=1 -++++ config.resources[981]threads_per_task=1 -++++ config.resources[982]memory=96GB -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n '' ]] -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export memory -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.oceanice_products[11]export MAX_TASKS=25 -+++ config.oceanice_products[11]MAX_TASKS=25 -+++ config.oceanice_products[13]export OCEANICEPRODUCTS_CONFIG=/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml -+++ config.oceanice_products[13]OCEANICEPRODUCTS_CONFIG=/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml -+++ config.oceanice_products[16]export NFHRS_PER_GROUP=3 -+++ config.oceanice_products[16]NFHRS_PER_GROUP=3 -+++ config.oceanice_products[18]echo 'END: config.oceanice_products' -END: config.oceanice_products -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env oceanice_products -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=oceanice_products -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 100663296 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 1 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 1 ]] -+++ HERCULES.env[34]max_threads_per_task=80 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 80 ]] -+++ HERCULES.env[40][[ 1 -gt 80 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 1' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[286]export NTHREADS_OCNICEPOST=1 -+++ HERCULES.env[286]NTHREADS_OCNICEPOST=1 -+++ HERCULES.env[287]export 'APRUN_OCNICEPOST=srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1' -+++ HERCULES.env[287]APRUN_OCNICEPOST='srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1' -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_OCEANICE_PRODUCTS[11]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[11]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[11]declare_from_tmpl -rx COMIN_ICE_HISTORY:COM_ICE_HISTORY_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_ICE_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/ice/history -+ JGLOBAL_OCEANICE_PRODUCTS[12]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[12]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[12]declare_from_tmpl -rx COMOUT_ICE_GRIB:COM_ICE_GRIB_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_ICE_GRIB=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ice/grib2 -+ JGLOBAL_OCEANICE_PRODUCTS[13]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[13]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[13]declare_from_tmpl -rx COMOUT_ICE_NETCDF:COM_ICE_NETCDF_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_ICE_NETCDF=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ice/netcdf -+ JGLOBAL_OCEANICE_PRODUCTS[17]/work2/noaa/global/mterry/global-workflow_forked/scripts/exglobal_oceanice_products.py -2025-07-28 21:34:54,440 - INFO - root : BEGIN: __main__.main -2025-07-28 21:34:54,440 - DEBUG - root : ( ) -2025-07-28 21:34:54,448 - INFO - oceanice_products: BEGIN: OceanIceProducts.__init__ -2025-07-28 21:34:54,448 - DEBUG - oceanice_products: ( , {'SHELL': '/bin/bash', 'PDY': datetime.datetime(2021, 3, 23, 0, 0), 'DO_JEDIATMENS': False, 'sqlite_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4', 'COPYGB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/copygb', 'WORK_hercules': '/work/hercules', 'TOCGRIB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib2', 'G2_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/include_d', 'PIO_TYPENAME_VALID_VALUES': ['netcdf', 'netcdf4p', 'netcdf4c', 'pnetcdf'], 'WGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/wgrib', 'SERIAL_CXX': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icpc', 'FHMAX_WAV': 9, 'FIXlut': '/work2/noaa/global/mterry/global-workflow_forked/fix/lut', 'fftw_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj', 'snappy_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx', '_ModuleTable016_': 'bGxOYW1lID0gImdldHRleHQvMC4yMS4xIiwKbG9hZE9yZGVyID0gMTEsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ2V0dGV4dC8wLjIxLjEiLAp3ViA9ICIwMDAwMDAwMDAuMDAwMDAwMDIxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJnaXQtbGZzIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvZ2l0LWxmcy8zLjEuMi5sdWEiLApmdWxsTmFtZSA9ICJnaXQtbGZzLzMuMS4yIiwK', 'MAKE_NSSTBUFR': False, 'cdo_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e', 'curl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops', 'PIO_VERSION_MAJOR': 2, 'threads_per_task': 1, 'SLURM_JOB_USER': 'mterry', '_ModuleTable067_': 'NC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3NuYXBweS8xLjEuMTAubHVhIiwKZnVsbE5hbWUgPSAic25hcHB5LzEuMS4xMCIsCmxvYWRPcmRlciA9IDI4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInNuYXBweS8xLjEuMTAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAxLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKc3AgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45', 'COM_ICE_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}', 'COM_OCEAN_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}', '_ModuleTable022_': 'ZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9oZGY1LzEuMTQuMC5sdWEiLApmdWxsTmFtZSA9ICJoZGY1LzEuMTQuMCIsCmxvYWRPcmRlciA9IDI3LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImhkZjUvMS4xNC4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAxNC4qemZpbmFsIiwKfSwKWyJpbnRlbC1vbmVhcGktY29tcGlsZXJzIl0gPSB7CmZuID0gIi9hcHBzL3NwYWNrLW1hbmFnZWQvbW9kdWxlZmlsZXMvbGludXgtcm9ja3k5LXg4Nl82NC9Db3JlL2ludGVsLW9uZWFwaS1jb21waWxlcnMvMjAyMy4xLjAubHVhIiwKZnVsbE5hbWUgPSAiaW50ZWwtb25lYXBpLWNvbXBpbGVy', 'SLURM_TASKS_PER_NODE': 1, 'COM_ATMOS_OZNMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon', 'COM_ATMOS_GEMPAK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}', 'SLURM_JOB_UID': 9583, 'HISTCONTROL': 'ignoredups', 'SLURM_EXPORT_ENV': 'NONE', 'g2c_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp', 'COM_OCEAN_NETCDF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf', 'GRB2INDEX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grb2index', 'l4densvar': True, 'FHOUT_HF_GFS': 1, 'I_MPI_EXTRA_FILESYSTEM': 1, 'SLURM_TASK_PID': 357858, 'EXECprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/exec', 'g2tmpl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel', 'MPI_BUFS_PER_PROC': 2048, 'IAU_DELTHRS_ENKF': 6, 'COM_WAVE_WMO_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo', '_ModuleTable027_': 'LTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9saWJqcGVnLzIuMS4wLmx1YSIsCmZ1bGxOYW1lID0gImxpYmpwZWcvMi4xLjAiLApsb2FkT3JkZXIgPSAxNiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJsaWJqcGVnLzIuMS4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKbGlicG5nID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4w', 'NMEM_ENS': 0, '__LMOD_REF_COUNT_CLASSPATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/mpi.jar:1', '_ModuleTable050_': 'ZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbWFya3Vwc2FmZS8yLjEuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1tYXJrdXBzYWZlLzIuMS4zIiwKbG9hZE9yZGVyID0gNjEsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktbWFya3Vwc2FmZS8yLjEuMyIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDEuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LW5ldGNkZjQiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFj', 'err': 0, '_ModuleTable032_': 'CnVzZXJOYW1lID0gIm1ldHBsdXMvMy4xLjEiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJtb2R1bGVfYmFzZS5oZXJjdWxlcyJdID0gewpmbiA9ICIvd29yazIvbm9hYS9nbG9iYWwvbXRlcnJ5L2dsb2JhbC13b3JrZmxvd19mb3JrZWQvbW9kdWxlZmlsZXMvbW9kdWxlX2Jhc2UuaGVyY3VsZXMubHVhIiwKZnVsbE5hbWUgPSAibW9kdWxlX2Jhc2UuaGVyY3VsZXMiLApsb2FkT3JkZXIgPSA4NCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDAsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJtb2R1bGVfYmFzZS5oZXJjdWxlcyIsCndWID0gIk0uKnpmaW5hbCIsCn0sCm5jbyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGlj', 'CHGRP_CMD': 'chgrp rstprod', 'PKG_CONFIG_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib/pkgconfig:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/pkgconfig:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/lib/pkgconfig', 'SLURM_LOCALID': 0, 'MAX_TASKS': 25, '_ModuleTable040_': 'aW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9wYXJhbGxlbC1uZXRjZGYvMS4xMi4yLmx1YSIsCmZ1bGxOYW1lID0gInBhcmFsbGVsLW5ldGNkZi8xLjEyLjIiLApsb2FkT3JkZXIgPSAzMiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwYXJhbGxlbC1uZXRjZGYvMS4xMi4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAxMi4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCnBhcmFsbGVsaW8gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0', 'SLURM_SUBMIT_DIR': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', '_ModuleTable077_': 'LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvenN0ZC8xLjUuMi5sdWEiLApmdWxsTmFtZSA9ICJ6c3RkLzEuNS4yIiwKbG9hZE9yZGVyID0gOSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMywKc3RhY2tEZXB0aCA9IDQsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJ6c3RkLzEuNS4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwNS4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCn0sCm1wYXRoQSA9IHsKIi93b3JrMi9ub2FhL2dsb2JhbC9yb2xlLWdsb2JhbC9naXQvRml0Mk9icy92MS4xLjcuMS9tb2R1bGVmaWxlcyIKLCAiL3dvcmsyL25vYWEvZ2xvYmFsL3JvbGUtZ2xvYmFsL2dpdC9wcmVwb2JzL3YxLjEuMC9tb2R1bGVmaWxl', 'HOSTNAME': 'hercules-08-38', 'HISTSIZE': 1000, 'wgrib2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln', 'FRAC_GRID': True, 'G2_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/include_4', 'intel_oneapi_mpi_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse', 'SLURMD_NODENAME': 'hercules-08-38', '_ModuleTable080_': 'bGVzL2xpbnV4LXJvY2t5OS14ODZfNjQvb25lYXBpLzIwMjMuMS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL0NvcmUiCiwgIi93b3JrMi9ub2FhL2dsb2JhbC9tdGVycnkvZ2xvYmFsLXdvcmtmbG93X2ZvcmtlZC9tb2R1bGVmaWxlcyIKLCAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L0NvcmUiCiwgIi9hcHBzL290aGVyL21vZHVsZWZpbGVzIiwgIi9hcHBzL2NvbnRhaW5lcnMvbW9kdWxlZmlsZXMiCiwgIi9hcHBzL2xpY2Vuc2VkL21vZHVsZWZpbGVzIiwg', '__LMOD_REF_COUNT_OCL_ICD_FILENAMES': 'libintelocl_emu.so:1;libalteracl.so:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64/libintelocl.so:1', 'c_blosc_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn', 'FHOUT_OCN': 3, 'COM_OBS_JEDI': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi', 'eccodes_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj', '_ModuleTable037_': 'ay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9uZ2h0dHAyLzEuNTcuMC5sdWEiLApmdWxsTmFtZSA9ICJuZ2h0dHAyLzEuNTcuMCIsCmxvYWRPcmRlciA9IDIyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gNCwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm5naHR0cDIvMS41Ny4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDA1Ny4qemZpbmFsIiwKfSwKb3BlbmJsYXMgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2kt', 'netcdf_c_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy', '_ModuleTable070_': 'c3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzdGFjay1pbnRlbC8yMDIxLjkuMCIsCndWID0gIjAwMDAwMjAyMS4wMDAwMDAwMDkuKnpmaW5hbCIsCn0sClsic3RhY2staW50ZWwtb25lYXBpLW1waSJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3N0YWNrLWludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAubHVhIiwKZnVsbE5hbWUgPSAic3RhY2staW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMCIsCmxvYWRPcmRlciA9IDUsCnByb3BUID0g', 'SLURM_JOB_START_TIME': 1753756458, '_ModuleTable011_': 'b3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZXNtZi84LjYuMCIsCndWID0gIjAwMDAwMDAwOC4wMDAwMDAwMDYuKnpmaW5hbCIsCn0sCmZmdHcgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9mZnR3LzMuMy4xMC5sdWEiLApmdWxsTmFtZSA9ICJmZnR3LzMuMy4xMCIsCmxvYWRPcmRlciA9IDIxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwK', 'DO_NEST': False, 'STMP': '/work2/noaa/stmp/mterry/HERCULES', 'HYDRA_LAUNCHER_EXTRA_ARGS': '--external-launcher', 'ACL_BOARD_VENDOR_PATH': '/opt/Intel/OpenCLFPGA/oneAPI/Boards', '_ModuleTable060_': 'ZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1weXlhbWwvNi4wIiwKd1YgPSAiMDAwMDAwMDA2Lip6ZmluYWwiLAp9LApbInB5LXNldHVwdG9vbHMiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1zZXR1cHRvb2xzLzYzLjQuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1zZXR1cHRvb2xzLzYzLjQuMyIsCmxvYWRPcmRlciA9IDUxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIs', '_ModuleTable007_': 'ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9jcnRtLzIuNC4wLjEubHVhIiwKZnVsbE5hbWUgPSAiY3J0bS8yLjQuMC4xIiwKbG9hZE9yZGVyID0gNDksCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY3J0bS8yLjQuMC4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNC4wMDAwMDAwMDAuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbImNydG0tZml4Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45', 'G2TMPL_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib/libg2tmpl.a', 'DO_WAVE': True, 'SCRgfs': '/work2/noaa/global/mterry/global-workflow_forked/scripts', 'SP_INC8': 'include_8', '_ModuleTable002_': 'b2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9hbnRsci8yLjcuNy5sdWEiLApmdWxsTmFtZSA9ICJhbnRsci8yLjcuNyIsCmxvYWRPcmRlciA9IDM1LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImFudGxyLzIuNy43IiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4wMDAwMDAwMDcuKnpmaW5hbCIsCn0sCmJhY2lvID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2JhY2lvLzIuNC4xLmx1', 'MPICC': 'mpiicc', 'HDF5_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh', '_ModuleTable055_': 'L21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXBhY2thZ2luZy8yMy4xLmx1YSIsCmZ1bGxOYW1lID0gInB5LXBhY2thZ2luZy8yMy4xIiwKbG9hZE9yZGVyID0gNzksCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcGFja2FnaW5nLzIzLjEiLAp3ViA9ICIwMDAwMDAwMjMuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbInB5LXBhbmRhcyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEu', 'SP_INC4': 'include_4', 'FPGA_VARS_DIR': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga', 'FHOUT': 3, '__LMOD_REF_COUNT_MODULEPATH': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles:1;/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0-a66eaip/g2tmpl/1.13.0/intel/2021.9.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/g2tmpl/1.13.0/intel/2021.9.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/intel-oneapi-mpi/2021.9.0-a66eaip/oneapi/2023.1.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core:1;/work2/noaa/global/mterry/global-workflow_forked/modulefiles:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core:1;/apps/other/modulefiles:1;/apps/containers/modulefiles:1;/apps/licensed/modulefiles:1;/apps/contrib/modulefiles:1', 'cycle': 't12z', '_ModuleTable043_': 'aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcmVwb2JzLzEuMS4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKcHJvZF91dGlsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3Byb2RfdXRpbC8yLjEuMS5sdWEiLApmdWxsTmFtZSA9ICJwcm9kX3V0aWwvMi4xLjEiLApsb2FkT3JkZXIgPSA0MCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcm9kX3V0aWwvMi4x', 'MPI_CC': 'mpiicc', 'PARTITION_DTN': '', 'sp_ver': '2.5.0', 'SHELLOPTS': 'braceexpand:errexit:hashall:interactive-comments:nounset:xtrace', 'DO_AWIPS': False, 'EDATE': datetime.datetime(2021, 3, 23, 12, 0), '__LMOD_REF_COUNT___INTEL_POST_CFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1'], 'I_MPI_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0', 'CRTM_FIX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2/fix', '_ModuleTable062_': 'MTYuMCIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMTYuKnpmaW5hbCIsCn0sClsicHkteGFycmF5Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHkteGFycmF5LzIwMjMuNy4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXhhcnJheS8yMDIzLjcuMCIsCmxvYWRPcmRlciA9IDgwLApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXhhcnJheS8yMDIzLjcuMCIsCndWID0gIjAwMDAwMjAyMy4wMDAwMDAwMDcuKnpm', 'proj_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez', 'FSYNC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file', '_ModuleTable005_': 'cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJjLWJsb3NjLzEuMjEuNSIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMjEuMDAwMDAwMDA1Lip6ZmluYWwiLAp9LApjZG8gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9jZG8vMi4yLjAubHVhIiwKZnVsbE5hbWUgPSAiY2RvLzIuMi4wIiwKbG9hZE9yZGVyID0gMjYsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY2RvLzIu', 'g2tmpl_ver': '1.13.0', 'FHOUT_GOES': 3, '_ModuleTable021_': 'MC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nc2wvMi43LjEubHVhIiwKZnVsbE5hbWUgPSAiZ3NsLzIuNy4xIiwKbG9hZE9yZGVyID0gMzYsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ3NsLzIuNy4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sCmhkZjUgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxl', 'EUPD_CYC': 'gdas', 'fms_ver': '2023.02.01', 'util_linux_uuid_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj', '_ModuleTable052_': 'Y2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbnVtZXhwci8yLjguNC5sdWEiLApmdWxsTmFtZSA9ICJweS1udW1leHByLzIuOC40IiwKbG9hZE9yZGVyID0gNjQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktbnVtZXhwci8yLjguNCIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDguMDAwMDAwMDA0Lip6ZmluYWwiLAp9LApbInB5LW51bXB5Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2kt', '_ModuleTable030_': 'cFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJsaWJ5YW1sLzAuMi41IiwKd1YgPSAiMDAwMDAwMDAwLjAwMDAwMDAwMi4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCm1ldCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL21ldC85LjEuMy5sdWEiLApmdWxsTmFtZSA9ICJtZXQvOS4xLjMiLApsb2FkT3JkZXIgPSA3NywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0', 'COM_ATMOS_IMAGERY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery', 'I_MPI_EXTRA_FILESYSTEM_LIST': 'lustre', 'COM_WAVE_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history', 'SLURM_CLUSTER_NAME': 'hercules', '__LMOD_REF_COUNT_ACLOCAL_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/aclocal:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/aclocal:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/aclocal:2', 'SERIAL_CC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icc', 'COM_OBSPROC_TMPL': '${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos', 'SLURM_JOB_END_TIME': 1753757358, '_ModuleTable075_': 'ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3dncmliMi8yLjAuOC5sdWEiLApmdWxsTmFtZSA9ICJ3Z3JpYjIvMi4wLjgiLApsb2FkT3JkZXIgPSA1NSwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJ3Z3JpYjIvMi4wLjgiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKemxpYiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJj', 'FHMAX_GFS': 120, 'SENDAWIP': False, 'G2_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64/libg2_d.a', 'spack_mod_path': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core', 'OCNRES': 500, 'COM_MED_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart', 'SLURM_CPUS_ON_NODE': 1, 'LMOD_FAMILY_METAMPI': 'stack-intel-oneapi-mpi', 'LEVS': 128, 'FIXgsi': '/work2/noaa/global/mterry/global-workflow_forked/fix/gsi', 'NFHRS_PER_GROUP': 3, 'COM_WAVE_PREP_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep', 'FI_PROVIDER_PATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib/prov:/usr/lib64/libfabric', 'python_ver': '3.11.6', 'IAU_OFFSET': 0, 'py_packaging_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6', 'DIAGUTIL_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/sys_check/sys_check.sh', 'python_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2', 'SCRIPTScfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/scripts', 'py_markupsafe_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7', 'G2C_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include', 'py_xlrd_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f', 'LMOD_G2TMPLVIRT_VERSION': '1.13.0', 'pid': 358493, 'OUTPUT_GRID': 'gaussian_grid', 'jobid': 'oceanice_products_ice_f072.357871', 'SLURM_JOB_CPUS_PER_NODE': 1, 'spack_env': 'gsi-addon-dev-fms-2024.01', 'INTEL_ONEAPI_MPI_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse', 'LMOD_FAMILY_METACOMPILER': 'stack-intel', 'SP_INCd': 'include_d', 'py_six_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o', 'COM_ATMOS_GOES_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim', 'MPICXX': 'mpiicpc', 'restart_interval_gdas': 3, 'fit2obs_ver': '1.1.7.1', 'LMOD_DIR': '/apps/other/lmod/lmod/libexec', '_ModuleTable068_': 'LjAvc3AvMi41LjAubHVhIiwKZnVsbE5hbWUgPSAic3AvMi41LjAiLApsb2FkT3JkZXIgPSA0MiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMiwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzcC8yLjUuMCIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCnNxbGl0ZSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9zcWxpdGUvMy40My4yLmx1YSIsCmZ1bGxOYW1lID0gInNxbGl0ZS8zLjQzLjIiLApsb2FkT3Jk', 'REPLAY_ICS': False, 'COM_ICE_NETCDF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf', 'G2_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64/libg2_4.a', '_ModuleTable058_': 'InB5LXB5dHovMjAyMy4zIiwKbG9hZE9yZGVyID0gNjcsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcHl0ei8yMDIzLjMiLAp3ViA9ICIwMDAwMDIwMjMuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LXB5eGxzYiJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXB5eGxzYi8xLjAuMTAubHVhIiwKZnVsbE5hbWUgPSAicHktcHl4bHNiLzEuMC4xMCIsCmxvYWRPcmRlciA9', 'py_setuptools_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr', 'FHMIN': 0, 'COM_ICE_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input', '_ModuleTable048_': 'cHktZjkwbm1sIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktZjkwbm1sLzEuNC4zLmx1YSIsCmZ1bGxOYW1lID0gInB5LWY5MG5tbC8xLjQuMyIsCmxvYWRPcmRlciA9IDU2LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWY5MG5tbC8xLjQuMyIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMDQuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LWppbmphMiJdID0gewpmbiA9ICIvd29yay9ub2FhL2Vw', 'intel_mkl_ver': '2023.1.0', 'PRTE_MCA_plm_slurm_args': '--external-launcher', 'PWD': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ice_f072.357871', 'SLURM_GTIDS': 0, 'LOGNAME': 'mterry', 'MAKE_ACFTBUFR': False, 'W3EMC_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_8', 'HOMEDIR': '/work2/noaa/global/mterry', 'IAU_DELTHRS': 6, 'W3EMC_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_4', 'XDG_SESSION_TYPE': 'unspecified', 'ESMFMKFILE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib/esmf.mk', 'FHMAX': 9, 'WRITE_NSFLIP': True, 'SLURM_JOB_PARTITION': 'hercules', 'YAML_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x', 'PARTITION_SERVICE': 'service', 'MODULESHOME': '/apps/other/lmod/lmod', '__LMOD_REF_COUNT_DYLD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:1', 'COM_WAVE_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart', 'BUFR_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_4.so', 'BUFR_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_8.so', 'HOMEcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1', 'COM_OCEAN_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon', 'DEBUG_POSTSCRIPT': False, 'MANPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/share/man:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/man:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/documentation/en/man/common:/apps/other/lmod/lmod/share/man:/usr/share/man:/apps/share/man:/apps/man:/opt/slurm/share/man::', 'libyaml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x', 'jasper_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt', 'py_xlsxwriter_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok', 'SLURM_TRES_PER_TASK': 'cpu=1', 'nco_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq', 'stack_intel_ver': '2021.9.0', 'SLURM_OOM_KILL_STEP': 0, 'BUFR_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m', 'bufr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m', 'CLUSTERS_DTN': '', 'grib_util_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35', 'nghttp2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky', 'DO_TEST_MODE': True, 'PACKAGEROOT': '/work2/noaa/global/role-global/nwpara', 'lobsdiag_forenkf': True, 'SLURM_JOB_NUM_NODES': 1, 'py_netcdf4_ver': '1.5.8', 'ICERES': 500, 'CXX': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icpc', 'GRIB2GRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grib2grib2', 'DMPDIR': '/work/noaa/rstprod/dump', 'LSOIL_INCR': 2, 'metplus_ver': '3.1.1', '_ModuleTable008_': 'LjAvY3J0bS1maXgvMi40LjAuMV9lbWMubHVhIiwKZnVsbE5hbWUgPSAiY3J0bS1maXgvMi40LjAuMV9lbWMiLApsb2FkT3JkZXIgPSA0NywKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJjcnRtLWZpeC8yLjQuMC4xX2VtYyIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDQuMDAwMDAwMDAwLjAwMDAwMDAwMS4qXy4qZW1jLip6ZmluYWwiLAp9LApjdXJsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIw', 'COM_OCEAN_LETKF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf', 'zlib_ver': '1.2.13', 'BUFR_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_d.so', 'ENKF_SPREAD': True, 'zlib_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j', '__LMOD_REF_COUNT_CMAKE_PREFIX_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r:2;/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/IntelDPCPP:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72:1', '_ModuleTable035_': 'c3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9uZXRjZGYtY3h4NC80LjMuMS5sdWEiLApmdWxsTmFtZSA9ICJuZXRjZGYtY3h4NC80LjMuMSIsCmxvYWRPcmRlciA9IDc2LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm5ldGNkZi1jeHg0LzQuMy4xIiwKd1YgPSAiMDAwMDAwMDA0LjAwMDAwMDAwMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsibmV0Y2RmLWZvcnRyYW4iXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFj', 'FIXgfs': '/work2/noaa/global/mterry/global-workflow_forked/fix', 'DO_COUPLED': True, 'SLURM_JOBID': 5951734, 'SERIAL_FC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'COM_ATMOS_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input', 'W3EMC_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_d', '_ModuleTable013_': 'IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nMi8zLjQuNS5sdWEiLApmdWxsTmFtZSA9ICJnMi8zLjQuNSIsCmxvYWRPcmRlciA9IDQxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImcyLzMuNC41IiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAwNC4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCmcyYyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJj', 'COM_ATMOS_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history', 'FIXam': '/work2/noaa/global/mterry/global-workflow_forked/fix/am', 'hdf5_ver': '1.14.0', 'restart_interval_enkfgdas': 3, 'COM_ATMOS_RADMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon', 'DO_CALC_INCREMENT': False, 'DO_STARTMEM_FROM_JEDIICE': False, 'FIXprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/fix', 'SLURM_JOB_QOS': 'batch', 'I_MPI_HYDRA_BOOTSTRAP_EXEC_EXTRA_ARGS': '--external-launcher', 'obsproc_run_ver': '1.2.0', 'COM_ATMOS_MINMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon', 'crtm_ver': '2.4.0.1', 'FIXorog': '/work2/noaa/global/mterry/global-workflow_forked/fix/orog', 'COM_OCEAN_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history', '__LMOD_REF_COUNT_PATH': '/apps/other/globus-cli-3.35.2/bin:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r/ush:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/bin:2;/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/bin:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/bin/intel64:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/bin:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/bin:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/bin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/bin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin:1;/usr/sbin:1;/usr/bin:1;/apps/sbin:1;/apps/bin:1;/opt/slurm/bin:1;/home/gfekete/sven/bin:1', '_ModuleTable018_': 'KnpmaW5hbCIsCn0sCmdtYWtlID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2dtYWtlLzQuMi4xLmx1YSIsCmZ1bGxOYW1lID0gImdtYWtlLzQuMi4xIiwKbG9hZE9yZGVyID0gNTQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ21ha2UvNC4yLjEiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDAyLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJncmliLXV0aWwiXSA9IHsKZm4gPSAiL3dv', 'COMIN_ICE_HISTORY': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/ice/history', 'COMROOT': '/work2/noaa/global/mterry/RUNTESTS/COMROOT', 'HOME': '/home/mterry', 'FI_PROVIDER': 'mlx', 'met_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b', 'COM_RTOFS_TMPL': '${DMPDIR}', '_ModuleTable015_': 'bnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2cydG1wbC8xLjEzLjAubHVhIiwKZnVsbE5hbWUgPSAiZzJ0bXBsLzEuMTMuMCIsCmxvYWRPcmRlciA9IDQ1LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImcydG1wbC8xLjEzLjAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDEzLip6ZmluYWwiLAp9LApnZXR0ZXh0ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2dldHRleHQvMC4yMS4xLmx1YSIsCmZ1', '_ModuleTable_Sz_': 81, 'LANG': 'C.UTF-8', '__LMOD_REF_COUNT_LIBRARY_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:1', 'DO_BUFRSND': False, 'memory': '96GB', '_ModuleTable064_': 'PSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHkteGxzeHdyaXRlci8zLjEuNy5sdWEiLApmdWxsTmFtZSA9ICJweS14bHN4d3JpdGVyLzMuMS43IiwKbG9hZE9yZGVyID0gNzAsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHkteGxzeHdyaXRlci8zLjEuNyIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDA3Lip6ZmluYWwiLAp9LApbInB5LXhsd3QiXSA9IHsKZm4gPSAi', 'DYLD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib', 'PARTITION_BATCH': 'hercules', '_ModuleTable023_': 'cy8yMDIzLjEuMCIsCmxvYWRPcmRlciA9IDIsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiaW50ZWwtb25lYXBpLWNvbXBpbGVycy8yMDIzLjEuMCIsCndWID0gIjAwMDAwMjAyMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsiaW50ZWwtb25lYXBpLW1rbCJdID0gewpmbiA9ICIvYXBwcy9zcGFjay1tYW5hZ2VkL21vZHVsZWZpbGVzL2xpbnV4LXJvY2t5OS14ODZfNjQvQ29yZS9pbnRlbC1vbmVhcGktbWtsLzIwMjMuMS4wLmx1YSIsCmZ1bGxOYW1lID0gImludGVsLW9uZWFwaS1ta2wvMjAyMy4xLjAiLApsb2FkT3JkZXIgPSA2LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVz', '_ModuleTable081_': 'Ii9hcHBzL2NvbnRyaWIvbW9kdWxlZmlsZXMiLAp9LApzeXN0ZW1CYXNlTVBBVEggPSAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L0NvcmU6L2FwcHMvb3RoZXIvbW9kdWxlZmlsZXM6L2FwcHMvY29udGFpbmVycy9tb2R1bGVmaWxlczovYXBwcy9saWNlbnNlZC9tb2R1bGVmaWxlcyIsCn0K', 'REDERR': '2>', 'PSLOT': 'C48_S2SW', '__LMOD_REF_COUNT_PKG_CONFIG_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib/pkgconfig:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/pkgconfig:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib/pkgconfig:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/pkgconfig:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib/pkgconfig:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/lib/pkgconfig:1', '_ModuleTable036_': 'ay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL25ldGNkZi1mb3J0cmFuLzQuNi4xLmx1YSIsCmZ1bGxOYW1lID0gIm5ldGNkZi1mb3J0cmFuLzQuNi4xIiwKbG9hZE9yZGVyID0gMzEsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAibmV0Y2RmLWZvcnRyYW4vNC42LjEiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDA2LjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKbmdodHRwMiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFj', 'COM_ATMOS_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}', 'USHcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/ush', 'DO_GEMPAK': False, 'SLURM_PROCID': 0, 'USHgfs': '/work2/noaa/global/mterry/global-workflow_forked/ush', 'DOIBP_WAV': False, 'prepobs_run_ver': '1.1.0', 'EXPDIR': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', 'job': 'oceanice_products', 'imp_physics': 8, 'LMOD_SHELL_PRGM': 'bash', 'IAU_FHROT': 0, 'MPIF90': 'mpiifort', 'IP_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_4.a', 'waveGRD': 'uglo_100km', 'COM_CHEM_BMAT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix', 'jasper_ver': '2.0.32', 'SCRATCH': '/scratch/hercules', 'IP_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_8.a', 'CLIENT_GLOBUS_UUID': '869912fe-f6de-46c0-af10-b22efd84a022', 'ACCOUNT': 'fv3-cpu', 'HOMEgfs': '/work2/noaa/global/mterry/global-workflow_forked', 'NHOUR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/nhour', 'binary_diag': False, 'COM_WAVE_GRID_RES_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}', 'SCRIPTSprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/scripts', 'MP_STDOUTMODE': 'ORDERED', 'sigio_ver': '2.3.2', '__LMOD_REF_COUNT_CPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/include:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/include:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/include:1', 'LMOD_SETTARG_FULL_SUPPORT': False, 'COMINsyn': '/work2/noaa/global/role-global/com/gfs/prod/syndat', 'OFFSET_START_HOUR': 0, 'REDOUT': '1>', 'PTMP': '/work2/noaa/stmp/mterry/HERCULES', '_ModuleTable076_': 'dWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC96bGliLzEuMi4xMy5sdWEiLApmdWxsTmFtZSA9ICJ6bGliLzEuMi4xMyIsCmxvYWRPcmRlciA9IDcsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEwLApzdGFja0RlcHRoID0gNSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInpsaWIvMS4yLjEzIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMi4wMDAwMDAwMTMuKnpmaW5hbCIsCn0sCnpzdGQgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2', 'g2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato', 'MODE': 'forecast-only', 'DATAROOT': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312', 'bacio_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj', 'APP': 'S2SW', '_ModuleTable029_': 'LzQuNC4zNSIsCmxvYWRPcmRlciA9IDEyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImxpYnhjcnlwdC80LjQuMzUiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDA0LjAwMDAwMDAzNS4qemZpbmFsIiwKfSwKbGlieWFtbCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9saWJ5YW1sLzAuMi41Lmx1YSIsCmZ1bGxOYW1lID0gImxpYnlhbWwvMC4yLjUiLApsb2FkT3JkZXIgPSA1OSwKcHJv', 'openblas_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f', 'stack_impi_ver': '2021.9.0', 'COM_CHEM_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem', 'TMPDIR': '/local/scratch/mterry/5951734', 'HDF5_PLUGIN_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/plugins', 'DO_TRACKER': True, 'CMAKE_PREFIX_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r:/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/IntelDPCPP:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72', 'crtm_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r', '_ModuleTable038_': 'YWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvb3BlbmJsYXMvMC4zLjI0Lmx1YSIsCmZ1bGxOYW1lID0gIm9wZW5ibGFzLzAuMy4yNCIsCmxvYWRPcmRlciA9IDUwLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm9wZW5ibGFzLzAuMy4yNCIsCndWID0gIjAwMDAwMDAwMC4wMDAwMDAwMDMuMDAwMDAwMDI0Lip6ZmluYWwiLAp9LApvcGVuanBlZyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5z', 'nco_ver': '5.0.6', '_ModuleTable046_': 'IiwKfSwKWyJweS1jZnRpbWUiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1jZnRpbWUvMS4wLjMuNC5sdWEiLApmdWxsTmFtZSA9ICJweS1jZnRpbWUvMS4wLjMuNCIsCmxvYWRPcmRlciA9IDU3LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWNmdGltZS8xLjAuMy40IiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMC4wMDAwMDAwMDMuMDAwMDAwMDA0Lip6ZmluYWwiLAp9', 'GRBINDEX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grbindex', 'libjpeg_turbo_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy', 'MPIF77': 'mpiifort', '_ModuleTable073_': 'LjAwMDAwMDAwMi4wMDAwMDAwMjguKnpmaW5hbCIsCn0sClsidXRpbC1saW51eC11dWlkIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvdXRpbC1saW51eC11dWlkLzIuMzguMS5sdWEiLApmdWxsTmFtZSA9ICJ1dGlsLWxpbnV4LXV1aWQvMi4zOC4xIiwKbG9hZE9yZGVyID0gMTQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAidXRpbC1saW51eC11dWlkLzIuMzguMSIsCndWID0gIjAw', '_ModuleTable010_': 'Mi4wIiwKbG9hZE9yZGVyID0gMjAsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZWNjb2Rlcy8yLjMyLjAiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDMyLip6ZmluYWwiLAp9LAplc21mID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAvZXNtZi84LjYuMC5sdWEiLApmdWxsTmFtZSA9ICJlc21mLzguNi4wIiwKbG9hZE9yZGVyID0gMzQsCnBy', 'NTHSTACK': 1024000000, 'SLURM_CPUS_PER_TASK': 1, 'FIXcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/fix', 'SLURM_NTASKS': 1, 'DO_ICE': True, 'NET': 'gfs', 'COM_ATMOS_GENESIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital', 'HOMEpost': '/work2/noaa/global/mterry/global-workflow_forked', '_ModuleTable061_': 'CnVzZXJOYW1lID0gInB5LXNldHVwdG9vbHMvNjMuNC4zIiwKd1YgPSAiMDAwMDAwMDYzLjAwMDAwMDAwNC4wMDAwMDAwMDMuKnpmaW5hbCIsCn0sClsicHktc2l4Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktc2l4LzEuMTYuMC5sdWEiLApmdWxsTmFtZSA9ICJweS1zaXgvMS4xNi4wIiwKbG9hZE9yZGVyID0gNzMsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktc2l4LzEu', 'bacio_ver': '2.4.1', 'cmake_ver': '3.23.1', 'UTILROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq', '_ModuleTable024_': 'ZXJOYW1lID0gImludGVsLW9uZWFwaS1ta2wvMjAyMy4xLjAiLAp3ViA9ICIwMDAwMDIwMjMuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbImludGVsLW9uZWFwaS1tcGkiXSA9IHsKZm4gPSAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L29uZWFwaS8yMDIzLjEuMC9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLmx1YSIsCmZ1bGxOYW1lID0gImludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAiLApsb2FkT3JkZXIgPSA0LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAiLAp3ViA9ICIwMDAwMDIwMjEuMDAwMDAwMDA5Lip6ZmluYWwi', 'tar_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths', 'ntasks': 1, 'max_tasks_per_node': 80, 'NCDUMP': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump', '_ModuleTable057_': 'LjIubHVhIiwKZnVsbE5hbWUgPSAicHktcHl0aG9uLWRhdGV1dGlsLzIuOC4yIiwKbG9hZE9yZGVyID0gNzQsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcHl0aG9uLWRhdGV1dGlsLzIuOC4yIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwOC4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sClsicHktcHl0eiJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXB5dHovMjAyMy4zLmx1YSIsCmZ1bGxOYW1lID0g', 'py_numpy_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne', 'QUILTING': True, 'IP_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_d.a', 'SLURM_TOPOLOGY_ADDR': 'hercules-08-38', 'libxcrypt_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr', 'CRTM_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include', 'LMOD_VERSION': '8.7.14', 'spack_stack_ver': '1.6.0', '_ModuleTable041_': 'LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9wYXJhbGxlbGlvLzIuNS4xMC5sdWEiLApmdWxsTmFtZSA9ICJwYXJhbGxlbGlvLzIuNS4xMCIsCmxvYWRPcmRlciA9IDMzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInBhcmFsbGVsaW8vMi41LjEwIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNS4wMDAwMDAwMTAuKnpmaW5hbCIsCn0sCnBpZ3ogPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0y', 'HOMEprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0', 'COMPONENT': 'ice', 'BASE_GIT': '/work2/noaa/global/role-global/git', 'FCST_SEGMENTS': [0, 120], 'libpng_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu', '_ModuleTable044_': 'LjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKcHJvaiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9wcm9qLzkuMi4xLmx1YSIsCmZ1bGxOYW1lID0gInByb2ovOS4yLjEiLApsb2FkT3JkZXIgPSAyNCwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcm9qLzkuMi4xIiwKd1YgPSAiMDAwMDAwMDA5LjAwMDAwMDAwMi4wMDAwMDAwMDEuKnpm', 'DO_GENESIS': True, 'BUFR_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_4', '__LMOD_REF_COUNT_DIAGUTIL_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/sys_check/sys_check.sh:1', 'COM_WAVE_GEMPAK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak', 'COM_ATMOS_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2', 'RUN': 'gfs', 'BUFR_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_8', 'COM_ICE_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon', 'py_pandas_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw', '__INTEL_POST_CFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64'], 'ARCDIR': '/work2/noaa/global/mterry/archive/C48_S2SW', '__LMOD_REF_COUNT_NLSPATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64/locale/%l_%t/%N:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin/locale/%l_%t/%N:1', '_ModuleTable069_': 'ZXIgPSAxMywKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMiwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzcWxpdGUvMy40My4yIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDA0My4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sClsic3RhY2staW50ZWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9Db3JlL3N0YWNrLWludGVsLzIwMjEuOS4wLmx1YSIsCmZ1bGxOYW1lID0gInN0YWNrLWludGVsLzIwMjEuOS4wIiwKbG9hZE9yZGVyID0gMywKcHJvcFQgPSB7fSwK', 'SENDECF': False, 'parallelio_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed', 'py_pytz_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s', 'ILPOST': 3, 'WGRIB2_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib/libwgrib2.a', 'parallel_netcdf_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3', 'CMPLR_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0', 'prod_util_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq', 'COMINukmet': '/work2/noaa/global/role-global/data/external_gempak/ukmet', 'libpng_ver': '1.6.37', 'COPYGB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/copygb2', 'NMV': '/bin/mv', 'NOSCRUB': '/work2/noaa/global/mterry', 'SCRATCH_hercules': '/scratch/hercules', 'W3EMC_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_d.a', 'SENDSDM': False, 'envir': 'prod', 'DO_METP': False, 'CRTM_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib/libcrtm.a', 'LOGSCRIPT': '', 'INTEL_ONEAPI_COMPILERS_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72', '_ModuleTable074_': 'MDAwMDAwMi4wMDAwMDAwMzguMDAwMDAwMDAxLip6ZmluYWwiLAp9LAp3M2VtYyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC93M2VtYy8yLjEwLjAubHVhIiwKZnVsbE5hbWUgPSAidzNlbWMvMi4xMC4wIiwKbG9hZE9yZGVyID0gMzksCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAidzNlbWMvMi4xMC4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKd2dyaWIy', 'py_jinja2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj', 'FPGA_VARS_ARGS': '', 'COMOUT_ICE_NETCDF': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ice/netcdf', 'INCVARS_ZERO_STRAT': ["'sphum_inc'", "'liq_wat_inc'", "'icmr_inc'", "'rwmr_inc'", "'snmr_inc'", "'grle_inc'"], 'COM_CONF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf', '_ModuleTable003_': 'YSIsCmZ1bGxOYW1lID0gImJhY2lvLzIuNC4xIiwKbG9hZE9yZGVyID0gMzgsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAzLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiYmFjaW8vMi40LjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDA0LjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKYnVmciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9idWZyLzExLjcuMC5sdWEiLApmdWxsTmFtZSA9ICJidWZyLzExLjcuMCIsCmxvYWRPcmRlciA9', 'HYDRA_BOOTSTRAP': 'slurm', '__LMOD_REF_COUNT___INTEL_POST_FFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1'], 'COM_CHEM_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon', 'MDATE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/mdate', 'WGRIB2': 'wgrib2', '_ModuleTable054_': 'LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktb3BlbnB5eGwvMy4xLjIubHVhIiwKZnVsbE5hbWUgPSAicHktb3BlbnB5eGwvMy4xLjIiLApsb2FkT3JkZXIgPSA2NiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1vcGVucHl4bC8zLjEuMiIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbInB5LXBhY2thZ2luZyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxs', 'EXP_WARM_START': False, 'MODULEPATH_ROOT': '/apps/other/modulefiles', 'COM_SNOW_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon', 'CHGRP_RSTPROD': True, 'py_pyxlsb_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge', 'BACIO_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/include_8', 'restart_interval_gfs': 12, 'DO_GOES': False, 'pgm': '', 'SLURM_TOPOLOGY_ADDR_PATTERN': 'node', 'DO_LAND_IAU': False, 'ntiles': 6, 'FHCYC': 24, 'ROTDIR': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW', 'FHOUT_AERO': 3, 'py_jinja2_ver': '3.1.2', 'BACIO_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/include_4', 'INCREMENTS_TO_ZERO': ["'liq_wat_inc'", "'icmr_inc'", "'rwmr_inc'", "'snmr_inc'", "'grle_inc'"], 'FHR_LIST': 72, 'USHprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/ush', 'BUFR_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_d', 'hdf5_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh', 'INTEL_ONEAPI_MKL_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci', 'SCRIPTSfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/scripts', '__LMOD_REF_COUNT_FI_PROVIDER_PATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib/prov:1;/usr/lib64/libfabric:1', 'FIXcice': '/work2/noaa/global/mterry/global-workflow_forked/fix/cice', 'IAUFHRS': [6], 'XDG_SESSION_CLASS': 'background', 'INTERVAL_GFS': 6, 'py_openpyxl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe', 'FHMAX_FITS': 120, 'COM_ATMOS_MASTER_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master', 'LMOD_PKG': '/apps/other/lmod/lmod', 'MPI_MEMMAP_OFF': 1, 'UUID_HERCULES_DTN': '869912fe-f6de-46c0-af10-b22efd84a022', '_ModuleTable079_': 'L3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAiCiwgIi9hcHBzL3NwYWNrLW1hbmFnZWQvbW9kdWxlZmlsZXMvbGludXgtcm9ja3k5LXg4Nl82NC9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLWE2NmVhaXAvb25lYXBpLzIwMjMuMS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wIgosICIvYXBwcy9zcGFjay1tYW5hZ2VkL21vZHVsZWZp', 'SDATE': datetime.datetime(2021, 3, 23, 12, 0), 'SLURM_SCRIPT_CONTEXT': 'prolog_task', 'MPI_CXX': 'mpiicpc', 'lwrite4danl': True, 'SLURM_MEM_PER_NODE': 98304, 'CASE_ENS': '{{ CASE_ENS }}', '_ModuleTable049_': 'aWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LWppbmphMi8zLjEuMi5sdWEiLApmdWxsTmFtZSA9ICJweS1qaW5qYTIvMy4xLjIiLApsb2FkT3JkZXIgPSA2MiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1qaW5qYTIvMy4xLjIiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDAxLjAwMDAwMDAwMi4qemZpbmFsIiwKfSwKWyJweS1tYXJrdXBzYWZlIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3Vs', 'YAML_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include', 'PYTHONPATH': '/apps/other/globus-cli-3.35.2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:/work2/noaa/global/mterry/global-workflow_forked/sorc/wxflow/src:/work2/noaa/global/mterry/global-workflow_forked/ush/python', '__LMOD_REF_COUNT_PYTHONPATH': '/apps/other/globus-cli-3.35.2:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:3;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:1', 'py_xarray_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4', 'nst_anl': True, 'FHOUT_GFS': 3, 'WORK': '/work/hercules', 'W3EMC_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_4.a', 'F77': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'BASE_DATA': '/work2/noaa/global/role-global/data', 'py_xlwt_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56', 'ACLOCAL_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/aclocal:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/aclocal:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/aclocal', 'QUEUE_SERVICE': 'batch', 'W3EMC_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_8.a', 'DATA': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ice_f072.357871', 'LESSOPEN': '||/usr/bin/lesspipe.sh %s', 'DO_VERFOZN': True, 'antlr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2', 'crtm_fix_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2', 'TOCGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib', 'USER': 'mterry', 'HPSS_PROJECT': 'emc-global', 'FHMAX_WAV_GFS': 120, 'FIXgdas': '/work2/noaa/global/mterry/global-workflow_forked/fix/gdas', 'launcher': 'srun -l --export=ALL --hint=nomultithread', 'NDATE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/ndate', 'LIBRARY_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib', 'gsl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z', 'SLURM_NODELIST': 'hercules-08-38', 'DO_FIT2OBS': True, 'IP_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_4', 'DOIAU_ENKF': True, 'COM_ATMOS_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos', 'py_xarray_ver': '2023.7.0', '__INTEL_POST_FFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64'], 'netcdf_diag': True, 'DOBNDPNT_WAVE': True, 'FETCHDIR': '/NCEPDEV/emc-global/1year/David.Grumm/test_data', 'ENVIRONMENT': 'BATCH', 'gsi_ncdiag_ver': '1.1.2', 'IP_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_8', 'py_bottleneck_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a', 'w3emc_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo', 'py_netcdf4_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i', 'wgrib2_ver': '2.0.8', 'TOCGRIB2SUPER': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib2super', 'COM_ICE_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2', 'KEEPDATA': False, 'udunits_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo', 'cdo_ver': '2.2.0', 'COM_OCEAN_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input', 'FNTSFA': ' ', 'LOADEDMODULES': 'contrib/0.1:intel-oneapi-compilers/2023.1.0:stack-intel/2021.9.0:intel-oneapi-mpi/2021.9.0:stack-intel-oneapi-mpi/2021.9.0:intel-oneapi-mkl/2023.1.0:zlib/1.2.13:pigz/2.7:zstd/1.5.2:tar/1.34:gettext/0.21.1:libxcrypt/4.4.35:sqlite/3.43.2:util-linux-uuid/2.38.1:python/3.11.6:libjpeg/2.1.0:jasper/2.0.32:libpng/1.6.37:openjpeg/2.3.1:eccodes/2.32.0:fftw/3.3.10:nghttp2/1.57.0:curl/8.4.0:proj/9.2.1:udunits/2.2.28:cdo/2.2.0:hdf5/1.14.0:snappy/1.1.10:c-blosc/1.21.5:netcdf-c/4.9.2:netcdf-fortran/4.6.1:parallel-netcdf/1.12.2:parallelio/2.5.10:esmf/8.6.0:antlr/2.7.7:gsl/2.7.1:nco/5.0.6:bacio/2.4.1:w3emc/2.10.0:prod_util/2.1.1:g2/3.4.5:sp/2.5.0:ip/4.3.0:grib-util/1.3.0:g2tmpl/1.13.0:gsi-ncdiag/1.1.2:crtm-fix/2.4.0.1_emc:git-lfs/3.1.2:crtm/2.4.0.1:openblas/0.3.24:py-setuptools/63.4.3:py-numpy/1.23.4:bufr/11.7.0:gmake/4.2.1:wgrib2/2.0.8:py-f90nml/1.4.3:py-cftime/1.0.3.4:py-netcdf4/1.5.8:libyaml/0.2.5:py-pyyaml/6.0:py-markupsafe/2.1.3:py-jinja2/3.1.2:py-bottleneck/1.3.7:py-numexpr/2.8.4:py-et-xmlfile/1.0.1:py-openpyxl/3.1.2:py-pytz/2023.3:py-pyxlsb/1.0.10:py-xlrd/2.0.1:py-xlsxwriter/3.1.7:py-xlwt/1.3.0:py-pandas/1.5.3:py-six/1.16.0:py-python-dateutil/2.8.2:g2c/1.8.0:netcdf-cxx4/4.3.1:met/9.1.3:metplus/3.1.1:py-packaging/23.1:py-xarray/2023.7.0:prepobs/1.1.0:fit2obs/1.1.7.1:globus-cli/3.35.2:module_base.hercules', 'SLURM_JOB_ACCOUNT': 'fv3-cpu', 'DO_OCN': True, 'SLURM_PRIO_PROCESS': 0, 'HOMEfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1', 'gmake_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq', 'FIXfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/fix', 'py_python_dateutil_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy', 'tasks_per_node': 1, 'SLURM_NPROCS': 1, 'COM_CHEM_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history', 'LMOD_ROOT': '/apps/other/lmod', 'DOHYBVAR': '{{ DOHYBVAR }}', 'GSL_ROOT_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z', 'SERIAL_F77': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'IP_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_d', 'SHLVL': 4, '_ModuleTable071_': 'e30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAic3RhY2staW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMCIsCndWID0gIjAwMDAwMjAyMS4wMDAwMDAwMDkuKnpmaW5hbCIsCn0sCnRhciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC90YXIvMS4zNC5sdWEiLApmdWxsTmFtZSA9ICJ0YXIvMS4zNCIsCmxvYWRPcmRlciA9IDEwLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIs', '_ModuleTable012_': 'c3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImZmdHcvMy4zLjEwIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAwMy4wMDAwMDAwMTAuKnpmaW5hbCIsCn0sCmZpdDJvYnMgPSB7CmZuID0gIi93b3JrMi9ub2FhL2dsb2JhbC9yb2xlLWdsb2JhbC9naXQvRml0Mk9icy92MS4xLjcuMS9tb2R1bGVmaWxlcy9maXQyb2JzLzEuMS43LjEubHVhIiwKZnVsbE5hbWUgPSAiZml0Mm9icy8xLjEuNy4xIiwKbG9hZE9yZGVyID0gODIsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZml0Mm9icy8xLjEuNy4xIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4wMDAwMDAwMDcuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApnMiA9', 'INCVARS_EFOLD': 5, 'SLURM_NNODES': 1, 'RESERVATION': '', 'FHMIN_GFS': 0, 'FHMAX_HF_WAV': 48, 'FIXugwd': '/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd', 'BASH_ENV': '/apps/other/lmod/lmod/init/bash', 'CDATE': datetime.datetime(2021, 3, 23, 12, 0), '_ModuleTable063_': 'aW5hbCIsCn0sClsicHkteGxyZCJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXhscmQvMi4wLjEubHVhIiwKZnVsbE5hbWUgPSAicHkteGxyZC8yLjAuMSIsCmxvYWRPcmRlciA9IDY5LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXhscmQvMi4wLjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJweS14bHN4d3JpdGVyIl0g', '_ModuleTable006_': 'Mi4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMi4qemZpbmFsIiwKfSwKY29udHJpYiA9IHsKZm4gPSAiL2FwcHMvb3RoZXIvbW9kdWxlZmlsZXMvY29udHJpYi8wLjEiLApmdWxsTmFtZSA9ICJjb250cmliLzAuMSIsCmxvYWRPcmRlciA9IDEsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAwLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY29udHJpYiIsCndWID0gIjAwMDAwMDAwMC4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sCmNydG0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50', 'FHOUT_OCN_GFS': 6, '_ModuleTable026_': 'cmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2phc3Blci8yLjAuMzIubHVhIiwKZnVsbE5hbWUgPSAiamFzcGVyLzIuMC4zMiIsCmxvYWRPcmRlciA9IDE3LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImphc3Blci8yLjAuMzIiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAzMi4qemZpbmFsIiwKfSwKbGlianBlZyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1z', 'DONST': True, '_ModuleTable051_': 'ay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL3B5LW5ldGNkZjQvMS41LjgubHVhIiwKZnVsbE5hbWUgPSAicHktbmV0Y2RmNC8xLjUuOCIsCmxvYWRPcmRlciA9IDU4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LW5ldGNkZjQvMS41LjgiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA1LjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKWyJweS1udW1leHByIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3Rh', 'LMOD_sys': 'Linux', '_ModuleTable033_': 'L3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL25jby81LjAuNi5sdWEiLApmdWxsTmFtZSA9ICJuY28vNS4wLjYiLApsb2FkT3JkZXIgPSAzNywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJuY28vNS4wLjYiLAp3ViA9ICIwMDAwMDAwMDUuMDAwMDAwMDAwLjAwMDAwMDAwNi4qemZpbmFsIiwKfSwKWyJuZXRjZGYtYyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1', 'FHMIN_WAV': 0, 'LMOD_FAMILY_METACOMPILER_VERSION': '2021.9.0', 'I_MPI_PMI_LIBRARY': '/opt/slurm/lib/libpmi2.so', 'machine': 'HERCULES', 'py_pyyaml_ver': 6.0, 'OCL_ICD_FILENAMES': 'libintelocl_emu.so:libalteracl.so:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64/libintelocl.so', '_ModuleTable017_': 'bG9hZE9yZGVyID0gNDgsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ2l0LWxmcy8zLjEuMiIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbImdsb2J1cy1jbGkiXSA9IHsKZm4gPSAiL2FwcHMvb3RoZXIvbW9kdWxlZmlsZXMvZ2xvYnVzLWNsaS8zLjM1LjIubHVhIiwKZnVsbE5hbWUgPSAiZ2xvYnVzLWNsaS8zLjM1LjIiLApsb2FkT3JkZXIgPSA4MywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJnbG9idXMtY2xpIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAzNS4wMDAwMDAwMDIu', 'DO_FETCH_HPSS': False, 'PROJ_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/proj', 'FHOUT_ICE': 3, 'COM_WAVE_STATION_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station', 'COMINecmwf': '/work2/noaa/global/role-global/data/external_gempak/ecmwf', 'COM_TOP_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}', 'WGRIB2_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include', 'mpmd_opt': '--multi-prog --output=mpmd.%j.%t.out', 'DO_PREP_SFC': False, 'LMOD_G2TMPLVIRT_NAME': 'g2tmpl', '_ModuleTable066_': 'bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5dGhvbi8zLjExLjYubHVhIiwKZnVsbE5hbWUgPSAicHl0aG9uLzMuMTEuNiIsCmxvYWRPcmRlciA9IDE1LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5dGhvbi8zLjExLjYiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDExLjAwMDAwMDAwNi4qemZpbmFsIiwKfSwKc25hcHB5ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAy', '_ModuleTable009_': 'MjEuOS4wL2N1cmwvOC40LjAubHVhIiwKZnVsbE5hbWUgPSAiY3VybC84LjQuMCIsCmxvYWRPcmRlciA9IDIzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImN1cmwvOC40LjAiLAp3ViA9ICIwMDAwMDAwMDguMDAwMDAwMDA0Lip6ZmluYWwiLAp9LAplY2NvZGVzID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2VjY29kZXMvMi4zMi4wLmx1YSIsCmZ1bGxOYW1lID0gImVjY29kZXMvMi4z', 'py_numexpr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq', '__LMOD_REF_COUNT_MANPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/share/man:2;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/man:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/documentation/en/man/common:1;/apps/other/lmod/lmod/share/man:1;/usr/share/man:1;/apps/share/man:1;/apps/man:1;/opt/slurm/share/man:1', 'DOLETKF_OCN': False, 'MPI_GROUP_MAX': 256, 'py_cftime_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i', 'XDG_SESSION_ID': 'c23', 'zstd_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4', 'CLUSTERS': '', '_ModuleTable034_': 'bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAvbmV0Y2RmLWMvNC45LjIubHVhIiwKZnVsbE5hbWUgPSAibmV0Y2RmLWMvNC45LjIiLApsb2FkT3JkZXIgPSAzMCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJuZXRjZGYtYy80LjkuMiIsCndWID0gIjAwMDAwMDAwNC4wMDAwMDAwMDkuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbIm5ldGNkZi1jeHg0Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2st', 'COM_OCEAN_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean', 'CLUSTERS_SERVICE': '', 'pgmout': 'OUTPUT.358493', 'EXECfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/exec', '_ModuleTable031_': 'aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJtZXQvOS4xLjMiLAp3ViA9ICIwMDAwMDAwMDkuMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKbWV0cGx1cyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL21ldHBsdXMvMy4xLjEubHVhIiwKZnVsbE5hbWUgPSAibWV0cGx1cy8zLjEuMSIsCmxvYWRPcmRlciA9IDc4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIs', 'COMOUT_ICE_GRIB': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ice/grib2', 'COM_ATMOS_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon', 'HOMEobsproc': '/work2/noaa/global/role-global/git/obsproc/v1.2.0', 'esmf_ver': '8.6.0', 'DO_ARCHCOM': False, 'SLURM_SUBMIT_HOST': 'hercules-login-1.hpc.msstate.edu', 'WRITE_DOPOST': True, 'EXECcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/exec', 'DUMP_SUFFIX': '', '_ModuleTable072_': 'CnVzZXJOYW1lID0gInRhci8xLjM0IiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAzNC4qemZpbmFsIiwKfSwKdWR1bml0cyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC91ZHVuaXRzLzIuMi4yOC5sdWEiLApmdWxsTmFtZSA9ICJ1ZHVuaXRzLzIuMi4yOCIsCmxvYWRPcmRlciA9IDI1LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInVkdW5pdHMvMi4yLjI4IiwKd1YgPSAiMDAwMDAwMDAy', 'CASE': 'C48', 'SENDDBN': False, 'gettext_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43', 'COM_ICE_BMATRIX_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice', 'git_lfs_ROOT': '/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l', '_ModuleTable001_': 'X01vZHVsZVRhYmxlXyA9IHsKTVR2ZXJzaW9uID0gMywKY19yZWJ1aWxkVGltZSA9IGZhbHNlLApjX3Nob3J0VGltZSA9IGZhbHNlLApkZXB0aFQgPSB7fSwKZmFtaWx5ID0gewpNZXRhQ29tcGlsZXIgPSAic3RhY2staW50ZWwiLApNZXRhTVBJID0gInN0YWNrLWludGVsLW9uZWFwaS1tcGkiLApjb21waWxlciA9ICJpbnRlbC1vbmVhcGktY29tcGlsZXJzIiwKZzJ0bXBsdmlydCA9ICJnMnRtcGwiLAptcGkgPSAiaW50ZWwtb25lYXBpLW1waSIsCn0sCm1UID0gewphbnRsciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9t', 'COM_ICE_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart', '_ModuleTable025_': 'LAp9LAppcCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9pcC80LjMuMC5sdWEiLApmdWxsTmFtZSA9ICJpcC80LjMuMCIsCmxvYWRPcmRlciA9IDQzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImlwLzQuMy4wIiwKd1YgPSAiMDAwMDAwMDA0LjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKamFzcGVyID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hl', 'COM_OCEAN_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart', 'ARCHCOM_TO': 'globus_hpss', 'COM_ATMOS_WMO_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo', 'ip_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf', 'CLASSPATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/mpi.jar', 'INTELFPGAOCLSDKROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga', 'LD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/host/linux64/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/lib:/usr/lib64:/usr/lib:/opt/slurm/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib', 'DO_CA': True, 'g2_ver': '3.4.5', 'LMOD_FAMILY_COMPILER': 'intel-oneapi-compilers', 'XDG_RUNTIME_DIR': '/run/user/9583', 'BASE_ENV': '/work2/noaa/global/mterry/global-workflow_forked/env', 'SLURM_JOB_ID': 5951734, '_ModuleTable014_': 'dWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nMmMvMS44LjAubHVhIiwKZnVsbE5hbWUgPSAiZzJjLzEuOC4wIiwKbG9hZE9yZGVyID0gNzUsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZzJjLzEuOC4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKZzJ0bXBsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9p', 'OCEANICEPRODUCTS_CONFIG': '/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml', 'NTHREADS_OCNICEPOST': 1, 'DO_JEDIOCNVAR': False, 'FIXreg2grb2': '/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2', 'SLURM_NODEID': 0, 'ip_ver': '4.3.0', 'KMP_AFFINITY': 'scatter', 'FHOUT_ICE_GFS': 6, '_ModuleTable065_': 'L3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS14bHd0LzEuMy4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXhsd3QvMS4zLjAiLApsb2FkT3JkZXIgPSA3MSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS14bHd0LzEuMy4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKcHl0aG9uID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1', 'DO_ATM': True, '_ModuleTable004_': 'IDUzLApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImJ1ZnIvMTEuNy4wIiwKd1YgPSAiMDAwMDAwMDExLjAwMDAwMDAwNy4qemZpbmFsIiwKfSwKWyJjLWJsb3NjIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvYy1ibG9zYy8xLjIxLjUubHVhIiwKZnVsbE5hbWUgPSAiYy1ibG9zYy8xLjIxLjUiLApsb2FkT3JkZXIgPSAyOSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1', 'esmf_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep', 'LMOD_FAMILY_MPI_VERSION': '2021.9.0', '_ModuleTable020_': 'c3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9nc2ktbmNkaWFnLzEuMS4yLmx1YSIsCmZ1bGxOYW1lID0gImdzaS1uY2RpYWcvMS4xLjIiLApsb2FkT3JkZXIgPSA0NiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJnc2ktbmNkaWFnLzEuMS4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCmdzbCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYu', 'DO_JEDIATMVAR': False, 'DO_AERO_ANL': False, 'NCLEN': '/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen', 'PS4': '+ $(basename ${BASH_SOURCE[0]:-${FUNCNAME[0]:-"Unknown"}})[${LINENO}]', 'COM_SNOW_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow', '_ModuleTable053_': 'YWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbnVtcHkvMS4yMy40Lmx1YSIsCmZ1bGxOYW1lID0gInB5LW51bXB5LzEuMjMuNCIsCmxvYWRPcmRlciA9IDUyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSA4LApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LW51bXB5LzEuMjMuNCIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMjMuMDAwMDAwMDA0Lip6ZmluYWwiLAp9LApbInB5LW9wZW5weXhsIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0', '__LMOD_REF_COUNT_ACL_BOARD_VENDOR_PATH': '/opt/Intel/OpenCLFPGA/oneAPI/Boards:1', 'LMOD_FAMILY_METAMPI_VERSION': '2021.9.0', '_ModuleTable059_': 'IDY4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXB5eGxzYi8xLjAuMTAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAwLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKWyJweS1weXlhbWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1weXlhbWwvNi4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXB5eWFtbC82LjAiLApsb2FkT3JkZXIgPSA2MCwKcHJvcFQgPSB7fSwKc3RhY2tE', 'DO_CALC_INCREMENT_ENKF_GFS': False, 'MKLROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0', 'DEBUGINFOD_URLS': 'https://debuginfod.centos.org/ ', 'sven_root_path': '/home/gfekete/sven', '_ModuleTable047_': 'LApbInB5LWV0LXhtbGZpbGUiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1ldC14bWxmaWxlLzEuMC4xLmx1YSIsCmZ1bGxOYW1lID0gInB5LWV0LXhtbGZpbGUvMS4wLjEiLApsb2FkT3JkZXIgPSA2NSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1ldC14bWxmaWxlLzEuMC4xIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMC4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsi', 'restart_interval_enkfgfs': 3, 'SENDDBN_NTC': False, 'FIXaer': '/work2/noaa/global/mterry/global-workflow_forked/fix/aer', 'FHMAX_HF_GFS': 48, 'COM_ICE_LETKF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf', 'LMOD_FAMILY_MPI': 'intel-oneapi-mpi', 'netcdf_fortran_ver': '4.6.1', 'FIXmom': '/work2/noaa/global/mterry/global-workflow_forked/fix/mom6', 'cyc': 12, 'BACIO_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib/libbacio_4.a', 'FC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'FHOUT_WAV_GFS': 3, 'SDATE_GFS': datetime.datetime(2021, 3, 23, 12, 0), 'BACIO_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib/libbacio_8.a', 'sp_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr', 'OPS_RES': 'C768', 'NMEM_ENS_GFS': 30, 'which_declare': 'declare -f', '_ModuleTable056_': 'OS4wL3B5LXBhbmRhcy8xLjUuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1wYW5kYXMvMS41LjMiLApsb2FkT3JkZXIgPSA3MiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1wYW5kYXMvMS41LjMiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA1LjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKWyJweS1weXRob24tZGF0ZXV0aWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1weXRob24tZGF0ZXV0aWwvMi44', 'COM_WAVE_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded', 'LMOD_FAMILY_G2TMPLVIRT': 'g2tmpl', 'QUEUE_DTN': 'batch', 'DO_GENESIS_FSU': False, 'prod_util_ver': '2.1.1', 'MPI_F90': 'mpiifort', 'LMOD_FAMILY_COMPILER_VERSION': '2023.1.0', '_ModuleTable042_': 'MDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcGlnei8yLjcubHVhIiwKZnVsbE5hbWUgPSAicGlnei8yLjciLApsb2FkT3JkZXIgPSA4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gNCwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInBpZ3ovMi43IiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4qemZpbmFsIiwKfSwKcHJlcG9icyA9IHsKZm4gPSAiL3dvcmsyL25vYWEvZ2xvYmFsL3JvbGUtZ2xvYmFsL2dpdC9wcmVwb2JzL3YxLjEuMC9tb2R1bGVmaWxlcy9wcmVwb2JzLzEuMS4wLmx1YSIsCmZ1bGxOYW1lID0gInByZXBvYnMvMS4xLjAiLApsb2FkT3JkZXIgPSA4MSwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0', 'openjpeg_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu', 'USHfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/ush', 'YAML_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib', 'DO_GSISOILDA': False, 'nemsio_ver': '2.5.4', 'DO_NPOESS': False, 'SLURM_MPI_TYPE': 'pmi2', 'OMP_STACKSIZE': 2048000, 'py_et_xmlfile_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq', 'FORECAST_HOUR': 72, 'ATARDIR': '/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW', 'EXECgfs': '/work2/noaa/global/mterry/global-workflow_forked/exec', 'py_python_dateutil_ver': '2.8.2', 'assim_freq': 6, 'DOIAU': True, 'FHOUT_WAV': 1, 'py_pyyaml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2', 'met_ver': '9.1.3', 'gsi_ncdiag_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe', 'DO_FETCH_LOCAL': False, 'NLSPATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64/locale/%l_%t/%N:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin/locale/%l_%t/%N', 'netcdf_fortran_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe', 'SP_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_4.a', 'COM_OCEAN_BMATRIX_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean', 'SLURM_CONF': '/var/spool/slurmd/conf-cache/slurm.conf', 'PATH': '/apps/other/globus-cli-3.35.2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r/ush:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/bin:/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/bin/intel64:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/bin:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/bin:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin:/usr/sbin:/usr/bin:/apps/sbin:/apps/bin:/opt/slurm/bin:/home/gfekete/sven/bin', 'SP_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_8.a', 'netcdf_c_ver': '4.9.2', 'SLURM_JOB_NAME': 'C48_S2SW_gfs_ice_prod_f072_12', 'COM_ICE_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history', 'NLN': '/bin/ln -sf', 'MODULEPATH': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles:/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0-a66eaip/g2tmpl/1.13.0/intel/2021.9.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/g2tmpl/1.13.0/intel/2021.9.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/intel-oneapi-mpi/2021.9.0-a66eaip/oneapi/2023.1.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core:/work2/noaa/global/mterry/global-workflow_forked/modulefiles:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core:/apps/other/modulefiles:/apps/containers/modulefiles:/apps/licensed/modulefiles:/apps/contrib/modulefiles', 'CC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icc', 'grib_util_ver': '1.3.0', 'py_f90nml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z', 'DO_MERGENSST': False, 'SLURM_NTASKS_PER_NODE': 1, '_LMFILES_': '/apps/other/modulefiles/contrib/0.1:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core/intel-oneapi-compilers/2023.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core/stack-intel/2021.9.0.lua:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0/intel-oneapi-mpi/2021.9.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/stack-intel-oneapi-mpi/2021.9.0.lua:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core/intel-oneapi-mkl/2023.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/zlib/1.2.13.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/pigz/2.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/zstd/1.5.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/tar/1.34.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gettext/0.21.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libxcrypt/4.4.35.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/sqlite/3.43.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/util-linux-uuid/2.38.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/python/3.11.6.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libjpeg/2.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/jasper/2.0.32.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libpng/1.6.37.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/openjpeg/2.3.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/eccodes/2.32.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/fftw/3.3.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/nghttp2/1.57.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/curl/8.4.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/proj/9.2.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/udunits/2.2.28.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/cdo/2.2.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/hdf5/1.14.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/snappy/1.1.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/c-blosc/1.21.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-c/4.9.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-fortran/4.6.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/parallel-netcdf/1.12.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/parallelio/2.5.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/esmf/8.6.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/antlr/2.7.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gsl/2.7.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/nco/5.0.6.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/bacio/2.4.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/w3emc/2.10.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/prod_util/2.1.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2/3.4.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/sp/2.5.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/ip/4.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/grib-util/1.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2tmpl/1.13.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/gsi-ncdiag/1.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/crtm-fix/2.4.0.1_emc.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/git-lfs/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/crtm/2.4.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/openblas/0.3.24.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-setuptools/63.4.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-numpy/1.23.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/bufr/11.7.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gmake/4.2.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/wgrib2/2.0.8.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-f90nml/1.4.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-cftime/1.0.3.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/py-netcdf4/1.5.8.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libyaml/0.2.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pyyaml/6.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-markupsafe/2.1.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-jinja2/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-bottleneck/1.3.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-numexpr/2.8.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-et-xmlfile/1.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-openpyxl/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pytz/2023.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pyxlsb/1.0.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlrd/2.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlsxwriter/3.1.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlwt/1.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pandas/1.5.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-six/1.16.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-python-dateutil/2.8.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2c/1.8.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-cxx4/4.3.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/met/9.1.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/metplus/3.1.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-packaging/23.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xarray/2023.7.0.lua:/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles/prepobs/1.1.0.lua:/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles/fit2obs/1.1.7.1.lua:/apps/other/modulefiles/globus-cli/3.35.2.lua:/work2/noaa/global/mterry/global-workflow_forked/modulefiles/module_base.hercules.lua', 'globus_cli_ver': 3.27, '_ModuleTable028_': 'L2xpYnBuZy8xLjYuMzcubHVhIiwKZnVsbE5hbWUgPSAibGlicG5nLzEuNi4zNyIsCmxvYWRPcmRlciA9IDE4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImxpYnBuZy8xLjYuMzciLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA2LjAwMDAwMDAzNy4qemZpbmFsIiwKfSwKbGlieGNyeXB0ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2xpYnhjcnlwdC80LjQuMzUubHVhIiwKZnVsbE5hbWUgPSAibGlieGNyeXB0', 'py_numpy_ver': '1.23.4', '__LMOD_REF_COUNT_HDF5_PLUGIN_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/plugins:1', 'bufr_ver': '11.7.0', 'COMINnam': '/work2/noaa/global/role-global/data/external_gempak/nam', 'DBUS_SESSION_BUS_ADDRESS': 'unix:path=/run/user/9583/bus', 'DO_AERO_FCST': False, 'py_pandas_ver': '1.5.3', '_ModuleTable039_': 'dGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9vcGVuanBlZy8yLjMuMS5sdWEiLApmdWxsTmFtZSA9ICJvcGVuanBlZy8yLjMuMSIsCmxvYWRPcmRlciA9IDE5LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm9wZW5qcGVnLzIuMy4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsicGFyYWxsZWwtbmV0Y2RmIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMv', 'COM_ICE_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice', 'UUID_ORION_DTN': '8a10dd4f-24ee-4794-a39d-9c313ab6a34b', 'MPI_F77': 'mpiifort', 'LMOD_CMD': '/apps/other/lmod/lmod/libexec/lmod', '_ModuleTable045_': 'aW5hbCIsCn0sClsicHktYm90dGxlbmVjayJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LWJvdHRsZW5lY2svMS4zLjcubHVhIiwKZnVsbE5hbWUgPSAicHktYm90dGxlbmVjay8xLjMuNyIsCmxvYWRPcmRlciA9IDYzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWJvdHRsZW5lY2svMS4zLjciLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAzLjAwMDAwMDAwNy4qemZpbmFs', 'LMOD_MPI_VERSION': '2021.9.0-a66eaip', 'LMOD_SYSTEM_NAME': 'hercules', 'RUN_ENVIR': 'emc', 'w3emc_ver': '2.10.0', '__LMOD_REF_COUNT_LD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/host/linux64/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/lib:1;/usr/lib64:1;/usr/lib:1;/opt/slurm/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:2', 'DOHYBVAR_OCN': False, 'NUM_SND_COLLECTIVES': 9, 'walltime': '00:15:00', 'netcdf_cxx4_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu', 'ens_tracker_ver': 'v1.2.0', 'LMOD_SYSTEM_DEFAULT_MODULES': 'contrib', 'OMPI_MCA_plm_slurm_args': '--external-launcher', '_ModuleTable078_': 'cyIKLCAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLWE2NmVhaXAvZzJ0bXBsLzEuMTMuMC9pbnRlbC8yMDIxLjkuMCIKLCAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9nMnRtcGwvMS4xMy4wL2ludGVsLzIwMjEuOS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVz', 'FIXcpl': '/work2/noaa/global/mterry/global-workflow_forked/fix/cpl', 'G2C_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64/libg2c.so', '_ModuleTable019_': 'cmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9ncmliLXV0aWwvMS4zLjAubHVhIiwKZnVsbE5hbWUgPSAiZ3JpYi11dGlsLzEuMy4wIiwKbG9hZE9yZGVyID0gNDQsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ3JpYi11dGlsLzEuMy4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKWyJnc2ktbmNkaWFnIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMv', 'SLURM_JOB_GID': 17000, 'SLURM_GET_USER_ENV': 1, 'IAUFHRS_ENKF': [3, 6, 9], 'CPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/include:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/include:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/include', 'G2TMPL_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include', 'DO_JEDISNOWDA': False, 'SP_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_d.a', 'pigz_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq', 'COM_ATMOS_BUFR_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr', 'COM_OCEAN_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2', 'CNVGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/cnvgrib', 'COM_ATMOS_TRACK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks', 'DO_VERFRAD': True, 'pgmerr': 'errfile', 'MPI_BUFS_PER_HOST': 2048, 'DO_PREP_OBS_AERO': False, 'LMOD_FAMILY_G2TMPLVIRT_VERSION': '1.13.0', 'PNG_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu', 'COM_ATMOS_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart', 'DEGRIB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/degrib2', 'DO_ANLSTAT': False, 'LMOD_MPI_NAME': 'intel-oneapi-mpi', 'py_f90nml_ver': '1.4.3', 'NMEM_ENS_GFS_OFFSET': 20, 'SMOOTH_ENKF': False, 'FHMAX_GOES': 120, 'OLDPWD': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', 'QUEUE': 'batch', 'COM_OBS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/obs', 'DO_VMINMON': True, 'metplus_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r', 'VERBOSE': True, 'SLURM_JOB_NODELIST': 'hercules-08-38', 'CLUSTER': 'hercules', 'DBNROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn', 'FHOUT_HF_WAV': 1, 'PARMgfs': '/work2/noaa/global/mterry/global-workflow_forked/parm', 'I_MPI_HYDRA_BOOTSTRAP': 'slurm', 'APRUN_OCNICEPOST': 'srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1', 'BASH_FUNC_ml%%': '() { eval "$($LMOD_DIR/ml_cmd "$@")"\n}', 'BASH_FUNC_which%%': '() { ( alias;\n eval ${which_declare} ) | /usr/bin/which --tty-only --read-alias --read-functions --show-tilde --show-dot $@\n}', 'BASH_FUNC_module%%': '() { if [ -z "${LMOD_SH_DBG_ON+x}" ]; then\n case "$-" in \n *v*x*)\n __lmod_sh_dbg=\'vx\'\n ;;\n *v*)\n __lmod_sh_dbg=\'v\'\n ;;\n *x*)\n __lmod_sh_dbg=\'x\'\n ;;\n esac;\n fi;\n if [ -n "${__lmod_sh_dbg:-}" ]; then\n set +$__lmod_sh_dbg;\n echo "Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for Lmod\'s output" 1>&2;\n fi;\n eval "$($LMOD_CMD $LMOD_SHELL_PRGM "$@")" && eval "$(${LMOD_SETTARG_CMD:-:} -s sh)";\n __lmod_my_status=$?;\n if [ -n "${__lmod_sh_dbg:-}" ]; then\n echo "Shell debugging restarted" 1>&2;\n set -$__lmod_sh_dbg;\n fi;\n unset __lmod_sh_dbg;\n return $__lmod_my_status\n}', 'BASH_FUNC_declare_from_tmpl%%': '() { if [[ ${DEBUG_WORKFLOW:-"NO"} == "NO" ]]; then\n set +x;\n fi;\n local opts="-g";\n local OPTIND=1;\n while getopts "rx" option; do\n opts="${opts}${option}";\n done;\n shift $((OPTIND-1));\n for input in "$@";\n do\n IFS=\':\' read -ra args <<< "${input}";\n local com_var="${args[0]}";\n local template;\n local value;\n if (( ${#args[@]} > 1 )); then\n template="${args[1]}";\n else\n template="${com_var}_TMPL";\n fi;\n if [[ ! -v "${template}" ]]; then\n echo "FATAL ERROR in declare_from_tmpl: Requested template ${template} not defined!";\n exit 2;\n fi;\n value=$(echo "${!template}" | envsubst);\n declare ${opts} "${com_var}"="${value}";\n echo "declare_from_tmpl :: ${com_var}=${value}";\n done;\n set_trace\n}', 'BASH_FUNC_err_exit%%': ['() { set +eux;\n msg1=${*:-Job ${jobid} failed};\n if [[ -n "${pgm}" ]]; then\n msg1+="', 'ERROR IN ${pgm}";\n fi;\n if [[ -n "${err}" ]]; then\n msg1+=" RETURN CODE ${err}";\n fi;\n msg2="\n -------------------------------------------------------------\n -- FATAL ERROR: ${msg1}\n -- ABNORMAL EXIT at $(date) on ${HOSTNAME}\n -------------------------------------------------------------\n ";\n echo "${msg2}" 1>&2;\n module list;\n echo "" 1>&2;\n echo "${msg1}" 1>&2;\n if [[ -n "${DATA}" ]]; then\n echo "${DATA}" 1>&2;\n ls -ltr "${DATA}" 1>&2;\n else\n echo "WARNING: DATA variable not defined" 1>&2;\n fi;\n if [[ -n "${pgmout}" ]]; then\n if [[ -s errfile ]]; then\n echo "----- contents of errfile -----" >> "${pgmout}";\n cat errfile >> "${pgmout}";\n fi;\n cat "${pgmout}" 1>&2;\n else\n if [[ -s errfile ]]; then\n cat errfile 1>&2;\n fi;\n fi;\n if [[ "${SENDECF}" == "YES" ]]; then\n timeout 30 ecflow_client --msg "${ECF_NAME}: ${msg1}";\n timeout 30 ssh "${ECF_HOST}" "echo \\"${msg}2\\" >> ${ECF_JOBOUT:?}";\n fi;\n if [[ "${SENDECF}" == "YES" ]]; then\n ecflow_client --kill="${ECF_NAME:?}";\n fi;\n if [[ -n "${PBS_JOBID}" ]]; then\n qdel "${PBS_JOBID}";\n else\n if [[ -n "${SLURM_JOB_ID}" ]]; then\n scancel "${SLURM_JOB_ID}";\n fi;\n fi\n}'], 'BASH_FUNC_wait_for_file%%': '() { set +x;\n local file_name=${1:?"wait_for_file() requires a file name"};\n local sleep_interval=${2:-60};\n local max_tries=${3:-100};\n for ((iter=0; iter' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.oceanice_products -+++ config.oceanice_products[5]echo 'BEGIN: config.oceanice_products' -BEGIN: config.oceanice_products -+++ config.oceanice_products[8]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources oceanice_products -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=oceanice_products -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[978]walltime=00:15:00 -++++ config.resources[979]ntasks=1 -++++ config.resources[980]tasks_per_node=1 -++++ config.resources[981]threads_per_task=1 -++++ config.resources[982]memory=96GB -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n '' ]] -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export memory -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.oceanice_products[11]export MAX_TASKS=25 -+++ config.oceanice_products[11]MAX_TASKS=25 -+++ config.oceanice_products[13]export OCEANICEPRODUCTS_CONFIG=/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml -+++ config.oceanice_products[13]OCEANICEPRODUCTS_CONFIG=/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml -+++ config.oceanice_products[16]export NFHRS_PER_GROUP=3 -+++ config.oceanice_products[16]NFHRS_PER_GROUP=3 -+++ config.oceanice_products[18]echo 'END: config.oceanice_products' -END: config.oceanice_products -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env oceanice_products -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=oceanice_products -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 100663296 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 1 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 1 ]] -+++ HERCULES.env[34]max_threads_per_task=80 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 80 ]] -+++ HERCULES.env[40][[ 1 -gt 80 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 1' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[286]export NTHREADS_OCNICEPOST=1 -+++ HERCULES.env[286]NTHREADS_OCNICEPOST=1 -+++ HERCULES.env[287]export 'APRUN_OCNICEPOST=srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1' -+++ HERCULES.env[287]APRUN_OCNICEPOST='srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1' -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_OCEANICE_PRODUCTS[11]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[11]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[11]declare_from_tmpl -rx COMIN_ICE_HISTORY:COM_ICE_HISTORY_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_ICE_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/ice/history -+ JGLOBAL_OCEANICE_PRODUCTS[12]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[12]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[12]declare_from_tmpl -rx COMOUT_ICE_GRIB:COM_ICE_GRIB_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_ICE_GRIB=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ice/grib2 -+ JGLOBAL_OCEANICE_PRODUCTS[13]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[13]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[13]declare_from_tmpl -rx COMOUT_ICE_NETCDF:COM_ICE_NETCDF_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_ICE_NETCDF=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ice/netcdf -+ JGLOBAL_OCEANICE_PRODUCTS[17]/work2/noaa/global/mterry/global-workflow_forked/scripts/exglobal_oceanice_products.py -2025-07-28 21:34:54,440 - INFO - root : BEGIN: __main__.main -2025-07-28 21:34:54,440 - DEBUG - root : ( ) -2025-07-28 21:34:54,448 - INFO - oceanice_products: BEGIN: OceanIceProducts.__init__ -2025-07-28 21:34:54,448 - DEBUG - oceanice_products: ( , {'SHELL': '/bin/bash', 'PDY': datetime.datetime(2021, 3, 23, 0, 0), 'DO_JEDIATMENS': False, 'sqlite_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4', 'COPYGB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/copygb', 'WORK_hercules': '/work/hercules', 'TOCGRIB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib2', 'G2_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/include_d', 'PIO_TYPENAME_VALID_VALUES': ['netcdf', 'netcdf4p', 'netcdf4c', 'pnetcdf'], 'WGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/wgrib', 'SERIAL_CXX': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icpc', 'FHMAX_WAV': 9, 'FIXlut': '/work2/noaa/global/mterry/global-workflow_forked/fix/lut', 'fftw_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj', 'snappy_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx', '_ModuleTable016_': 'bGxOYW1lID0gImdldHRleHQvMC4yMS4xIiwKbG9hZE9yZGVyID0gMTEsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ2V0dGV4dC8wLjIxLjEiLAp3ViA9ICIwMDAwMDAwMDAuMDAwMDAwMDIxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJnaXQtbGZzIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvZ2l0LWxmcy8zLjEuMi5sdWEiLApmdWxsTmFtZSA9ICJnaXQtbGZzLzMuMS4yIiwK', 'MAKE_NSSTBUFR': False, 'cdo_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e', 'curl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops', 'PIO_VERSION_MAJOR': 2, 'threads_per_task': 1, 'SLURM_JOB_USER': 'mterry', '_ModuleTable067_': 'NC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3NuYXBweS8xLjEuMTAubHVhIiwKZnVsbE5hbWUgPSAic25hcHB5LzEuMS4xMCIsCmxvYWRPcmRlciA9IDI4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInNuYXBweS8xLjEuMTAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAxLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKc3AgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45', 'COM_ICE_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}', 'COM_OCEAN_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}', '_ModuleTable022_': 'ZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9oZGY1LzEuMTQuMC5sdWEiLApmdWxsTmFtZSA9ICJoZGY1LzEuMTQuMCIsCmxvYWRPcmRlciA9IDI3LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImhkZjUvMS4xNC4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAxNC4qemZpbmFsIiwKfSwKWyJpbnRlbC1vbmVhcGktY29tcGlsZXJzIl0gPSB7CmZuID0gIi9hcHBzL3NwYWNrLW1hbmFnZWQvbW9kdWxlZmlsZXMvbGludXgtcm9ja3k5LXg4Nl82NC9Db3JlL2ludGVsLW9uZWFwaS1jb21waWxlcnMvMjAyMy4xLjAubHVhIiwKZnVsbE5hbWUgPSAiaW50ZWwtb25lYXBpLWNvbXBpbGVy', 'SLURM_TASKS_PER_NODE': 1, 'COM_ATMOS_OZNMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon', 'COM_ATMOS_GEMPAK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}', 'SLURM_JOB_UID': 9583, 'HISTCONTROL': 'ignoredups', 'SLURM_EXPORT_ENV': 'NONE', 'g2c_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp', 'COM_OCEAN_NETCDF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf', 'GRB2INDEX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grb2index', 'l4densvar': True, 'FHOUT_HF_GFS': 1, 'I_MPI_EXTRA_FILESYSTEM': 1, 'SLURM_TASK_PID': 357856, 'EXECprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/exec', 'g2tmpl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel', 'MPI_BUFS_PER_PROC': 2048, 'IAU_DELTHRS_ENKF': 6, 'COM_WAVE_WMO_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo', '_ModuleTable027_': 'LTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9saWJqcGVnLzIuMS4wLmx1YSIsCmZ1bGxOYW1lID0gImxpYmpwZWcvMi4xLjAiLApsb2FkT3JkZXIgPSAxNiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJsaWJqcGVnLzIuMS4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKbGlicG5nID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4w', 'NMEM_ENS': 0, '__LMOD_REF_COUNT_CLASSPATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/mpi.jar:1', '_ModuleTable050_': 'ZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbWFya3Vwc2FmZS8yLjEuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1tYXJrdXBzYWZlLzIuMS4zIiwKbG9hZE9yZGVyID0gNjEsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktbWFya3Vwc2FmZS8yLjEuMyIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDEuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LW5ldGNkZjQiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFj', 'err': 0, '_ModuleTable032_': 'CnVzZXJOYW1lID0gIm1ldHBsdXMvMy4xLjEiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJtb2R1bGVfYmFzZS5oZXJjdWxlcyJdID0gewpmbiA9ICIvd29yazIvbm9hYS9nbG9iYWwvbXRlcnJ5L2dsb2JhbC13b3JrZmxvd19mb3JrZWQvbW9kdWxlZmlsZXMvbW9kdWxlX2Jhc2UuaGVyY3VsZXMubHVhIiwKZnVsbE5hbWUgPSAibW9kdWxlX2Jhc2UuaGVyY3VsZXMiLApsb2FkT3JkZXIgPSA4NCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDAsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJtb2R1bGVfYmFzZS5oZXJjdWxlcyIsCndWID0gIk0uKnpmaW5hbCIsCn0sCm5jbyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGlj', 'CHGRP_CMD': 'chgrp rstprod', 'PKG_CONFIG_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib/pkgconfig:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/pkgconfig:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/lib/pkgconfig', 'SLURM_LOCALID': 0, 'MAX_TASKS': 25, '_ModuleTable040_': 'aW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9wYXJhbGxlbC1uZXRjZGYvMS4xMi4yLmx1YSIsCmZ1bGxOYW1lID0gInBhcmFsbGVsLW5ldGNkZi8xLjEyLjIiLApsb2FkT3JkZXIgPSAzMiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwYXJhbGxlbC1uZXRjZGYvMS4xMi4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAxMi4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCnBhcmFsbGVsaW8gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0', 'SLURM_SUBMIT_DIR': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', '_ModuleTable077_': 'LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvenN0ZC8xLjUuMi5sdWEiLApmdWxsTmFtZSA9ICJ6c3RkLzEuNS4yIiwKbG9hZE9yZGVyID0gOSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMywKc3RhY2tEZXB0aCA9IDQsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJ6c3RkLzEuNS4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwNS4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCn0sCm1wYXRoQSA9IHsKIi93b3JrMi9ub2FhL2dsb2JhbC9yb2xlLWdsb2JhbC9naXQvRml0Mk9icy92MS4xLjcuMS9tb2R1bGVmaWxlcyIKLCAiL3dvcmsyL25vYWEvZ2xvYmFsL3JvbGUtZ2xvYmFsL2dpdC9wcmVwb2JzL3YxLjEuMC9tb2R1bGVmaWxl', 'HOSTNAME': 'hercules-08-38', 'HISTSIZE': 1000, 'wgrib2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln', 'FRAC_GRID': True, 'G2_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/include_4', 'intel_oneapi_mpi_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse', 'SLURMD_NODENAME': 'hercules-08-38', '_ModuleTable080_': 'bGVzL2xpbnV4LXJvY2t5OS14ODZfNjQvb25lYXBpLzIwMjMuMS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL0NvcmUiCiwgIi93b3JrMi9ub2FhL2dsb2JhbC9tdGVycnkvZ2xvYmFsLXdvcmtmbG93X2ZvcmtlZC9tb2R1bGVmaWxlcyIKLCAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L0NvcmUiCiwgIi9hcHBzL290aGVyL21vZHVsZWZpbGVzIiwgIi9hcHBzL2NvbnRhaW5lcnMvbW9kdWxlZmlsZXMiCiwgIi9hcHBzL2xpY2Vuc2VkL21vZHVsZWZpbGVzIiwg', '__LMOD_REF_COUNT_OCL_ICD_FILENAMES': 'libintelocl_emu.so:1;libalteracl.so:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64/libintelocl.so:1', 'c_blosc_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn', 'FHOUT_OCN': 3, 'COM_OBS_JEDI': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi', 'eccodes_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj', '_ModuleTable037_': 'ay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9uZ2h0dHAyLzEuNTcuMC5sdWEiLApmdWxsTmFtZSA9ICJuZ2h0dHAyLzEuNTcuMCIsCmxvYWRPcmRlciA9IDIyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gNCwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm5naHR0cDIvMS41Ny4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDA1Ny4qemZpbmFsIiwKfSwKb3BlbmJsYXMgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2kt', 'netcdf_c_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy', '_ModuleTable070_': 'c3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzdGFjay1pbnRlbC8yMDIxLjkuMCIsCndWID0gIjAwMDAwMjAyMS4wMDAwMDAwMDkuKnpmaW5hbCIsCn0sClsic3RhY2staW50ZWwtb25lYXBpLW1waSJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3N0YWNrLWludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAubHVhIiwKZnVsbE5hbWUgPSAic3RhY2staW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMCIsCmxvYWRPcmRlciA9IDUsCnByb3BUID0g', 'SLURM_JOB_START_TIME': 1753756458, '_ModuleTable011_': 'b3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZXNtZi84LjYuMCIsCndWID0gIjAwMDAwMDAwOC4wMDAwMDAwMDYuKnpmaW5hbCIsCn0sCmZmdHcgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9mZnR3LzMuMy4xMC5sdWEiLApmdWxsTmFtZSA9ICJmZnR3LzMuMy4xMCIsCmxvYWRPcmRlciA9IDIxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwK', 'DO_NEST': False, 'STMP': '/work2/noaa/stmp/mterry/HERCULES', 'HYDRA_LAUNCHER_EXTRA_ARGS': '--external-launcher', 'ACL_BOARD_VENDOR_PATH': '/opt/Intel/OpenCLFPGA/oneAPI/Boards', '_ModuleTable060_': 'ZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1weXlhbWwvNi4wIiwKd1YgPSAiMDAwMDAwMDA2Lip6ZmluYWwiLAp9LApbInB5LXNldHVwdG9vbHMiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1zZXR1cHRvb2xzLzYzLjQuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1zZXR1cHRvb2xzLzYzLjQuMyIsCmxvYWRPcmRlciA9IDUxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIs', '_ModuleTable007_': 'ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9jcnRtLzIuNC4wLjEubHVhIiwKZnVsbE5hbWUgPSAiY3J0bS8yLjQuMC4xIiwKbG9hZE9yZGVyID0gNDksCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY3J0bS8yLjQuMC4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNC4wMDAwMDAwMDAuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbImNydG0tZml4Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45', 'G2TMPL_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib/libg2tmpl.a', 'DO_WAVE': True, 'SCRgfs': '/work2/noaa/global/mterry/global-workflow_forked/scripts', 'SP_INC8': 'include_8', '_ModuleTable002_': 'b2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9hbnRsci8yLjcuNy5sdWEiLApmdWxsTmFtZSA9ICJhbnRsci8yLjcuNyIsCmxvYWRPcmRlciA9IDM1LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImFudGxyLzIuNy43IiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4wMDAwMDAwMDcuKnpmaW5hbCIsCn0sCmJhY2lvID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2JhY2lvLzIuNC4xLmx1', 'MPICC': 'mpiicc', 'HDF5_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh', '_ModuleTable055_': 'L21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXBhY2thZ2luZy8yMy4xLmx1YSIsCmZ1bGxOYW1lID0gInB5LXBhY2thZ2luZy8yMy4xIiwKbG9hZE9yZGVyID0gNzksCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcGFja2FnaW5nLzIzLjEiLAp3ViA9ICIwMDAwMDAwMjMuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbInB5LXBhbmRhcyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEu', 'SP_INC4': 'include_4', 'FPGA_VARS_DIR': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga', 'FHOUT': 3, '__LMOD_REF_COUNT_MODULEPATH': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles:1;/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0-a66eaip/g2tmpl/1.13.0/intel/2021.9.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/g2tmpl/1.13.0/intel/2021.9.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/intel-oneapi-mpi/2021.9.0-a66eaip/oneapi/2023.1.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core:1;/work2/noaa/global/mterry/global-workflow_forked/modulefiles:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core:1;/apps/other/modulefiles:1;/apps/containers/modulefiles:1;/apps/licensed/modulefiles:1;/apps/contrib/modulefiles:1', 'cycle': 't12z', '_ModuleTable043_': 'aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcmVwb2JzLzEuMS4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKcHJvZF91dGlsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3Byb2RfdXRpbC8yLjEuMS5sdWEiLApmdWxsTmFtZSA9ICJwcm9kX3V0aWwvMi4xLjEiLApsb2FkT3JkZXIgPSA0MCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcm9kX3V0aWwvMi4x', 'MPI_CC': 'mpiicc', 'PARTITION_DTN': '', 'sp_ver': '2.5.0', 'SHELLOPTS': 'braceexpand:errexit:hashall:interactive-comments:nounset:xtrace', 'DO_AWIPS': False, 'EDATE': datetime.datetime(2021, 3, 23, 12, 0), '__LMOD_REF_COUNT___INTEL_POST_CFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1'], 'I_MPI_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0', 'CRTM_FIX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2/fix', '_ModuleTable062_': 'MTYuMCIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMTYuKnpmaW5hbCIsCn0sClsicHkteGFycmF5Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHkteGFycmF5LzIwMjMuNy4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXhhcnJheS8yMDIzLjcuMCIsCmxvYWRPcmRlciA9IDgwLApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXhhcnJheS8yMDIzLjcuMCIsCndWID0gIjAwMDAwMjAyMy4wMDAwMDAwMDcuKnpm', 'proj_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez', 'FSYNC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file', '_ModuleTable005_': 'cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJjLWJsb3NjLzEuMjEuNSIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMjEuMDAwMDAwMDA1Lip6ZmluYWwiLAp9LApjZG8gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9jZG8vMi4yLjAubHVhIiwKZnVsbE5hbWUgPSAiY2RvLzIuMi4wIiwKbG9hZE9yZGVyID0gMjYsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY2RvLzIu', 'g2tmpl_ver': '1.13.0', 'FHOUT_GOES': 3, '_ModuleTable021_': 'MC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nc2wvMi43LjEubHVhIiwKZnVsbE5hbWUgPSAiZ3NsLzIuNy4xIiwKbG9hZE9yZGVyID0gMzYsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ3NsLzIuNy4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sCmhkZjUgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxl', 'EUPD_CYC': 'gdas', 'fms_ver': '2023.02.01', 'util_linux_uuid_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj', '_ModuleTable052_': 'Y2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbnVtZXhwci8yLjguNC5sdWEiLApmdWxsTmFtZSA9ICJweS1udW1leHByLzIuOC40IiwKbG9hZE9yZGVyID0gNjQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktbnVtZXhwci8yLjguNCIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDguMDAwMDAwMDA0Lip6ZmluYWwiLAp9LApbInB5LW51bXB5Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2kt', '_ModuleTable030_': 'cFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJsaWJ5YW1sLzAuMi41IiwKd1YgPSAiMDAwMDAwMDAwLjAwMDAwMDAwMi4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCm1ldCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL21ldC85LjEuMy5sdWEiLApmdWxsTmFtZSA9ICJtZXQvOS4xLjMiLApsb2FkT3JkZXIgPSA3NywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0', 'COM_ATMOS_IMAGERY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery', 'I_MPI_EXTRA_FILESYSTEM_LIST': 'lustre', 'COM_WAVE_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history', 'SLURM_CLUSTER_NAME': 'hercules', '__LMOD_REF_COUNT_ACLOCAL_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/aclocal:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/aclocal:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/aclocal:2', 'SERIAL_CC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icc', 'COM_OBSPROC_TMPL': '${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos', 'SLURM_JOB_END_TIME': 1753757358, '_ModuleTable075_': 'ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3dncmliMi8yLjAuOC5sdWEiLApmdWxsTmFtZSA9ICJ3Z3JpYjIvMi4wLjgiLApsb2FkT3JkZXIgPSA1NSwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJ3Z3JpYjIvMi4wLjgiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKemxpYiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJj', 'FHMAX_GFS': 120, 'SENDAWIP': False, 'G2_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64/libg2_d.a', 'spack_mod_path': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core', 'OCNRES': 500, 'COM_MED_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart', 'SLURM_CPUS_ON_NODE': 1, 'LMOD_FAMILY_METAMPI': 'stack-intel-oneapi-mpi', 'LEVS': 128, 'FIXgsi': '/work2/noaa/global/mterry/global-workflow_forked/fix/gsi', 'NFHRS_PER_GROUP': 3, 'COM_WAVE_PREP_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep', 'FI_PROVIDER_PATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib/prov:/usr/lib64/libfabric', 'python_ver': '3.11.6', 'IAU_OFFSET': 0, 'py_packaging_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6', 'DIAGUTIL_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/sys_check/sys_check.sh', 'python_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2', 'SCRIPTScfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/scripts', 'py_markupsafe_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7', 'G2C_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include', 'py_xlrd_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f', 'LMOD_G2TMPLVIRT_VERSION': '1.13.0', 'pid': 358494, 'OUTPUT_GRID': 'gaussian_grid', 'jobid': 'oceanice_products_ice_f078.357885', 'SLURM_JOB_CPUS_PER_NODE': 1, 'spack_env': 'gsi-addon-dev-fms-2024.01', 'INTEL_ONEAPI_MPI_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse', 'LMOD_FAMILY_METACOMPILER': 'stack-intel', 'SP_INCd': 'include_d', 'py_six_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o', 'COM_ATMOS_GOES_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim', 'MPICXX': 'mpiicpc', 'restart_interval_gdas': 3, 'fit2obs_ver': '1.1.7.1', 'LMOD_DIR': '/apps/other/lmod/lmod/libexec', '_ModuleTable068_': 'LjAvc3AvMi41LjAubHVhIiwKZnVsbE5hbWUgPSAic3AvMi41LjAiLApsb2FkT3JkZXIgPSA0MiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMiwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzcC8yLjUuMCIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCnNxbGl0ZSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9zcWxpdGUvMy40My4yLmx1YSIsCmZ1bGxOYW1lID0gInNxbGl0ZS8zLjQzLjIiLApsb2FkT3Jk', 'REPLAY_ICS': False, 'COM_ICE_NETCDF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf', 'G2_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64/libg2_4.a', '_ModuleTable058_': 'InB5LXB5dHovMjAyMy4zIiwKbG9hZE9yZGVyID0gNjcsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcHl0ei8yMDIzLjMiLAp3ViA9ICIwMDAwMDIwMjMuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LXB5eGxzYiJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXB5eGxzYi8xLjAuMTAubHVhIiwKZnVsbE5hbWUgPSAicHktcHl4bHNiLzEuMC4xMCIsCmxvYWRPcmRlciA9', 'py_setuptools_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr', 'FHMIN': 0, 'COM_ICE_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input', '_ModuleTable048_': 'cHktZjkwbm1sIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktZjkwbm1sLzEuNC4zLmx1YSIsCmZ1bGxOYW1lID0gInB5LWY5MG5tbC8xLjQuMyIsCmxvYWRPcmRlciA9IDU2LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWY5MG5tbC8xLjQuMyIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMDQuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LWppbmphMiJdID0gewpmbiA9ICIvd29yay9ub2FhL2Vw', 'intel_mkl_ver': '2023.1.0', 'PRTE_MCA_plm_slurm_args': '--external-launcher', 'PWD': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ice_f078.357885', 'SLURM_GTIDS': 0, 'LOGNAME': 'mterry', 'MAKE_ACFTBUFR': False, 'W3EMC_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_8', 'HOMEDIR': '/work2/noaa/global/mterry', 'IAU_DELTHRS': 6, 'W3EMC_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_4', 'XDG_SESSION_TYPE': 'unspecified', 'ESMFMKFILE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib/esmf.mk', 'FHMAX': 9, 'WRITE_NSFLIP': True, 'SLURM_JOB_PARTITION': 'hercules', 'YAML_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x', 'PARTITION_SERVICE': 'service', 'MODULESHOME': '/apps/other/lmod/lmod', '__LMOD_REF_COUNT_DYLD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:1', 'COM_WAVE_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart', 'BUFR_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_4.so', 'BUFR_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_8.so', 'HOMEcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1', 'COM_OCEAN_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon', 'DEBUG_POSTSCRIPT': False, 'MANPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/share/man:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/man:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/documentation/en/man/common:/apps/other/lmod/lmod/share/man:/usr/share/man:/apps/share/man:/apps/man:/opt/slurm/share/man::', 'libyaml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x', 'jasper_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt', 'py_xlsxwriter_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok', 'SLURM_TRES_PER_TASK': 'cpu=1', 'nco_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq', 'stack_intel_ver': '2021.9.0', 'SLURM_OOM_KILL_STEP': 0, 'BUFR_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m', 'bufr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m', 'CLUSTERS_DTN': '', 'grib_util_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35', 'nghttp2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky', 'DO_TEST_MODE': True, 'PACKAGEROOT': '/work2/noaa/global/role-global/nwpara', 'lobsdiag_forenkf': True, 'SLURM_JOB_NUM_NODES': 1, 'py_netcdf4_ver': '1.5.8', 'ICERES': 500, 'CXX': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icpc', 'GRIB2GRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grib2grib2', 'DMPDIR': '/work/noaa/rstprod/dump', 'LSOIL_INCR': 2, 'metplus_ver': '3.1.1', '_ModuleTable008_': 'LjAvY3J0bS1maXgvMi40LjAuMV9lbWMubHVhIiwKZnVsbE5hbWUgPSAiY3J0bS1maXgvMi40LjAuMV9lbWMiLApsb2FkT3JkZXIgPSA0NywKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJjcnRtLWZpeC8yLjQuMC4xX2VtYyIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDQuMDAwMDAwMDAwLjAwMDAwMDAwMS4qXy4qZW1jLip6ZmluYWwiLAp9LApjdXJsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIw', 'COM_OCEAN_LETKF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf', 'zlib_ver': '1.2.13', 'BUFR_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_d.so', 'ENKF_SPREAD': True, 'zlib_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j', '__LMOD_REF_COUNT_CMAKE_PREFIX_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r:2;/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/IntelDPCPP:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72:1', '_ModuleTable035_': 'c3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9uZXRjZGYtY3h4NC80LjMuMS5sdWEiLApmdWxsTmFtZSA9ICJuZXRjZGYtY3h4NC80LjMuMSIsCmxvYWRPcmRlciA9IDc2LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm5ldGNkZi1jeHg0LzQuMy4xIiwKd1YgPSAiMDAwMDAwMDA0LjAwMDAwMDAwMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsibmV0Y2RmLWZvcnRyYW4iXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFj', 'FIXgfs': '/work2/noaa/global/mterry/global-workflow_forked/fix', 'DO_COUPLED': True, 'SLURM_JOBID': 5951735, 'SERIAL_FC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'COM_ATMOS_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input', 'W3EMC_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_d', '_ModuleTable013_': 'IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nMi8zLjQuNS5sdWEiLApmdWxsTmFtZSA9ICJnMi8zLjQuNSIsCmxvYWRPcmRlciA9IDQxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImcyLzMuNC41IiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAwNC4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCmcyYyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJj', 'COM_ATMOS_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history', 'FIXam': '/work2/noaa/global/mterry/global-workflow_forked/fix/am', 'hdf5_ver': '1.14.0', 'restart_interval_enkfgdas': 3, 'COM_ATMOS_RADMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon', 'DO_CALC_INCREMENT': False, 'DO_STARTMEM_FROM_JEDIICE': False, 'FIXprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/fix', 'SLURM_JOB_QOS': 'batch', 'I_MPI_HYDRA_BOOTSTRAP_EXEC_EXTRA_ARGS': '--external-launcher', 'obsproc_run_ver': '1.2.0', 'COM_ATMOS_MINMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon', 'crtm_ver': '2.4.0.1', 'FIXorog': '/work2/noaa/global/mterry/global-workflow_forked/fix/orog', 'COM_OCEAN_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history', '__LMOD_REF_COUNT_PATH': '/apps/other/globus-cli-3.35.2/bin:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r/ush:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/bin:2;/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/bin:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/bin/intel64:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/bin:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/bin:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/bin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/bin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin:1;/usr/sbin:1;/usr/bin:1;/apps/sbin:1;/apps/bin:1;/opt/slurm/bin:1;/home/gfekete/sven/bin:1', '_ModuleTable018_': 'KnpmaW5hbCIsCn0sCmdtYWtlID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2dtYWtlLzQuMi4xLmx1YSIsCmZ1bGxOYW1lID0gImdtYWtlLzQuMi4xIiwKbG9hZE9yZGVyID0gNTQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ21ha2UvNC4yLjEiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDAyLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJncmliLXV0aWwiXSA9IHsKZm4gPSAiL3dv', 'COMIN_ICE_HISTORY': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/ice/history', 'COMROOT': '/work2/noaa/global/mterry/RUNTESTS/COMROOT', 'HOME': '/home/mterry', 'FI_PROVIDER': 'mlx', 'met_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b', 'COM_RTOFS_TMPL': '${DMPDIR}', '_ModuleTable015_': 'bnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2cydG1wbC8xLjEzLjAubHVhIiwKZnVsbE5hbWUgPSAiZzJ0bXBsLzEuMTMuMCIsCmxvYWRPcmRlciA9IDQ1LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImcydG1wbC8xLjEzLjAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDEzLip6ZmluYWwiLAp9LApnZXR0ZXh0ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2dldHRleHQvMC4yMS4xLmx1YSIsCmZ1', '_ModuleTable_Sz_': 81, 'LANG': 'C.UTF-8', '__LMOD_REF_COUNT_LIBRARY_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:1', 'DO_BUFRSND': False, 'memory': '96GB', '_ModuleTable064_': 'PSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHkteGxzeHdyaXRlci8zLjEuNy5sdWEiLApmdWxsTmFtZSA9ICJweS14bHN4d3JpdGVyLzMuMS43IiwKbG9hZE9yZGVyID0gNzAsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHkteGxzeHdyaXRlci8zLjEuNyIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDA3Lip6ZmluYWwiLAp9LApbInB5LXhsd3QiXSA9IHsKZm4gPSAi', 'DYLD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib', 'PARTITION_BATCH': 'hercules', '_ModuleTable023_': 'cy8yMDIzLjEuMCIsCmxvYWRPcmRlciA9IDIsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiaW50ZWwtb25lYXBpLWNvbXBpbGVycy8yMDIzLjEuMCIsCndWID0gIjAwMDAwMjAyMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsiaW50ZWwtb25lYXBpLW1rbCJdID0gewpmbiA9ICIvYXBwcy9zcGFjay1tYW5hZ2VkL21vZHVsZWZpbGVzL2xpbnV4LXJvY2t5OS14ODZfNjQvQ29yZS9pbnRlbC1vbmVhcGktbWtsLzIwMjMuMS4wLmx1YSIsCmZ1bGxOYW1lID0gImludGVsLW9uZWFwaS1ta2wvMjAyMy4xLjAiLApsb2FkT3JkZXIgPSA2LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVz', '_ModuleTable081_': 'Ii9hcHBzL2NvbnRyaWIvbW9kdWxlZmlsZXMiLAp9LApzeXN0ZW1CYXNlTVBBVEggPSAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L0NvcmU6L2FwcHMvb3RoZXIvbW9kdWxlZmlsZXM6L2FwcHMvY29udGFpbmVycy9tb2R1bGVmaWxlczovYXBwcy9saWNlbnNlZC9tb2R1bGVmaWxlcyIsCn0K', 'REDERR': '2>', 'PSLOT': 'C48_S2SW', '__LMOD_REF_COUNT_PKG_CONFIG_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib/pkgconfig:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/pkgconfig:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib/pkgconfig:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/pkgconfig:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib/pkgconfig:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/lib/pkgconfig:1', '_ModuleTable036_': 'ay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL25ldGNkZi1mb3J0cmFuLzQuNi4xLmx1YSIsCmZ1bGxOYW1lID0gIm5ldGNkZi1mb3J0cmFuLzQuNi4xIiwKbG9hZE9yZGVyID0gMzEsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAibmV0Y2RmLWZvcnRyYW4vNC42LjEiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDA2LjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKbmdodHRwMiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFj', 'COM_ATMOS_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}', 'USHcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/ush', 'DO_GEMPAK': False, 'SLURM_PROCID': 0, 'USHgfs': '/work2/noaa/global/mterry/global-workflow_forked/ush', 'DOIBP_WAV': False, 'prepobs_run_ver': '1.1.0', 'EXPDIR': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', 'job': 'oceanice_products', 'imp_physics': 8, 'LMOD_SHELL_PRGM': 'bash', 'IAU_FHROT': 0, 'MPIF90': 'mpiifort', 'IP_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_4.a', 'waveGRD': 'uglo_100km', 'COM_CHEM_BMAT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix', 'jasper_ver': '2.0.32', 'SCRATCH': '/scratch/hercules', 'IP_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_8.a', 'CLIENT_GLOBUS_UUID': '869912fe-f6de-46c0-af10-b22efd84a022', 'ACCOUNT': 'fv3-cpu', 'HOMEgfs': '/work2/noaa/global/mterry/global-workflow_forked', 'NHOUR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/nhour', 'binary_diag': False, 'COM_WAVE_GRID_RES_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}', 'SCRIPTSprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/scripts', 'MP_STDOUTMODE': 'ORDERED', 'sigio_ver': '2.3.2', '__LMOD_REF_COUNT_CPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/include:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/include:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/include:1', 'LMOD_SETTARG_FULL_SUPPORT': False, 'COMINsyn': '/work2/noaa/global/role-global/com/gfs/prod/syndat', 'OFFSET_START_HOUR': 0, 'REDOUT': '1>', 'PTMP': '/work2/noaa/stmp/mterry/HERCULES', '_ModuleTable076_': 'dWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC96bGliLzEuMi4xMy5sdWEiLApmdWxsTmFtZSA9ICJ6bGliLzEuMi4xMyIsCmxvYWRPcmRlciA9IDcsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEwLApzdGFja0RlcHRoID0gNSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInpsaWIvMS4yLjEzIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMi4wMDAwMDAwMTMuKnpmaW5hbCIsCn0sCnpzdGQgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2', 'g2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato', 'MODE': 'forecast-only', 'DATAROOT': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312', 'bacio_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj', 'APP': 'S2SW', '_ModuleTable029_': 'LzQuNC4zNSIsCmxvYWRPcmRlciA9IDEyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImxpYnhjcnlwdC80LjQuMzUiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDA0LjAwMDAwMDAzNS4qemZpbmFsIiwKfSwKbGlieWFtbCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9saWJ5YW1sLzAuMi41Lmx1YSIsCmZ1bGxOYW1lID0gImxpYnlhbWwvMC4yLjUiLApsb2FkT3JkZXIgPSA1OSwKcHJv', 'openblas_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f', 'stack_impi_ver': '2021.9.0', 'COM_CHEM_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem', 'TMPDIR': '/local/scratch/mterry/5951735', 'HDF5_PLUGIN_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/plugins', 'DO_TRACKER': True, 'CMAKE_PREFIX_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r:/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/IntelDPCPP:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72', 'crtm_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r', '_ModuleTable038_': 'YWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvb3BlbmJsYXMvMC4zLjI0Lmx1YSIsCmZ1bGxOYW1lID0gIm9wZW5ibGFzLzAuMy4yNCIsCmxvYWRPcmRlciA9IDUwLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm9wZW5ibGFzLzAuMy4yNCIsCndWID0gIjAwMDAwMDAwMC4wMDAwMDAwMDMuMDAwMDAwMDI0Lip6ZmluYWwiLAp9LApvcGVuanBlZyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5z', 'nco_ver': '5.0.6', '_ModuleTable046_': 'IiwKfSwKWyJweS1jZnRpbWUiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1jZnRpbWUvMS4wLjMuNC5sdWEiLApmdWxsTmFtZSA9ICJweS1jZnRpbWUvMS4wLjMuNCIsCmxvYWRPcmRlciA9IDU3LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWNmdGltZS8xLjAuMy40IiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMC4wMDAwMDAwMDMuMDAwMDAwMDA0Lip6ZmluYWwiLAp9', 'GRBINDEX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grbindex', 'libjpeg_turbo_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy', 'MPIF77': 'mpiifort', '_ModuleTable073_': 'LjAwMDAwMDAwMi4wMDAwMDAwMjguKnpmaW5hbCIsCn0sClsidXRpbC1saW51eC11dWlkIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvdXRpbC1saW51eC11dWlkLzIuMzguMS5sdWEiLApmdWxsTmFtZSA9ICJ1dGlsLWxpbnV4LXV1aWQvMi4zOC4xIiwKbG9hZE9yZGVyID0gMTQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAidXRpbC1saW51eC11dWlkLzIuMzguMSIsCndWID0gIjAw', '_ModuleTable010_': 'Mi4wIiwKbG9hZE9yZGVyID0gMjAsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZWNjb2Rlcy8yLjMyLjAiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDMyLip6ZmluYWwiLAp9LAplc21mID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAvZXNtZi84LjYuMC5sdWEiLApmdWxsTmFtZSA9ICJlc21mLzguNi4wIiwKbG9hZE9yZGVyID0gMzQsCnBy', 'NTHSTACK': 1024000000, 'SLURM_CPUS_PER_TASK': 1, 'FIXcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/fix', 'SLURM_NTASKS': 1, 'DO_ICE': True, 'NET': 'gfs', 'COM_ATMOS_GENESIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital', 'HOMEpost': '/work2/noaa/global/mterry/global-workflow_forked', '_ModuleTable061_': 'CnVzZXJOYW1lID0gInB5LXNldHVwdG9vbHMvNjMuNC4zIiwKd1YgPSAiMDAwMDAwMDYzLjAwMDAwMDAwNC4wMDAwMDAwMDMuKnpmaW5hbCIsCn0sClsicHktc2l4Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktc2l4LzEuMTYuMC5sdWEiLApmdWxsTmFtZSA9ICJweS1zaXgvMS4xNi4wIiwKbG9hZE9yZGVyID0gNzMsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktc2l4LzEu', 'bacio_ver': '2.4.1', 'cmake_ver': '3.23.1', 'UTILROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq', '_ModuleTable024_': 'ZXJOYW1lID0gImludGVsLW9uZWFwaS1ta2wvMjAyMy4xLjAiLAp3ViA9ICIwMDAwMDIwMjMuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbImludGVsLW9uZWFwaS1tcGkiXSA9IHsKZm4gPSAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L29uZWFwaS8yMDIzLjEuMC9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLmx1YSIsCmZ1bGxOYW1lID0gImludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAiLApsb2FkT3JkZXIgPSA0LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAiLAp3ViA9ICIwMDAwMDIwMjEuMDAwMDAwMDA5Lip6ZmluYWwi', 'tar_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths', 'ntasks': 1, 'max_tasks_per_node': 80, 'NCDUMP': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump', '_ModuleTable057_': 'LjIubHVhIiwKZnVsbE5hbWUgPSAicHktcHl0aG9uLWRhdGV1dGlsLzIuOC4yIiwKbG9hZE9yZGVyID0gNzQsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcHl0aG9uLWRhdGV1dGlsLzIuOC4yIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwOC4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sClsicHktcHl0eiJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXB5dHovMjAyMy4zLmx1YSIsCmZ1bGxOYW1lID0g', 'py_numpy_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne', 'QUILTING': True, 'IP_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_d.a', 'SLURM_TOPOLOGY_ADDR': 'hercules-08-38', 'libxcrypt_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr', 'CRTM_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include', 'LMOD_VERSION': '8.7.14', 'spack_stack_ver': '1.6.0', '_ModuleTable041_': 'LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9wYXJhbGxlbGlvLzIuNS4xMC5sdWEiLApmdWxsTmFtZSA9ICJwYXJhbGxlbGlvLzIuNS4xMCIsCmxvYWRPcmRlciA9IDMzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInBhcmFsbGVsaW8vMi41LjEwIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNS4wMDAwMDAwMTAuKnpmaW5hbCIsCn0sCnBpZ3ogPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0y', 'HOMEprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0', 'COMPONENT': 'ice', 'BASE_GIT': '/work2/noaa/global/role-global/git', 'FCST_SEGMENTS': [0, 120], 'libpng_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu', '_ModuleTable044_': 'LjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKcHJvaiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9wcm9qLzkuMi4xLmx1YSIsCmZ1bGxOYW1lID0gInByb2ovOS4yLjEiLApsb2FkT3JkZXIgPSAyNCwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcm9qLzkuMi4xIiwKd1YgPSAiMDAwMDAwMDA5LjAwMDAwMDAwMi4wMDAwMDAwMDEuKnpm', 'DO_GENESIS': True, 'BUFR_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_4', '__LMOD_REF_COUNT_DIAGUTIL_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/sys_check/sys_check.sh:1', 'COM_WAVE_GEMPAK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak', 'COM_ATMOS_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2', 'RUN': 'gfs', 'BUFR_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_8', 'COM_ICE_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon', 'py_pandas_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw', '__INTEL_POST_CFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64'], 'ARCDIR': '/work2/noaa/global/mterry/archive/C48_S2SW', '__LMOD_REF_COUNT_NLSPATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64/locale/%l_%t/%N:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin/locale/%l_%t/%N:1', '_ModuleTable069_': 'ZXIgPSAxMywKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMiwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzcWxpdGUvMy40My4yIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDA0My4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sClsic3RhY2staW50ZWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9Db3JlL3N0YWNrLWludGVsLzIwMjEuOS4wLmx1YSIsCmZ1bGxOYW1lID0gInN0YWNrLWludGVsLzIwMjEuOS4wIiwKbG9hZE9yZGVyID0gMywKcHJvcFQgPSB7fSwK', 'SENDECF': False, 'parallelio_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed', 'py_pytz_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s', 'ILPOST': 3, 'WGRIB2_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib/libwgrib2.a', 'parallel_netcdf_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3', 'CMPLR_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0', 'prod_util_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq', 'COMINukmet': '/work2/noaa/global/role-global/data/external_gempak/ukmet', 'libpng_ver': '1.6.37', 'COPYGB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/copygb2', 'NMV': '/bin/mv', 'NOSCRUB': '/work2/noaa/global/mterry', 'SCRATCH_hercules': '/scratch/hercules', 'W3EMC_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_d.a', 'SENDSDM': False, 'envir': 'prod', 'DO_METP': False, 'CRTM_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib/libcrtm.a', 'LOGSCRIPT': '', 'INTEL_ONEAPI_COMPILERS_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72', '_ModuleTable074_': 'MDAwMDAwMi4wMDAwMDAwMzguMDAwMDAwMDAxLip6ZmluYWwiLAp9LAp3M2VtYyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC93M2VtYy8yLjEwLjAubHVhIiwKZnVsbE5hbWUgPSAidzNlbWMvMi4xMC4wIiwKbG9hZE9yZGVyID0gMzksCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAidzNlbWMvMi4xMC4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKd2dyaWIy', 'py_jinja2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj', 'FPGA_VARS_ARGS': '', 'COMOUT_ICE_NETCDF': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ice/netcdf', 'INCVARS_ZERO_STRAT': ["'sphum_inc'", "'liq_wat_inc'", "'icmr_inc'", "'rwmr_inc'", "'snmr_inc'", "'grle_inc'"], 'COM_CONF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf', '_ModuleTable003_': 'YSIsCmZ1bGxOYW1lID0gImJhY2lvLzIuNC4xIiwKbG9hZE9yZGVyID0gMzgsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAzLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiYmFjaW8vMi40LjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDA0LjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKYnVmciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9idWZyLzExLjcuMC5sdWEiLApmdWxsTmFtZSA9ICJidWZyLzExLjcuMCIsCmxvYWRPcmRlciA9', 'HYDRA_BOOTSTRAP': 'slurm', '__LMOD_REF_COUNT___INTEL_POST_FFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1'], 'COM_CHEM_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon', 'MDATE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/mdate', 'WGRIB2': 'wgrib2', '_ModuleTable054_': 'LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktb3BlbnB5eGwvMy4xLjIubHVhIiwKZnVsbE5hbWUgPSAicHktb3BlbnB5eGwvMy4xLjIiLApsb2FkT3JkZXIgPSA2NiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1vcGVucHl4bC8zLjEuMiIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbInB5LXBhY2thZ2luZyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxs', 'EXP_WARM_START': False, 'MODULEPATH_ROOT': '/apps/other/modulefiles', 'COM_SNOW_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon', 'CHGRP_RSTPROD': True, 'py_pyxlsb_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge', 'BACIO_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/include_8', 'restart_interval_gfs': 12, 'DO_GOES': False, 'pgm': '', 'SLURM_TOPOLOGY_ADDR_PATTERN': 'node', 'DO_LAND_IAU': False, 'ntiles': 6, 'FHCYC': 24, 'ROTDIR': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW', 'FHOUT_AERO': 3, 'py_jinja2_ver': '3.1.2', 'BACIO_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/include_4', 'INCREMENTS_TO_ZERO': ["'liq_wat_inc'", "'icmr_inc'", "'rwmr_inc'", "'snmr_inc'", "'grle_inc'"], 'FHR_LIST': 78, 'USHprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/ush', 'BUFR_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_d', 'hdf5_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh', 'INTEL_ONEAPI_MKL_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci', 'SCRIPTSfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/scripts', '__LMOD_REF_COUNT_FI_PROVIDER_PATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib/prov:1;/usr/lib64/libfabric:1', 'FIXcice': '/work2/noaa/global/mterry/global-workflow_forked/fix/cice', 'IAUFHRS': [6], 'XDG_SESSION_CLASS': 'background', 'INTERVAL_GFS': 6, 'py_openpyxl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe', 'FHMAX_FITS': 120, 'COM_ATMOS_MASTER_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master', 'LMOD_PKG': '/apps/other/lmod/lmod', 'MPI_MEMMAP_OFF': 1, 'UUID_HERCULES_DTN': '869912fe-f6de-46c0-af10-b22efd84a022', '_ModuleTable079_': 'L3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAiCiwgIi9hcHBzL3NwYWNrLW1hbmFnZWQvbW9kdWxlZmlsZXMvbGludXgtcm9ja3k5LXg4Nl82NC9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLWE2NmVhaXAvb25lYXBpLzIwMjMuMS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wIgosICIvYXBwcy9zcGFjay1tYW5hZ2VkL21vZHVsZWZp', 'SDATE': datetime.datetime(2021, 3, 23, 12, 0), 'SLURM_SCRIPT_CONTEXT': 'prolog_task', 'MPI_CXX': 'mpiicpc', 'lwrite4danl': True, 'SLURM_MEM_PER_NODE': 98304, 'CASE_ENS': '{{ CASE_ENS }}', '_ModuleTable049_': 'aWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LWppbmphMi8zLjEuMi5sdWEiLApmdWxsTmFtZSA9ICJweS1qaW5qYTIvMy4xLjIiLApsb2FkT3JkZXIgPSA2MiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1qaW5qYTIvMy4xLjIiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDAxLjAwMDAwMDAwMi4qemZpbmFsIiwKfSwKWyJweS1tYXJrdXBzYWZlIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3Vs', 'YAML_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include', 'PYTHONPATH': '/apps/other/globus-cli-3.35.2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:/work2/noaa/global/mterry/global-workflow_forked/sorc/wxflow/src:/work2/noaa/global/mterry/global-workflow_forked/ush/python', '__LMOD_REF_COUNT_PYTHONPATH': '/apps/other/globus-cli-3.35.2:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:3;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:1', 'py_xarray_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4', 'nst_anl': True, 'FHOUT_GFS': 3, 'WORK': '/work/hercules', 'W3EMC_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_4.a', 'F77': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'BASE_DATA': '/work2/noaa/global/role-global/data', 'py_xlwt_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56', 'ACLOCAL_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/aclocal:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/aclocal:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/aclocal', 'QUEUE_SERVICE': 'batch', 'W3EMC_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_8.a', 'DATA': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ice_f078.357885', 'LESSOPEN': '||/usr/bin/lesspipe.sh %s', 'DO_VERFOZN': True, 'antlr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2', 'crtm_fix_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2', 'TOCGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib', 'USER': 'mterry', 'HPSS_PROJECT': 'emc-global', 'FHMAX_WAV_GFS': 120, 'FIXgdas': '/work2/noaa/global/mterry/global-workflow_forked/fix/gdas', 'launcher': 'srun -l --export=ALL --hint=nomultithread', 'NDATE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/ndate', 'LIBRARY_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib', 'gsl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z', 'SLURM_NODELIST': 'hercules-08-38', 'DO_FIT2OBS': True, 'IP_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_4', 'DOIAU_ENKF': True, 'COM_ATMOS_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos', 'py_xarray_ver': '2023.7.0', '__INTEL_POST_FFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64'], 'netcdf_diag': True, 'DOBNDPNT_WAVE': True, 'FETCHDIR': '/NCEPDEV/emc-global/1year/David.Grumm/test_data', 'ENVIRONMENT': 'BATCH', 'gsi_ncdiag_ver': '1.1.2', 'IP_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_8', 'py_bottleneck_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a', 'w3emc_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo', 'py_netcdf4_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i', 'wgrib2_ver': '2.0.8', 'TOCGRIB2SUPER': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib2super', 'COM_ICE_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2', 'KEEPDATA': False, 'udunits_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo', 'cdo_ver': '2.2.0', 'COM_OCEAN_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input', 'FNTSFA': ' ', 'LOADEDMODULES': 'contrib/0.1:intel-oneapi-compilers/2023.1.0:stack-intel/2021.9.0:intel-oneapi-mpi/2021.9.0:stack-intel-oneapi-mpi/2021.9.0:intel-oneapi-mkl/2023.1.0:zlib/1.2.13:pigz/2.7:zstd/1.5.2:tar/1.34:gettext/0.21.1:libxcrypt/4.4.35:sqlite/3.43.2:util-linux-uuid/2.38.1:python/3.11.6:libjpeg/2.1.0:jasper/2.0.32:libpng/1.6.37:openjpeg/2.3.1:eccodes/2.32.0:fftw/3.3.10:nghttp2/1.57.0:curl/8.4.0:proj/9.2.1:udunits/2.2.28:cdo/2.2.0:hdf5/1.14.0:snappy/1.1.10:c-blosc/1.21.5:netcdf-c/4.9.2:netcdf-fortran/4.6.1:parallel-netcdf/1.12.2:parallelio/2.5.10:esmf/8.6.0:antlr/2.7.7:gsl/2.7.1:nco/5.0.6:bacio/2.4.1:w3emc/2.10.0:prod_util/2.1.1:g2/3.4.5:sp/2.5.0:ip/4.3.0:grib-util/1.3.0:g2tmpl/1.13.0:gsi-ncdiag/1.1.2:crtm-fix/2.4.0.1_emc:git-lfs/3.1.2:crtm/2.4.0.1:openblas/0.3.24:py-setuptools/63.4.3:py-numpy/1.23.4:bufr/11.7.0:gmake/4.2.1:wgrib2/2.0.8:py-f90nml/1.4.3:py-cftime/1.0.3.4:py-netcdf4/1.5.8:libyaml/0.2.5:py-pyyaml/6.0:py-markupsafe/2.1.3:py-jinja2/3.1.2:py-bottleneck/1.3.7:py-numexpr/2.8.4:py-et-xmlfile/1.0.1:py-openpyxl/3.1.2:py-pytz/2023.3:py-pyxlsb/1.0.10:py-xlrd/2.0.1:py-xlsxwriter/3.1.7:py-xlwt/1.3.0:py-pandas/1.5.3:py-six/1.16.0:py-python-dateutil/2.8.2:g2c/1.8.0:netcdf-cxx4/4.3.1:met/9.1.3:metplus/3.1.1:py-packaging/23.1:py-xarray/2023.7.0:prepobs/1.1.0:fit2obs/1.1.7.1:globus-cli/3.35.2:module_base.hercules', 'SLURM_JOB_ACCOUNT': 'fv3-cpu', 'DO_OCN': True, 'SLURM_PRIO_PROCESS': 0, 'HOMEfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1', 'gmake_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq', 'FIXfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/fix', 'py_python_dateutil_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy', 'tasks_per_node': 1, 'SLURM_NPROCS': 1, 'COM_CHEM_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history', 'LMOD_ROOT': '/apps/other/lmod', 'DOHYBVAR': '{{ DOHYBVAR }}', 'GSL_ROOT_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z', 'SERIAL_F77': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'IP_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_d', 'SHLVL': 4, '_ModuleTable071_': 'e30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAic3RhY2staW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMCIsCndWID0gIjAwMDAwMjAyMS4wMDAwMDAwMDkuKnpmaW5hbCIsCn0sCnRhciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC90YXIvMS4zNC5sdWEiLApmdWxsTmFtZSA9ICJ0YXIvMS4zNCIsCmxvYWRPcmRlciA9IDEwLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIs', '_ModuleTable012_': 'c3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImZmdHcvMy4zLjEwIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAwMy4wMDAwMDAwMTAuKnpmaW5hbCIsCn0sCmZpdDJvYnMgPSB7CmZuID0gIi93b3JrMi9ub2FhL2dsb2JhbC9yb2xlLWdsb2JhbC9naXQvRml0Mk9icy92MS4xLjcuMS9tb2R1bGVmaWxlcy9maXQyb2JzLzEuMS43LjEubHVhIiwKZnVsbE5hbWUgPSAiZml0Mm9icy8xLjEuNy4xIiwKbG9hZE9yZGVyID0gODIsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZml0Mm9icy8xLjEuNy4xIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4wMDAwMDAwMDcuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApnMiA9', 'INCVARS_EFOLD': 5, 'SLURM_NNODES': 1, 'RESERVATION': '', 'FHMIN_GFS': 0, 'FHMAX_HF_WAV': 48, 'FIXugwd': '/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd', 'BASH_ENV': '/apps/other/lmod/lmod/init/bash', 'CDATE': datetime.datetime(2021, 3, 23, 12, 0), '_ModuleTable063_': 'aW5hbCIsCn0sClsicHkteGxyZCJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXhscmQvMi4wLjEubHVhIiwKZnVsbE5hbWUgPSAicHkteGxyZC8yLjAuMSIsCmxvYWRPcmRlciA9IDY5LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXhscmQvMi4wLjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJweS14bHN4d3JpdGVyIl0g', '_ModuleTable006_': 'Mi4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMi4qemZpbmFsIiwKfSwKY29udHJpYiA9IHsKZm4gPSAiL2FwcHMvb3RoZXIvbW9kdWxlZmlsZXMvY29udHJpYi8wLjEiLApmdWxsTmFtZSA9ICJjb250cmliLzAuMSIsCmxvYWRPcmRlciA9IDEsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAwLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY29udHJpYiIsCndWID0gIjAwMDAwMDAwMC4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sCmNydG0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50', 'FHOUT_OCN_GFS': 6, '_ModuleTable026_': 'cmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2phc3Blci8yLjAuMzIubHVhIiwKZnVsbE5hbWUgPSAiamFzcGVyLzIuMC4zMiIsCmxvYWRPcmRlciA9IDE3LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImphc3Blci8yLjAuMzIiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAzMi4qemZpbmFsIiwKfSwKbGlianBlZyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1z', 'DONST': True, '_ModuleTable051_': 'ay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL3B5LW5ldGNkZjQvMS41LjgubHVhIiwKZnVsbE5hbWUgPSAicHktbmV0Y2RmNC8xLjUuOCIsCmxvYWRPcmRlciA9IDU4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LW5ldGNkZjQvMS41LjgiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA1LjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKWyJweS1udW1leHByIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3Rh', 'LMOD_sys': 'Linux', '_ModuleTable033_': 'L3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL25jby81LjAuNi5sdWEiLApmdWxsTmFtZSA9ICJuY28vNS4wLjYiLApsb2FkT3JkZXIgPSAzNywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJuY28vNS4wLjYiLAp3ViA9ICIwMDAwMDAwMDUuMDAwMDAwMDAwLjAwMDAwMDAwNi4qemZpbmFsIiwKfSwKWyJuZXRjZGYtYyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1', 'FHMIN_WAV': 0, 'LMOD_FAMILY_METACOMPILER_VERSION': '2021.9.0', 'I_MPI_PMI_LIBRARY': '/opt/slurm/lib/libpmi2.so', 'machine': 'HERCULES', 'py_pyyaml_ver': 6.0, 'OCL_ICD_FILENAMES': 'libintelocl_emu.so:libalteracl.so:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64/libintelocl.so', '_ModuleTable017_': 'bG9hZE9yZGVyID0gNDgsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ2l0LWxmcy8zLjEuMiIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbImdsb2J1cy1jbGkiXSA9IHsKZm4gPSAiL2FwcHMvb3RoZXIvbW9kdWxlZmlsZXMvZ2xvYnVzLWNsaS8zLjM1LjIubHVhIiwKZnVsbE5hbWUgPSAiZ2xvYnVzLWNsaS8zLjM1LjIiLApsb2FkT3JkZXIgPSA4MywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJnbG9idXMtY2xpIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAzNS4wMDAwMDAwMDIu', 'DO_FETCH_HPSS': False, 'PROJ_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/proj', 'FHOUT_ICE': 3, 'COM_WAVE_STATION_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station', 'COMINecmwf': '/work2/noaa/global/role-global/data/external_gempak/ecmwf', 'COM_TOP_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}', 'WGRIB2_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include', 'mpmd_opt': '--multi-prog --output=mpmd.%j.%t.out', 'DO_PREP_SFC': False, 'LMOD_G2TMPLVIRT_NAME': 'g2tmpl', '_ModuleTable066_': 'bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5dGhvbi8zLjExLjYubHVhIiwKZnVsbE5hbWUgPSAicHl0aG9uLzMuMTEuNiIsCmxvYWRPcmRlciA9IDE1LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5dGhvbi8zLjExLjYiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDExLjAwMDAwMDAwNi4qemZpbmFsIiwKfSwKc25hcHB5ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAy', '_ModuleTable009_': 'MjEuOS4wL2N1cmwvOC40LjAubHVhIiwKZnVsbE5hbWUgPSAiY3VybC84LjQuMCIsCmxvYWRPcmRlciA9IDIzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImN1cmwvOC40LjAiLAp3ViA9ICIwMDAwMDAwMDguMDAwMDAwMDA0Lip6ZmluYWwiLAp9LAplY2NvZGVzID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2VjY29kZXMvMi4zMi4wLmx1YSIsCmZ1bGxOYW1lID0gImVjY29kZXMvMi4z', 'py_numexpr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq', '__LMOD_REF_COUNT_MANPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/share/man:2;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/man:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/documentation/en/man/common:1;/apps/other/lmod/lmod/share/man:1;/usr/share/man:1;/apps/share/man:1;/apps/man:1;/opt/slurm/share/man:1', 'DOLETKF_OCN': False, 'MPI_GROUP_MAX': 256, 'py_cftime_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i', 'XDG_SESSION_ID': 'c24', 'zstd_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4', 'CLUSTERS': '', '_ModuleTable034_': 'bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAvbmV0Y2RmLWMvNC45LjIubHVhIiwKZnVsbE5hbWUgPSAibmV0Y2RmLWMvNC45LjIiLApsb2FkT3JkZXIgPSAzMCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJuZXRjZGYtYy80LjkuMiIsCndWID0gIjAwMDAwMDAwNC4wMDAwMDAwMDkuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbIm5ldGNkZi1jeHg0Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2st', 'COM_OCEAN_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean', 'CLUSTERS_SERVICE': '', 'pgmout': 'OUTPUT.358494', 'EXECfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/exec', '_ModuleTable031_': 'aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJtZXQvOS4xLjMiLAp3ViA9ICIwMDAwMDAwMDkuMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKbWV0cGx1cyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL21ldHBsdXMvMy4xLjEubHVhIiwKZnVsbE5hbWUgPSAibWV0cGx1cy8zLjEuMSIsCmxvYWRPcmRlciA9IDc4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIs', 'COMOUT_ICE_GRIB': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ice/grib2', 'COM_ATMOS_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon', 'HOMEobsproc': '/work2/noaa/global/role-global/git/obsproc/v1.2.0', 'esmf_ver': '8.6.0', 'DO_ARCHCOM': False, 'SLURM_SUBMIT_HOST': 'hercules-login-1.hpc.msstate.edu', 'WRITE_DOPOST': True, 'EXECcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/exec', 'DUMP_SUFFIX': '', '_ModuleTable072_': 'CnVzZXJOYW1lID0gInRhci8xLjM0IiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAzNC4qemZpbmFsIiwKfSwKdWR1bml0cyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC91ZHVuaXRzLzIuMi4yOC5sdWEiLApmdWxsTmFtZSA9ICJ1ZHVuaXRzLzIuMi4yOCIsCmxvYWRPcmRlciA9IDI1LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInVkdW5pdHMvMi4yLjI4IiwKd1YgPSAiMDAwMDAwMDAy', 'CASE': 'C48', 'SENDDBN': False, 'gettext_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43', 'COM_ICE_BMATRIX_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice', 'git_lfs_ROOT': '/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l', '_ModuleTable001_': 'X01vZHVsZVRhYmxlXyA9IHsKTVR2ZXJzaW9uID0gMywKY19yZWJ1aWxkVGltZSA9IGZhbHNlLApjX3Nob3J0VGltZSA9IGZhbHNlLApkZXB0aFQgPSB7fSwKZmFtaWx5ID0gewpNZXRhQ29tcGlsZXIgPSAic3RhY2staW50ZWwiLApNZXRhTVBJID0gInN0YWNrLWludGVsLW9uZWFwaS1tcGkiLApjb21waWxlciA9ICJpbnRlbC1vbmVhcGktY29tcGlsZXJzIiwKZzJ0bXBsdmlydCA9ICJnMnRtcGwiLAptcGkgPSAiaW50ZWwtb25lYXBpLW1waSIsCn0sCm1UID0gewphbnRsciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9t', 'COM_ICE_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart', '_ModuleTable025_': 'LAp9LAppcCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9pcC80LjMuMC5sdWEiLApmdWxsTmFtZSA9ICJpcC80LjMuMCIsCmxvYWRPcmRlciA9IDQzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImlwLzQuMy4wIiwKd1YgPSAiMDAwMDAwMDA0LjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKamFzcGVyID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hl', 'COM_OCEAN_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart', 'ARCHCOM_TO': 'globus_hpss', 'COM_ATMOS_WMO_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo', 'ip_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf', 'CLASSPATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/mpi.jar', 'INTELFPGAOCLSDKROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga', 'LD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/host/linux64/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/lib:/usr/lib64:/usr/lib:/opt/slurm/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib', 'DO_CA': True, 'g2_ver': '3.4.5', 'LMOD_FAMILY_COMPILER': 'intel-oneapi-compilers', 'XDG_RUNTIME_DIR': '/run/user/9583', 'BASE_ENV': '/work2/noaa/global/mterry/global-workflow_forked/env', 'SLURM_JOB_ID': 5951735, '_ModuleTable014_': 'dWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nMmMvMS44LjAubHVhIiwKZnVsbE5hbWUgPSAiZzJjLzEuOC4wIiwKbG9hZE9yZGVyID0gNzUsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZzJjLzEuOC4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKZzJ0bXBsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9p', 'OCEANICEPRODUCTS_CONFIG': '/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml', 'NTHREADS_OCNICEPOST': 1, 'DO_JEDIOCNVAR': False, 'FIXreg2grb2': '/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2', 'SLURM_NODEID': 0, 'ip_ver': '4.3.0', 'KMP_AFFINITY': 'scatter', 'FHOUT_ICE_GFS': 6, '_ModuleTable065_': 'L3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS14bHd0LzEuMy4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXhsd3QvMS4zLjAiLApsb2FkT3JkZXIgPSA3MSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS14bHd0LzEuMy4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKcHl0aG9uID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1', 'DO_ATM': True, '_ModuleTable004_': 'IDUzLApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImJ1ZnIvMTEuNy4wIiwKd1YgPSAiMDAwMDAwMDExLjAwMDAwMDAwNy4qemZpbmFsIiwKfSwKWyJjLWJsb3NjIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvYy1ibG9zYy8xLjIxLjUubHVhIiwKZnVsbE5hbWUgPSAiYy1ibG9zYy8xLjIxLjUiLApsb2FkT3JkZXIgPSAyOSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1', 'esmf_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep', 'LMOD_FAMILY_MPI_VERSION': '2021.9.0', '_ModuleTable020_': 'c3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9nc2ktbmNkaWFnLzEuMS4yLmx1YSIsCmZ1bGxOYW1lID0gImdzaS1uY2RpYWcvMS4xLjIiLApsb2FkT3JkZXIgPSA0NiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJnc2ktbmNkaWFnLzEuMS4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCmdzbCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYu', 'DO_JEDIATMVAR': False, 'DO_AERO_ANL': False, 'NCLEN': '/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen', 'PS4': '+ $(basename ${BASH_SOURCE[0]:-${FUNCNAME[0]:-"Unknown"}})[${LINENO}]', 'COM_SNOW_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow', '_ModuleTable053_': 'YWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbnVtcHkvMS4yMy40Lmx1YSIsCmZ1bGxOYW1lID0gInB5LW51bXB5LzEuMjMuNCIsCmxvYWRPcmRlciA9IDUyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSA4LApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LW51bXB5LzEuMjMuNCIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMjMuMDAwMDAwMDA0Lip6ZmluYWwiLAp9LApbInB5LW9wZW5weXhsIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0', '__LMOD_REF_COUNT_ACL_BOARD_VENDOR_PATH': '/opt/Intel/OpenCLFPGA/oneAPI/Boards:1', 'LMOD_FAMILY_METAMPI_VERSION': '2021.9.0', '_ModuleTable059_': 'IDY4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXB5eGxzYi8xLjAuMTAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAwLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKWyJweS1weXlhbWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1weXlhbWwvNi4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXB5eWFtbC82LjAiLApsb2FkT3JkZXIgPSA2MCwKcHJvcFQgPSB7fSwKc3RhY2tE', 'DO_CALC_INCREMENT_ENKF_GFS': False, 'MKLROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0', 'DEBUGINFOD_URLS': 'https://debuginfod.centos.org/ ', 'sven_root_path': '/home/gfekete/sven', '_ModuleTable047_': 'LApbInB5LWV0LXhtbGZpbGUiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1ldC14bWxmaWxlLzEuMC4xLmx1YSIsCmZ1bGxOYW1lID0gInB5LWV0LXhtbGZpbGUvMS4wLjEiLApsb2FkT3JkZXIgPSA2NSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1ldC14bWxmaWxlLzEuMC4xIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMC4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsi', 'restart_interval_enkfgfs': 3, 'SENDDBN_NTC': False, 'FIXaer': '/work2/noaa/global/mterry/global-workflow_forked/fix/aer', 'FHMAX_HF_GFS': 48, 'COM_ICE_LETKF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf', 'LMOD_FAMILY_MPI': 'intel-oneapi-mpi', 'netcdf_fortran_ver': '4.6.1', 'FIXmom': '/work2/noaa/global/mterry/global-workflow_forked/fix/mom6', 'cyc': 12, 'BACIO_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib/libbacio_4.a', 'FC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'FHOUT_WAV_GFS': 3, 'SDATE_GFS': datetime.datetime(2021, 3, 23, 12, 0), 'BACIO_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib/libbacio_8.a', 'sp_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr', 'OPS_RES': 'C768', 'NMEM_ENS_GFS': 30, 'which_declare': 'declare -f', '_ModuleTable056_': 'OS4wL3B5LXBhbmRhcy8xLjUuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1wYW5kYXMvMS41LjMiLApsb2FkT3JkZXIgPSA3MiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1wYW5kYXMvMS41LjMiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA1LjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKWyJweS1weXRob24tZGF0ZXV0aWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1weXRob24tZGF0ZXV0aWwvMi44', 'COM_WAVE_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded', 'LMOD_FAMILY_G2TMPLVIRT': 'g2tmpl', 'QUEUE_DTN': 'batch', 'DO_GENESIS_FSU': False, 'prod_util_ver': '2.1.1', 'MPI_F90': 'mpiifort', 'LMOD_FAMILY_COMPILER_VERSION': '2023.1.0', '_ModuleTable042_': 'MDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcGlnei8yLjcubHVhIiwKZnVsbE5hbWUgPSAicGlnei8yLjciLApsb2FkT3JkZXIgPSA4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gNCwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInBpZ3ovMi43IiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4qemZpbmFsIiwKfSwKcHJlcG9icyA9IHsKZm4gPSAiL3dvcmsyL25vYWEvZ2xvYmFsL3JvbGUtZ2xvYmFsL2dpdC9wcmVwb2JzL3YxLjEuMC9tb2R1bGVmaWxlcy9wcmVwb2JzLzEuMS4wLmx1YSIsCmZ1bGxOYW1lID0gInByZXBvYnMvMS4xLjAiLApsb2FkT3JkZXIgPSA4MSwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0', 'openjpeg_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu', 'USHfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/ush', 'YAML_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib', 'DO_GSISOILDA': False, 'nemsio_ver': '2.5.4', 'DO_NPOESS': False, 'SLURM_MPI_TYPE': 'pmi2', 'OMP_STACKSIZE': 2048000, 'py_et_xmlfile_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq', 'FORECAST_HOUR': 78, 'ATARDIR': '/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW', 'EXECgfs': '/work2/noaa/global/mterry/global-workflow_forked/exec', 'py_python_dateutil_ver': '2.8.2', 'assim_freq': 6, 'DOIAU': True, 'FHOUT_WAV': 1, 'py_pyyaml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2', 'met_ver': '9.1.3', 'gsi_ncdiag_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe', 'DO_FETCH_LOCAL': False, 'NLSPATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64/locale/%l_%t/%N:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin/locale/%l_%t/%N', 'netcdf_fortran_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe', 'SP_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_4.a', 'COM_OCEAN_BMATRIX_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean', 'SLURM_CONF': '/var/spool/slurmd/conf-cache/slurm.conf', 'PATH': '/apps/other/globus-cli-3.35.2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r/ush:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/bin:/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/bin/intel64:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/bin:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/bin:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin:/usr/sbin:/usr/bin:/apps/sbin:/apps/bin:/opt/slurm/bin:/home/gfekete/sven/bin', 'SP_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_8.a', 'netcdf_c_ver': '4.9.2', 'SLURM_JOB_NAME': 'C48_S2SW_gfs_ice_prod_f078_12', 'COM_ICE_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history', 'NLN': '/bin/ln -sf', 'MODULEPATH': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles:/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0-a66eaip/g2tmpl/1.13.0/intel/2021.9.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/g2tmpl/1.13.0/intel/2021.9.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/intel-oneapi-mpi/2021.9.0-a66eaip/oneapi/2023.1.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core:/work2/noaa/global/mterry/global-workflow_forked/modulefiles:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core:/apps/other/modulefiles:/apps/containers/modulefiles:/apps/licensed/modulefiles:/apps/contrib/modulefiles', 'CC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icc', 'grib_util_ver': '1.3.0', 'py_f90nml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z', 'DO_MERGENSST': False, 'SLURM_NTASKS_PER_NODE': 1, '_LMFILES_': '/apps/other/modulefiles/contrib/0.1:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core/intel-oneapi-compilers/2023.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core/stack-intel/2021.9.0.lua:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0/intel-oneapi-mpi/2021.9.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/stack-intel-oneapi-mpi/2021.9.0.lua:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core/intel-oneapi-mkl/2023.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/zlib/1.2.13.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/pigz/2.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/zstd/1.5.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/tar/1.34.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gettext/0.21.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libxcrypt/4.4.35.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/sqlite/3.43.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/util-linux-uuid/2.38.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/python/3.11.6.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libjpeg/2.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/jasper/2.0.32.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libpng/1.6.37.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/openjpeg/2.3.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/eccodes/2.32.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/fftw/3.3.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/nghttp2/1.57.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/curl/8.4.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/proj/9.2.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/udunits/2.2.28.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/cdo/2.2.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/hdf5/1.14.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/snappy/1.1.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/c-blosc/1.21.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-c/4.9.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-fortran/4.6.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/parallel-netcdf/1.12.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/parallelio/2.5.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/esmf/8.6.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/antlr/2.7.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gsl/2.7.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/nco/5.0.6.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/bacio/2.4.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/w3emc/2.10.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/prod_util/2.1.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2/3.4.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/sp/2.5.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/ip/4.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/grib-util/1.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2tmpl/1.13.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/gsi-ncdiag/1.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/crtm-fix/2.4.0.1_emc.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/git-lfs/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/crtm/2.4.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/openblas/0.3.24.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-setuptools/63.4.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-numpy/1.23.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/bufr/11.7.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gmake/4.2.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/wgrib2/2.0.8.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-f90nml/1.4.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-cftime/1.0.3.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/py-netcdf4/1.5.8.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libyaml/0.2.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pyyaml/6.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-markupsafe/2.1.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-jinja2/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-bottleneck/1.3.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-numexpr/2.8.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-et-xmlfile/1.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-openpyxl/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pytz/2023.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pyxlsb/1.0.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlrd/2.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlsxwriter/3.1.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlwt/1.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pandas/1.5.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-six/1.16.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-python-dateutil/2.8.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2c/1.8.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-cxx4/4.3.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/met/9.1.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/metplus/3.1.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-packaging/23.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xarray/2023.7.0.lua:/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles/prepobs/1.1.0.lua:/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles/fit2obs/1.1.7.1.lua:/apps/other/modulefiles/globus-cli/3.35.2.lua:/work2/noaa/global/mterry/global-workflow_forked/modulefiles/module_base.hercules.lua', 'globus_cli_ver': 3.27, '_ModuleTable028_': 'L2xpYnBuZy8xLjYuMzcubHVhIiwKZnVsbE5hbWUgPSAibGlicG5nLzEuNi4zNyIsCmxvYWRPcmRlciA9IDE4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImxpYnBuZy8xLjYuMzciLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA2LjAwMDAwMDAzNy4qemZpbmFsIiwKfSwKbGlieGNyeXB0ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2xpYnhjcnlwdC80LjQuMzUubHVhIiwKZnVsbE5hbWUgPSAibGlieGNyeXB0', 'py_numpy_ver': '1.23.4', '__LMOD_REF_COUNT_HDF5_PLUGIN_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/plugins:1', 'bufr_ver': '11.7.0', 'COMINnam': '/work2/noaa/global/role-global/data/external_gempak/nam', 'DBUS_SESSION_BUS_ADDRESS': 'unix:path=/run/user/9583/bus', 'DO_AERO_FCST': False, 'py_pandas_ver': '1.5.3', '_ModuleTable039_': 'dGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9vcGVuanBlZy8yLjMuMS5sdWEiLApmdWxsTmFtZSA9ICJvcGVuanBlZy8yLjMuMSIsCmxvYWRPcmRlciA9IDE5LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm9wZW5qcGVnLzIuMy4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsicGFyYWxsZWwtbmV0Y2RmIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMv', 'COM_ICE_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice', 'UUID_ORION_DTN': '8a10dd4f-24ee-4794-a39d-9c313ab6a34b', 'MPI_F77': 'mpiifort', 'LMOD_CMD': '/apps/other/lmod/lmod/libexec/lmod', '_ModuleTable045_': 'aW5hbCIsCn0sClsicHktYm90dGxlbmVjayJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LWJvdHRsZW5lY2svMS4zLjcubHVhIiwKZnVsbE5hbWUgPSAicHktYm90dGxlbmVjay8xLjMuNyIsCmxvYWRPcmRlciA9IDYzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWJvdHRsZW5lY2svMS4zLjciLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAzLjAwMDAwMDAwNy4qemZpbmFs', 'LMOD_MPI_VERSION': '2021.9.0-a66eaip', 'LMOD_SYSTEM_NAME': 'hercules', 'RUN_ENVIR': 'emc', 'w3emc_ver': '2.10.0', '__LMOD_REF_COUNT_LD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/host/linux64/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/lib:1;/usr/lib64:1;/usr/lib:1;/opt/slurm/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:2', 'DOHYBVAR_OCN': False, 'NUM_SND_COLLECTIVES': 9, 'walltime': '00:15:00', 'netcdf_cxx4_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu', 'ens_tracker_ver': 'v1.2.0', 'LMOD_SYSTEM_DEFAULT_MODULES': 'contrib', 'OMPI_MCA_plm_slurm_args': '--external-launcher', '_ModuleTable078_': 'cyIKLCAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLWE2NmVhaXAvZzJ0bXBsLzEuMTMuMC9pbnRlbC8yMDIxLjkuMCIKLCAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9nMnRtcGwvMS4xMy4wL2ludGVsLzIwMjEuOS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVz', 'FIXcpl': '/work2/noaa/global/mterry/global-workflow_forked/fix/cpl', 'G2C_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64/libg2c.so', '_ModuleTable019_': 'cmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9ncmliLXV0aWwvMS4zLjAubHVhIiwKZnVsbE5hbWUgPSAiZ3JpYi11dGlsLzEuMy4wIiwKbG9hZE9yZGVyID0gNDQsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ3JpYi11dGlsLzEuMy4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKWyJnc2ktbmNkaWFnIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMv', 'SLURM_JOB_GID': 17000, 'SLURM_GET_USER_ENV': 1, 'IAUFHRS_ENKF': [3, 6, 9], 'CPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/include:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/include:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/include', 'G2TMPL_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include', 'DO_JEDISNOWDA': False, 'SP_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_d.a', 'pigz_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq', 'COM_ATMOS_BUFR_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr', 'COM_OCEAN_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2', 'CNVGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/cnvgrib', 'COM_ATMOS_TRACK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks', 'DO_VERFRAD': True, 'pgmerr': 'errfile', 'MPI_BUFS_PER_HOST': 2048, 'DO_PREP_OBS_AERO': False, 'LMOD_FAMILY_G2TMPLVIRT_VERSION': '1.13.0', 'PNG_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu', 'COM_ATMOS_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart', 'DEGRIB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/degrib2', 'DO_ANLSTAT': False, 'LMOD_MPI_NAME': 'intel-oneapi-mpi', 'py_f90nml_ver': '1.4.3', 'NMEM_ENS_GFS_OFFSET': 20, 'SMOOTH_ENKF': False, 'FHMAX_GOES': 120, 'OLDPWD': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', 'QUEUE': 'batch', 'COM_OBS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/obs', 'DO_VMINMON': True, 'metplus_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r', 'VERBOSE': True, 'SLURM_JOB_NODELIST': 'hercules-08-38', 'CLUSTER': 'hercules', 'DBNROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn', 'FHOUT_HF_WAV': 1, 'PARMgfs': '/work2/noaa/global/mterry/global-workflow_forked/parm', 'I_MPI_HYDRA_BOOTSTRAP': 'slurm', 'APRUN_OCNICEPOST': 'srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1', 'BASH_FUNC_ml%%': '() { eval "$($LMOD_DIR/ml_cmd "$@")"\n}', 'BASH_FUNC_which%%': '() { ( alias;\n eval ${which_declare} ) | /usr/bin/which --tty-only --read-alias --read-functions --show-tilde --show-dot $@\n}', 'BASH_FUNC_module%%': '() { if [ -z "${LMOD_SH_DBG_ON+x}" ]; then\n case "$-" in \n *v*x*)\n __lmod_sh_dbg=\'vx\'\n ;;\n *v*)\n __lmod_sh_dbg=\'v\'\n ;;\n *x*)\n __lmod_sh_dbg=\'x\'\n ;;\n esac;\n fi;\n if [ -n "${__lmod_sh_dbg:-}" ]; then\n set +$__lmod_sh_dbg;\n echo "Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for Lmod\'s output" 1>&2;\n fi;\n eval "$($LMOD_CMD $LMOD_SHELL_PRGM "$@")" && eval "$(${LMOD_SETTARG_CMD:-:} -s sh)";\n __lmod_my_status=$?;\n if [ -n "${__lmod_sh_dbg:-}" ]; then\n echo "Shell debugging restarted" 1>&2;\n set -$__lmod_sh_dbg;\n fi;\n unset __lmod_sh_dbg;\n return $__lmod_my_status\n}', 'BASH_FUNC_declare_from_tmpl%%': '() { if [[ ${DEBUG_WORKFLOW:-"NO"} == "NO" ]]; then\n set +x;\n fi;\n local opts="-g";\n local OPTIND=1;\n while getopts "rx" option; do\n opts="${opts}${option}";\n done;\n shift $((OPTIND-1));\n for input in "$@";\n do\n IFS=\':\' read -ra args <<< "${input}";\n local com_var="${args[0]}";\n local template;\n local value;\n if (( ${#args[@]} > 1 )); then\n template="${args[1]}";\n else\n template="${com_var}_TMPL";\n fi;\n if [[ ! -v "${template}" ]]; then\n echo "FATAL ERROR in declare_from_tmpl: Requested template ${template} not defined!";\n exit 2;\n fi;\n value=$(echo "${!template}" | envsubst);\n declare ${opts} "${com_var}"="${value}";\n echo "declare_from_tmpl :: ${com_var}=${value}";\n done;\n set_trace\n}', 'BASH_FUNC_err_exit%%': ['() { set +eux;\n msg1=${*:-Job ${jobid} failed};\n if [[ -n "${pgm}" ]]; then\n msg1+="', 'ERROR IN ${pgm}";\n fi;\n if [[ -n "${err}" ]]; then\n msg1+=" RETURN CODE ${err}";\n fi;\n msg2="\n -------------------------------------------------------------\n -- FATAL ERROR: ${msg1}\n -- ABNORMAL EXIT at $(date) on ${HOSTNAME}\n -------------------------------------------------------------\n ";\n echo "${msg2}" 1>&2;\n module list;\n echo "" 1>&2;\n echo "${msg1}" 1>&2;\n if [[ -n "${DATA}" ]]; then\n echo "${DATA}" 1>&2;\n ls -ltr "${DATA}" 1>&2;\n else\n echo "WARNING: DATA variable not defined" 1>&2;\n fi;\n if [[ -n "${pgmout}" ]]; then\n if [[ -s errfile ]]; then\n echo "----- contents of errfile -----" >> "${pgmout}";\n cat errfile >> "${pgmout}";\n fi;\n cat "${pgmout}" 1>&2;\n else\n if [[ -s errfile ]]; then\n cat errfile 1>&2;\n fi;\n fi;\n if [[ "${SENDECF}" == "YES" ]]; then\n timeout 30 ecflow_client --msg "${ECF_NAME}: ${msg1}";\n timeout 30 ssh "${ECF_HOST}" "echo \\"${msg}2\\" >> ${ECF_JOBOUT:?}";\n fi;\n if [[ "${SENDECF}" == "YES" ]]; then\n ecflow_client --kill="${ECF_NAME:?}";\n fi;\n if [[ -n "${PBS_JOBID}" ]]; then\n qdel "${PBS_JOBID}";\n else\n if [[ -n "${SLURM_JOB_ID}" ]]; then\n scancel "${SLURM_JOB_ID}";\n fi;\n fi\n}'], 'BASH_FUNC_wait_for_file%%': '() { set +x;\n local file_name=${1:?"wait_for_file() requires a file name"};\n local sleep_interval=${2:-60};\n local max_tries=${3:-100};\n for ((iter=0; iter' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.oceanice_products -+++ config.oceanice_products[5]echo 'BEGIN: config.oceanice_products' -BEGIN: config.oceanice_products -+++ config.oceanice_products[8]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources oceanice_products -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=oceanice_products -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[978]walltime=00:15:00 -++++ config.resources[979]ntasks=1 -++++ config.resources[980]tasks_per_node=1 -++++ config.resources[981]threads_per_task=1 -++++ config.resources[982]memory=96GB -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n '' ]] -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export memory -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.oceanice_products[11]export MAX_TASKS=25 -+++ config.oceanice_products[11]MAX_TASKS=25 -+++ config.oceanice_products[13]export OCEANICEPRODUCTS_CONFIG=/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml -+++ config.oceanice_products[13]OCEANICEPRODUCTS_CONFIG=/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml -+++ config.oceanice_products[16]export NFHRS_PER_GROUP=3 -+++ config.oceanice_products[16]NFHRS_PER_GROUP=3 -+++ config.oceanice_products[18]echo 'END: config.oceanice_products' -END: config.oceanice_products -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env oceanice_products -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=oceanice_products -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 100663296 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 1 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 1 ]] -+++ HERCULES.env[34]max_threads_per_task=80 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 80 ]] -+++ HERCULES.env[40][[ 1 -gt 80 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 1' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[286]export NTHREADS_OCNICEPOST=1 -+++ HERCULES.env[286]NTHREADS_OCNICEPOST=1 -+++ HERCULES.env[287]export 'APRUN_OCNICEPOST=srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1' -+++ HERCULES.env[287]APRUN_OCNICEPOST='srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1' -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_OCEANICE_PRODUCTS[11]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[11]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[11]declare_from_tmpl -rx COMIN_ICE_HISTORY:COM_ICE_HISTORY_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_ICE_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/ice/history -+ JGLOBAL_OCEANICE_PRODUCTS[12]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[12]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[12]declare_from_tmpl -rx COMOUT_ICE_GRIB:COM_ICE_GRIB_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_ICE_GRIB=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ice/grib2 -+ JGLOBAL_OCEANICE_PRODUCTS[13]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[13]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[13]declare_from_tmpl -rx COMOUT_ICE_NETCDF:COM_ICE_NETCDF_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_ICE_NETCDF=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ice/netcdf -+ JGLOBAL_OCEANICE_PRODUCTS[17]/work2/noaa/global/mterry/global-workflow_forked/scripts/exglobal_oceanice_products.py -2025-07-28 21:34:54,440 - INFO - root : BEGIN: __main__.main -2025-07-28 21:34:54,441 - DEBUG - root : ( ) -2025-07-28 21:34:54,450 - INFO - oceanice_products: BEGIN: OceanIceProducts.__init__ -2025-07-28 21:34:54,450 - DEBUG - oceanice_products: ( , {'SHELL': '/bin/bash', 'PDY': datetime.datetime(2021, 3, 23, 0, 0), 'DO_JEDIATMENS': False, 'sqlite_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4', 'COPYGB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/copygb', 'WORK_hercules': '/work/hercules', 'TOCGRIB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib2', 'G2_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/include_d', 'PIO_TYPENAME_VALID_VALUES': ['netcdf', 'netcdf4p', 'netcdf4c', 'pnetcdf'], 'WGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/wgrib', 'SERIAL_CXX': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icpc', 'FHMAX_WAV': 9, 'FIXlut': '/work2/noaa/global/mterry/global-workflow_forked/fix/lut', 'fftw_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj', 'snappy_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx', '_ModuleTable016_': 'bGxOYW1lID0gImdldHRleHQvMC4yMS4xIiwKbG9hZE9yZGVyID0gMTEsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ2V0dGV4dC8wLjIxLjEiLAp3ViA9ICIwMDAwMDAwMDAuMDAwMDAwMDIxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJnaXQtbGZzIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvZ2l0LWxmcy8zLjEuMi5sdWEiLApmdWxsTmFtZSA9ICJnaXQtbGZzLzMuMS4yIiwK', 'MAKE_NSSTBUFR': False, 'cdo_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e', 'curl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops', 'PIO_VERSION_MAJOR': 2, 'threads_per_task': 1, 'SLURM_JOB_USER': 'mterry', '_ModuleTable067_': 'NC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3NuYXBweS8xLjEuMTAubHVhIiwKZnVsbE5hbWUgPSAic25hcHB5LzEuMS4xMCIsCmxvYWRPcmRlciA9IDI4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInNuYXBweS8xLjEuMTAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAxLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKc3AgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45', 'COM_ICE_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}', 'COM_OCEAN_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}', '_ModuleTable022_': 'ZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9oZGY1LzEuMTQuMC5sdWEiLApmdWxsTmFtZSA9ICJoZGY1LzEuMTQuMCIsCmxvYWRPcmRlciA9IDI3LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImhkZjUvMS4xNC4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAxNC4qemZpbmFsIiwKfSwKWyJpbnRlbC1vbmVhcGktY29tcGlsZXJzIl0gPSB7CmZuID0gIi9hcHBzL3NwYWNrLW1hbmFnZWQvbW9kdWxlZmlsZXMvbGludXgtcm9ja3k5LXg4Nl82NC9Db3JlL2ludGVsLW9uZWFwaS1jb21waWxlcnMvMjAyMy4xLjAubHVhIiwKZnVsbE5hbWUgPSAiaW50ZWwtb25lYXBpLWNvbXBpbGVy', 'SLURM_TASKS_PER_NODE': 1, 'COM_ATMOS_OZNMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon', 'COM_ATMOS_GEMPAK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}', 'SLURM_JOB_UID': 9583, 'HISTCONTROL': 'ignoredups', 'SLURM_EXPORT_ENV': 'NONE', 'g2c_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp', 'COM_OCEAN_NETCDF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf', 'GRB2INDEX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grb2index', 'l4densvar': True, 'FHOUT_HF_GFS': 1, 'I_MPI_EXTRA_FILESYSTEM': 1, 'SLURM_TASK_PID': 449818, 'EXECprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/exec', 'g2tmpl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel', 'MPI_BUFS_PER_PROC': 2048, 'IAU_DELTHRS_ENKF': 6, 'COM_WAVE_WMO_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo', '_ModuleTable027_': 'LTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9saWJqcGVnLzIuMS4wLmx1YSIsCmZ1bGxOYW1lID0gImxpYmpwZWcvMi4xLjAiLApsb2FkT3JkZXIgPSAxNiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJsaWJqcGVnLzIuMS4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKbGlicG5nID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4w', 'NMEM_ENS': 0, '__LMOD_REF_COUNT_CLASSPATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/mpi.jar:1', '_ModuleTable050_': 'ZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbWFya3Vwc2FmZS8yLjEuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1tYXJrdXBzYWZlLzIuMS4zIiwKbG9hZE9yZGVyID0gNjEsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktbWFya3Vwc2FmZS8yLjEuMyIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDEuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LW5ldGNkZjQiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFj', 'err': 0, '_ModuleTable032_': 'CnVzZXJOYW1lID0gIm1ldHBsdXMvMy4xLjEiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJtb2R1bGVfYmFzZS5oZXJjdWxlcyJdID0gewpmbiA9ICIvd29yazIvbm9hYS9nbG9iYWwvbXRlcnJ5L2dsb2JhbC13b3JrZmxvd19mb3JrZWQvbW9kdWxlZmlsZXMvbW9kdWxlX2Jhc2UuaGVyY3VsZXMubHVhIiwKZnVsbE5hbWUgPSAibW9kdWxlX2Jhc2UuaGVyY3VsZXMiLApsb2FkT3JkZXIgPSA4NCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDAsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJtb2R1bGVfYmFzZS5oZXJjdWxlcyIsCndWID0gIk0uKnpmaW5hbCIsCn0sCm5jbyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGlj', 'CHGRP_CMD': 'chgrp rstprod', 'PKG_CONFIG_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib/pkgconfig:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/pkgconfig:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/lib/pkgconfig', 'SLURM_LOCALID': 0, 'MAX_TASKS': 25, '_ModuleTable040_': 'aW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9wYXJhbGxlbC1uZXRjZGYvMS4xMi4yLmx1YSIsCmZ1bGxOYW1lID0gInBhcmFsbGVsLW5ldGNkZi8xLjEyLjIiLApsb2FkT3JkZXIgPSAzMiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwYXJhbGxlbC1uZXRjZGYvMS4xMi4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAxMi4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCnBhcmFsbGVsaW8gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0', 'SLURM_SUBMIT_DIR': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', '_ModuleTable077_': 'LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvenN0ZC8xLjUuMi5sdWEiLApmdWxsTmFtZSA9ICJ6c3RkLzEuNS4yIiwKbG9hZE9yZGVyID0gOSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMywKc3RhY2tEZXB0aCA9IDQsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJ6c3RkLzEuNS4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwNS4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCn0sCm1wYXRoQSA9IHsKIi93b3JrMi9ub2FhL2dsb2JhbC9yb2xlLWdsb2JhbC9naXQvRml0Mk9icy92MS4xLjcuMS9tb2R1bGVmaWxlcyIKLCAiL3dvcmsyL25vYWEvZ2xvYmFsL3JvbGUtZ2xvYmFsL2dpdC9wcmVwb2JzL3YxLjEuMC9tb2R1bGVmaWxl', 'HOSTNAME': 'hercules-08-55', 'HISTSIZE': 1000, 'wgrib2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln', 'FRAC_GRID': True, 'G2_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/include_4', 'intel_oneapi_mpi_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse', 'SLURMD_NODENAME': 'hercules-08-55', '_ModuleTable080_': 'bGVzL2xpbnV4LXJvY2t5OS14ODZfNjQvb25lYXBpLzIwMjMuMS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL0NvcmUiCiwgIi93b3JrMi9ub2FhL2dsb2JhbC9tdGVycnkvZ2xvYmFsLXdvcmtmbG93X2ZvcmtlZC9tb2R1bGVmaWxlcyIKLCAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L0NvcmUiCiwgIi9hcHBzL290aGVyL21vZHVsZWZpbGVzIiwgIi9hcHBzL2NvbnRhaW5lcnMvbW9kdWxlZmlsZXMiCiwgIi9hcHBzL2xpY2Vuc2VkL21vZHVsZWZpbGVzIiwg', '__LMOD_REF_COUNT_OCL_ICD_FILENAMES': 'libintelocl_emu.so:1;libalteracl.so:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64/libintelocl.so:1', 'c_blosc_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn', 'FHOUT_OCN': 3, 'COM_OBS_JEDI': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi', 'eccodes_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj', '_ModuleTable037_': 'ay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9uZ2h0dHAyLzEuNTcuMC5sdWEiLApmdWxsTmFtZSA9ICJuZ2h0dHAyLzEuNTcuMCIsCmxvYWRPcmRlciA9IDIyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gNCwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm5naHR0cDIvMS41Ny4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDA1Ny4qemZpbmFsIiwKfSwKb3BlbmJsYXMgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2kt', 'netcdf_c_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy', '_ModuleTable070_': 'c3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzdGFjay1pbnRlbC8yMDIxLjkuMCIsCndWID0gIjAwMDAwMjAyMS4wMDAwMDAwMDkuKnpmaW5hbCIsCn0sClsic3RhY2staW50ZWwtb25lYXBpLW1waSJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3N0YWNrLWludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAubHVhIiwKZnVsbE5hbWUgPSAic3RhY2staW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMCIsCmxvYWRPcmRlciA9IDUsCnByb3BUID0g', 'SLURM_JOB_START_TIME': 1753756458, '_ModuleTable011_': 'b3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZXNtZi84LjYuMCIsCndWID0gIjAwMDAwMDAwOC4wMDAwMDAwMDYuKnpmaW5hbCIsCn0sCmZmdHcgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9mZnR3LzMuMy4xMC5sdWEiLApmdWxsTmFtZSA9ICJmZnR3LzMuMy4xMCIsCmxvYWRPcmRlciA9IDIxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwK', 'DO_NEST': False, 'STMP': '/work2/noaa/stmp/mterry/HERCULES', 'HYDRA_LAUNCHER_EXTRA_ARGS': '--external-launcher', 'ACL_BOARD_VENDOR_PATH': '/opt/Intel/OpenCLFPGA/oneAPI/Boards', '_ModuleTable060_': 'ZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1weXlhbWwvNi4wIiwKd1YgPSAiMDAwMDAwMDA2Lip6ZmluYWwiLAp9LApbInB5LXNldHVwdG9vbHMiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1zZXR1cHRvb2xzLzYzLjQuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1zZXR1cHRvb2xzLzYzLjQuMyIsCmxvYWRPcmRlciA9IDUxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIs', '_ModuleTable007_': 'ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9jcnRtLzIuNC4wLjEubHVhIiwKZnVsbE5hbWUgPSAiY3J0bS8yLjQuMC4xIiwKbG9hZE9yZGVyID0gNDksCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY3J0bS8yLjQuMC4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNC4wMDAwMDAwMDAuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbImNydG0tZml4Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45', 'G2TMPL_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib/libg2tmpl.a', 'DO_WAVE': True, 'SCRgfs': '/work2/noaa/global/mterry/global-workflow_forked/scripts', 'SP_INC8': 'include_8', '_ModuleTable002_': 'b2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9hbnRsci8yLjcuNy5sdWEiLApmdWxsTmFtZSA9ICJhbnRsci8yLjcuNyIsCmxvYWRPcmRlciA9IDM1LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImFudGxyLzIuNy43IiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4wMDAwMDAwMDcuKnpmaW5hbCIsCn0sCmJhY2lvID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2JhY2lvLzIuNC4xLmx1', 'MPICC': 'mpiicc', 'HDF5_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh', '_ModuleTable055_': 'L21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXBhY2thZ2luZy8yMy4xLmx1YSIsCmZ1bGxOYW1lID0gInB5LXBhY2thZ2luZy8yMy4xIiwKbG9hZE9yZGVyID0gNzksCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcGFja2FnaW5nLzIzLjEiLAp3ViA9ICIwMDAwMDAwMjMuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbInB5LXBhbmRhcyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEu', 'SP_INC4': 'include_4', 'FPGA_VARS_DIR': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga', 'FHOUT': 3, '__LMOD_REF_COUNT_MODULEPATH': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles:1;/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0-a66eaip/g2tmpl/1.13.0/intel/2021.9.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/g2tmpl/1.13.0/intel/2021.9.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/intel-oneapi-mpi/2021.9.0-a66eaip/oneapi/2023.1.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core:1;/work2/noaa/global/mterry/global-workflow_forked/modulefiles:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core:1;/apps/other/modulefiles:1;/apps/containers/modulefiles:1;/apps/licensed/modulefiles:1;/apps/contrib/modulefiles:1', 'cycle': 't12z', '_ModuleTable043_': 'aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcmVwb2JzLzEuMS4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKcHJvZF91dGlsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3Byb2RfdXRpbC8yLjEuMS5sdWEiLApmdWxsTmFtZSA9ICJwcm9kX3V0aWwvMi4xLjEiLApsb2FkT3JkZXIgPSA0MCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcm9kX3V0aWwvMi4x', 'MPI_CC': 'mpiicc', 'PARTITION_DTN': '', 'sp_ver': '2.5.0', 'SHELLOPTS': 'braceexpand:errexit:hashall:interactive-comments:nounset:xtrace', 'DO_AWIPS': False, 'EDATE': datetime.datetime(2021, 3, 23, 12, 0), '__LMOD_REF_COUNT___INTEL_POST_CFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1'], 'I_MPI_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0', 'CRTM_FIX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2/fix', '_ModuleTable062_': 'MTYuMCIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMTYuKnpmaW5hbCIsCn0sClsicHkteGFycmF5Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHkteGFycmF5LzIwMjMuNy4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXhhcnJheS8yMDIzLjcuMCIsCmxvYWRPcmRlciA9IDgwLApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXhhcnJheS8yMDIzLjcuMCIsCndWID0gIjAwMDAwMjAyMy4wMDAwMDAwMDcuKnpm', 'proj_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez', 'FSYNC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file', '_ModuleTable005_': 'cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJjLWJsb3NjLzEuMjEuNSIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMjEuMDAwMDAwMDA1Lip6ZmluYWwiLAp9LApjZG8gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9jZG8vMi4yLjAubHVhIiwKZnVsbE5hbWUgPSAiY2RvLzIuMi4wIiwKbG9hZE9yZGVyID0gMjYsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY2RvLzIu', 'g2tmpl_ver': '1.13.0', 'FHOUT_GOES': 3, '_ModuleTable021_': 'MC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nc2wvMi43LjEubHVhIiwKZnVsbE5hbWUgPSAiZ3NsLzIuNy4xIiwKbG9hZE9yZGVyID0gMzYsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ3NsLzIuNy4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sCmhkZjUgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxl', 'EUPD_CYC': 'gdas', 'fms_ver': '2023.02.01', 'util_linux_uuid_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj', '_ModuleTable052_': 'Y2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbnVtZXhwci8yLjguNC5sdWEiLApmdWxsTmFtZSA9ICJweS1udW1leHByLzIuOC40IiwKbG9hZE9yZGVyID0gNjQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktbnVtZXhwci8yLjguNCIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDguMDAwMDAwMDA0Lip6ZmluYWwiLAp9LApbInB5LW51bXB5Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2kt', '_ModuleTable030_': 'cFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJsaWJ5YW1sLzAuMi41IiwKd1YgPSAiMDAwMDAwMDAwLjAwMDAwMDAwMi4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCm1ldCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL21ldC85LjEuMy5sdWEiLApmdWxsTmFtZSA9ICJtZXQvOS4xLjMiLApsb2FkT3JkZXIgPSA3NywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0', 'COM_ATMOS_IMAGERY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery', 'I_MPI_EXTRA_FILESYSTEM_LIST': 'lustre', 'COM_WAVE_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history', 'SLURM_CLUSTER_NAME': 'hercules', '__LMOD_REF_COUNT_ACLOCAL_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/aclocal:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/aclocal:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/aclocal:2', 'SERIAL_CC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icc', 'COM_OBSPROC_TMPL': '${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos', 'SLURM_JOB_END_TIME': 1753757358, '_ModuleTable075_': 'ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3dncmliMi8yLjAuOC5sdWEiLApmdWxsTmFtZSA9ICJ3Z3JpYjIvMi4wLjgiLApsb2FkT3JkZXIgPSA1NSwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJ3Z3JpYjIvMi4wLjgiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKemxpYiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJj', 'FHMAX_GFS': 120, 'SENDAWIP': False, 'G2_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64/libg2_d.a', 'spack_mod_path': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core', 'OCNRES': 500, 'COM_MED_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart', 'SLURM_CPUS_ON_NODE': 1, 'LMOD_FAMILY_METAMPI': 'stack-intel-oneapi-mpi', 'LEVS': 128, 'FIXgsi': '/work2/noaa/global/mterry/global-workflow_forked/fix/gsi', 'NFHRS_PER_GROUP': 3, 'COM_WAVE_PREP_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep', 'FI_PROVIDER_PATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib/prov:/usr/lib64/libfabric', 'python_ver': '3.11.6', 'IAU_OFFSET': 0, 'py_packaging_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6', 'DIAGUTIL_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/sys_check/sys_check.sh', 'python_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2', 'SCRIPTScfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/scripts', 'py_markupsafe_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7', 'G2C_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include', 'py_xlrd_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f', 'LMOD_G2TMPLVIRT_VERSION': '1.13.0', 'pid': 450170, 'OUTPUT_GRID': 'gaussian_grid', 'jobid': 'oceanice_products_ice_f102.449847', 'SLURM_JOB_CPUS_PER_NODE': 1, 'spack_env': 'gsi-addon-dev-fms-2024.01', 'INTEL_ONEAPI_MPI_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse', 'LMOD_FAMILY_METACOMPILER': 'stack-intel', 'SP_INCd': 'include_d', 'py_six_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o', 'COM_ATMOS_GOES_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim', 'MPICXX': 'mpiicpc', 'restart_interval_gdas': 3, 'fit2obs_ver': '1.1.7.1', 'LMOD_DIR': '/apps/other/lmod/lmod/libexec', '_ModuleTable068_': 'LjAvc3AvMi41LjAubHVhIiwKZnVsbE5hbWUgPSAic3AvMi41LjAiLApsb2FkT3JkZXIgPSA0MiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMiwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzcC8yLjUuMCIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCnNxbGl0ZSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9zcWxpdGUvMy40My4yLmx1YSIsCmZ1bGxOYW1lID0gInNxbGl0ZS8zLjQzLjIiLApsb2FkT3Jk', 'REPLAY_ICS': False, 'COM_ICE_NETCDF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf', 'G2_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64/libg2_4.a', '_ModuleTable058_': 'InB5LXB5dHovMjAyMy4zIiwKbG9hZE9yZGVyID0gNjcsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcHl0ei8yMDIzLjMiLAp3ViA9ICIwMDAwMDIwMjMuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LXB5eGxzYiJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXB5eGxzYi8xLjAuMTAubHVhIiwKZnVsbE5hbWUgPSAicHktcHl4bHNiLzEuMC4xMCIsCmxvYWRPcmRlciA9', 'py_setuptools_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr', 'FHMIN': 0, 'COM_ICE_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input', '_ModuleTable048_': 'cHktZjkwbm1sIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktZjkwbm1sLzEuNC4zLmx1YSIsCmZ1bGxOYW1lID0gInB5LWY5MG5tbC8xLjQuMyIsCmxvYWRPcmRlciA9IDU2LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWY5MG5tbC8xLjQuMyIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMDQuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LWppbmphMiJdID0gewpmbiA9ICIvd29yay9ub2FhL2Vw', 'intel_mkl_ver': '2023.1.0', 'PRTE_MCA_plm_slurm_args': '--external-launcher', 'PWD': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ice_f102.449847', 'SLURM_GTIDS': 0, 'LOGNAME': 'mterry', 'MAKE_ACFTBUFR': False, 'W3EMC_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_8', 'HOMEDIR': '/work2/noaa/global/mterry', 'IAU_DELTHRS': 6, 'W3EMC_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_4', 'XDG_SESSION_TYPE': 'unspecified', 'ESMFMKFILE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib/esmf.mk', 'FHMAX': 9, 'WRITE_NSFLIP': True, 'SLURM_JOB_PARTITION': 'hercules', 'YAML_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x', 'PARTITION_SERVICE': 'service', 'MODULESHOME': '/apps/other/lmod/lmod', '__LMOD_REF_COUNT_DYLD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:1', 'COM_WAVE_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart', 'BUFR_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_4.so', 'BUFR_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_8.so', 'HOMEcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1', 'COM_OCEAN_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon', 'DEBUG_POSTSCRIPT': False, 'MANPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/share/man:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/man:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/documentation/en/man/common:/apps/other/lmod/lmod/share/man:/usr/share/man:/apps/share/man:/apps/man:/opt/slurm/share/man::', 'libyaml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x', 'jasper_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt', 'py_xlsxwriter_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok', 'SLURM_TRES_PER_TASK': 'cpu=1', 'nco_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq', 'stack_intel_ver': '2021.9.0', 'SLURM_OOM_KILL_STEP': 0, 'BUFR_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m', 'bufr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m', 'CLUSTERS_DTN': '', 'grib_util_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35', 'nghttp2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky', 'DO_TEST_MODE': True, 'PACKAGEROOT': '/work2/noaa/global/role-global/nwpara', 'lobsdiag_forenkf': True, 'SLURM_JOB_NUM_NODES': 1, 'py_netcdf4_ver': '1.5.8', 'ICERES': 500, 'CXX': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icpc', 'GRIB2GRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grib2grib2', 'DMPDIR': '/work/noaa/rstprod/dump', 'LSOIL_INCR': 2, 'metplus_ver': '3.1.1', '_ModuleTable008_': 'LjAvY3J0bS1maXgvMi40LjAuMV9lbWMubHVhIiwKZnVsbE5hbWUgPSAiY3J0bS1maXgvMi40LjAuMV9lbWMiLApsb2FkT3JkZXIgPSA0NywKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJjcnRtLWZpeC8yLjQuMC4xX2VtYyIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDQuMDAwMDAwMDAwLjAwMDAwMDAwMS4qXy4qZW1jLip6ZmluYWwiLAp9LApjdXJsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIw', 'COM_OCEAN_LETKF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf', 'zlib_ver': '1.2.13', 'BUFR_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_d.so', 'ENKF_SPREAD': True, 'zlib_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j', '__LMOD_REF_COUNT_CMAKE_PREFIX_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r:2;/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/IntelDPCPP:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72:1', '_ModuleTable035_': 'c3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9uZXRjZGYtY3h4NC80LjMuMS5sdWEiLApmdWxsTmFtZSA9ICJuZXRjZGYtY3h4NC80LjMuMSIsCmxvYWRPcmRlciA9IDc2LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm5ldGNkZi1jeHg0LzQuMy4xIiwKd1YgPSAiMDAwMDAwMDA0LjAwMDAwMDAwMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsibmV0Y2RmLWZvcnRyYW4iXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFj', 'FIXgfs': '/work2/noaa/global/mterry/global-workflow_forked/fix', 'DO_COUPLED': True, 'SLURM_JOBID': 5951739, 'SERIAL_FC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'COM_ATMOS_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input', 'W3EMC_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_d', '_ModuleTable013_': 'IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nMi8zLjQuNS5sdWEiLApmdWxsTmFtZSA9ICJnMi8zLjQuNSIsCmxvYWRPcmRlciA9IDQxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImcyLzMuNC41IiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAwNC4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCmcyYyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJj', 'COM_ATMOS_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history', 'FIXam': '/work2/noaa/global/mterry/global-workflow_forked/fix/am', 'hdf5_ver': '1.14.0', 'restart_interval_enkfgdas': 3, 'COM_ATMOS_RADMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon', 'DO_CALC_INCREMENT': False, 'DO_STARTMEM_FROM_JEDIICE': False, 'FIXprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/fix', 'SLURM_JOB_QOS': 'batch', 'I_MPI_HYDRA_BOOTSTRAP_EXEC_EXTRA_ARGS': '--external-launcher', 'obsproc_run_ver': '1.2.0', 'COM_ATMOS_MINMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon', 'crtm_ver': '2.4.0.1', 'FIXorog': '/work2/noaa/global/mterry/global-workflow_forked/fix/orog', 'COM_OCEAN_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history', '__LMOD_REF_COUNT_PATH': '/apps/other/globus-cli-3.35.2/bin:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r/ush:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/bin:2;/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/bin:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/bin/intel64:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/bin:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/bin:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/bin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/bin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin:1;/usr/sbin:1;/usr/bin:1;/apps/sbin:1;/apps/bin:1;/opt/slurm/bin:1;/home/gfekete/sven/bin:1', '_ModuleTable018_': 'KnpmaW5hbCIsCn0sCmdtYWtlID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2dtYWtlLzQuMi4xLmx1YSIsCmZ1bGxOYW1lID0gImdtYWtlLzQuMi4xIiwKbG9hZE9yZGVyID0gNTQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ21ha2UvNC4yLjEiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDAyLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJncmliLXV0aWwiXSA9IHsKZm4gPSAiL3dv', 'COMIN_ICE_HISTORY': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/ice/history', 'COMROOT': '/work2/noaa/global/mterry/RUNTESTS/COMROOT', 'HOME': '/home/mterry', 'FI_PROVIDER': 'mlx', 'met_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b', 'COM_RTOFS_TMPL': '${DMPDIR}', '_ModuleTable015_': 'bnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2cydG1wbC8xLjEzLjAubHVhIiwKZnVsbE5hbWUgPSAiZzJ0bXBsLzEuMTMuMCIsCmxvYWRPcmRlciA9IDQ1LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImcydG1wbC8xLjEzLjAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDEzLip6ZmluYWwiLAp9LApnZXR0ZXh0ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2dldHRleHQvMC4yMS4xLmx1YSIsCmZ1', '_ModuleTable_Sz_': 81, 'LANG': 'C.UTF-8', '__LMOD_REF_COUNT_LIBRARY_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:1', 'DO_BUFRSND': False, 'memory': '96GB', '_ModuleTable064_': 'PSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHkteGxzeHdyaXRlci8zLjEuNy5sdWEiLApmdWxsTmFtZSA9ICJweS14bHN4d3JpdGVyLzMuMS43IiwKbG9hZE9yZGVyID0gNzAsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHkteGxzeHdyaXRlci8zLjEuNyIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDA3Lip6ZmluYWwiLAp9LApbInB5LXhsd3QiXSA9IHsKZm4gPSAi', 'DYLD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib', 'PARTITION_BATCH': 'hercules', '_ModuleTable023_': 'cy8yMDIzLjEuMCIsCmxvYWRPcmRlciA9IDIsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiaW50ZWwtb25lYXBpLWNvbXBpbGVycy8yMDIzLjEuMCIsCndWID0gIjAwMDAwMjAyMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsiaW50ZWwtb25lYXBpLW1rbCJdID0gewpmbiA9ICIvYXBwcy9zcGFjay1tYW5hZ2VkL21vZHVsZWZpbGVzL2xpbnV4LXJvY2t5OS14ODZfNjQvQ29yZS9pbnRlbC1vbmVhcGktbWtsLzIwMjMuMS4wLmx1YSIsCmZ1bGxOYW1lID0gImludGVsLW9uZWFwaS1ta2wvMjAyMy4xLjAiLApsb2FkT3JkZXIgPSA2LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVz', '_ModuleTable081_': 'Ii9hcHBzL2NvbnRyaWIvbW9kdWxlZmlsZXMiLAp9LApzeXN0ZW1CYXNlTVBBVEggPSAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L0NvcmU6L2FwcHMvb3RoZXIvbW9kdWxlZmlsZXM6L2FwcHMvY29udGFpbmVycy9tb2R1bGVmaWxlczovYXBwcy9saWNlbnNlZC9tb2R1bGVmaWxlcyIsCn0K', 'REDERR': '2>', 'PSLOT': 'C48_S2SW', '__LMOD_REF_COUNT_PKG_CONFIG_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib/pkgconfig:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/pkgconfig:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib/pkgconfig:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/pkgconfig:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib/pkgconfig:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/lib/pkgconfig:1', '_ModuleTable036_': 'ay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL25ldGNkZi1mb3J0cmFuLzQuNi4xLmx1YSIsCmZ1bGxOYW1lID0gIm5ldGNkZi1mb3J0cmFuLzQuNi4xIiwKbG9hZE9yZGVyID0gMzEsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAibmV0Y2RmLWZvcnRyYW4vNC42LjEiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDA2LjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKbmdodHRwMiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFj', 'COM_ATMOS_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}', 'USHcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/ush', 'DO_GEMPAK': False, 'SLURM_PROCID': 0, 'USHgfs': '/work2/noaa/global/mterry/global-workflow_forked/ush', 'DOIBP_WAV': False, 'prepobs_run_ver': '1.1.0', 'EXPDIR': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', 'job': 'oceanice_products', 'imp_physics': 8, 'LMOD_SHELL_PRGM': 'bash', 'IAU_FHROT': 0, 'MPIF90': 'mpiifort', 'IP_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_4.a', 'waveGRD': 'uglo_100km', 'COM_CHEM_BMAT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix', 'jasper_ver': '2.0.32', 'SCRATCH': '/scratch/hercules', 'IP_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_8.a', 'CLIENT_GLOBUS_UUID': '869912fe-f6de-46c0-af10-b22efd84a022', 'ACCOUNT': 'fv3-cpu', 'HOMEgfs': '/work2/noaa/global/mterry/global-workflow_forked', 'NHOUR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/nhour', 'binary_diag': False, 'COM_WAVE_GRID_RES_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}', 'SCRIPTSprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/scripts', 'MP_STDOUTMODE': 'ORDERED', 'sigio_ver': '2.3.2', '__LMOD_REF_COUNT_CPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/include:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/include:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/include:1', 'LMOD_SETTARG_FULL_SUPPORT': False, 'COMINsyn': '/work2/noaa/global/role-global/com/gfs/prod/syndat', 'OFFSET_START_HOUR': 0, 'REDOUT': '1>', 'PTMP': '/work2/noaa/stmp/mterry/HERCULES', '_ModuleTable076_': 'dWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC96bGliLzEuMi4xMy5sdWEiLApmdWxsTmFtZSA9ICJ6bGliLzEuMi4xMyIsCmxvYWRPcmRlciA9IDcsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEwLApzdGFja0RlcHRoID0gNSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInpsaWIvMS4yLjEzIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMi4wMDAwMDAwMTMuKnpmaW5hbCIsCn0sCnpzdGQgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2', 'g2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato', 'MODE': 'forecast-only', 'DATAROOT': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312', 'bacio_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj', 'APP': 'S2SW', '_ModuleTable029_': 'LzQuNC4zNSIsCmxvYWRPcmRlciA9IDEyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImxpYnhjcnlwdC80LjQuMzUiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDA0LjAwMDAwMDAzNS4qemZpbmFsIiwKfSwKbGlieWFtbCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9saWJ5YW1sLzAuMi41Lmx1YSIsCmZ1bGxOYW1lID0gImxpYnlhbWwvMC4yLjUiLApsb2FkT3JkZXIgPSA1OSwKcHJv', 'openblas_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f', 'stack_impi_ver': '2021.9.0', 'COM_CHEM_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem', 'TMPDIR': '/local/scratch/mterry/5951739', 'HDF5_PLUGIN_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/plugins', 'DO_TRACKER': True, 'CMAKE_PREFIX_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r:/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/IntelDPCPP:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72', 'crtm_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r', '_ModuleTable038_': 'YWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvb3BlbmJsYXMvMC4zLjI0Lmx1YSIsCmZ1bGxOYW1lID0gIm9wZW5ibGFzLzAuMy4yNCIsCmxvYWRPcmRlciA9IDUwLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm9wZW5ibGFzLzAuMy4yNCIsCndWID0gIjAwMDAwMDAwMC4wMDAwMDAwMDMuMDAwMDAwMDI0Lip6ZmluYWwiLAp9LApvcGVuanBlZyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5z', 'nco_ver': '5.0.6', '_ModuleTable046_': 'IiwKfSwKWyJweS1jZnRpbWUiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1jZnRpbWUvMS4wLjMuNC5sdWEiLApmdWxsTmFtZSA9ICJweS1jZnRpbWUvMS4wLjMuNCIsCmxvYWRPcmRlciA9IDU3LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWNmdGltZS8xLjAuMy40IiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMC4wMDAwMDAwMDMuMDAwMDAwMDA0Lip6ZmluYWwiLAp9', 'GRBINDEX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grbindex', 'libjpeg_turbo_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy', 'MPIF77': 'mpiifort', '_ModuleTable073_': 'LjAwMDAwMDAwMi4wMDAwMDAwMjguKnpmaW5hbCIsCn0sClsidXRpbC1saW51eC11dWlkIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvdXRpbC1saW51eC11dWlkLzIuMzguMS5sdWEiLApmdWxsTmFtZSA9ICJ1dGlsLWxpbnV4LXV1aWQvMi4zOC4xIiwKbG9hZE9yZGVyID0gMTQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAidXRpbC1saW51eC11dWlkLzIuMzguMSIsCndWID0gIjAw', '_ModuleTable010_': 'Mi4wIiwKbG9hZE9yZGVyID0gMjAsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZWNjb2Rlcy8yLjMyLjAiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDMyLip6ZmluYWwiLAp9LAplc21mID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAvZXNtZi84LjYuMC5sdWEiLApmdWxsTmFtZSA9ICJlc21mLzguNi4wIiwKbG9hZE9yZGVyID0gMzQsCnBy', 'NTHSTACK': 1024000000, 'SLURM_CPUS_PER_TASK': 1, 'FIXcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/fix', 'SLURM_NTASKS': 1, 'DO_ICE': True, 'NET': 'gfs', 'COM_ATMOS_GENESIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital', 'HOMEpost': '/work2/noaa/global/mterry/global-workflow_forked', '_ModuleTable061_': 'CnVzZXJOYW1lID0gInB5LXNldHVwdG9vbHMvNjMuNC4zIiwKd1YgPSAiMDAwMDAwMDYzLjAwMDAwMDAwNC4wMDAwMDAwMDMuKnpmaW5hbCIsCn0sClsicHktc2l4Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktc2l4LzEuMTYuMC5sdWEiLApmdWxsTmFtZSA9ICJweS1zaXgvMS4xNi4wIiwKbG9hZE9yZGVyID0gNzMsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktc2l4LzEu', 'bacio_ver': '2.4.1', 'cmake_ver': '3.23.1', 'UTILROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq', '_ModuleTable024_': 'ZXJOYW1lID0gImludGVsLW9uZWFwaS1ta2wvMjAyMy4xLjAiLAp3ViA9ICIwMDAwMDIwMjMuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbImludGVsLW9uZWFwaS1tcGkiXSA9IHsKZm4gPSAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L29uZWFwaS8yMDIzLjEuMC9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLmx1YSIsCmZ1bGxOYW1lID0gImludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAiLApsb2FkT3JkZXIgPSA0LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAiLAp3ViA9ICIwMDAwMDIwMjEuMDAwMDAwMDA5Lip6ZmluYWwi', 'tar_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths', 'ntasks': 1, 'max_tasks_per_node': 80, 'NCDUMP': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump', '_ModuleTable057_': 'LjIubHVhIiwKZnVsbE5hbWUgPSAicHktcHl0aG9uLWRhdGV1dGlsLzIuOC4yIiwKbG9hZE9yZGVyID0gNzQsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcHl0aG9uLWRhdGV1dGlsLzIuOC4yIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwOC4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sClsicHktcHl0eiJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXB5dHovMjAyMy4zLmx1YSIsCmZ1bGxOYW1lID0g', 'py_numpy_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne', 'QUILTING': True, 'IP_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_d.a', 'SLURM_TOPOLOGY_ADDR': 'hercules-08-55', 'libxcrypt_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr', 'CRTM_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include', 'LMOD_VERSION': '8.7.14', 'spack_stack_ver': '1.6.0', '_ModuleTable041_': 'LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9wYXJhbGxlbGlvLzIuNS4xMC5sdWEiLApmdWxsTmFtZSA9ICJwYXJhbGxlbGlvLzIuNS4xMCIsCmxvYWRPcmRlciA9IDMzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInBhcmFsbGVsaW8vMi41LjEwIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNS4wMDAwMDAwMTAuKnpmaW5hbCIsCn0sCnBpZ3ogPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0y', 'HOMEprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0', 'COMPONENT': 'ice', 'BASE_GIT': '/work2/noaa/global/role-global/git', 'FCST_SEGMENTS': [0, 120], 'libpng_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu', '_ModuleTable044_': 'LjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKcHJvaiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9wcm9qLzkuMi4xLmx1YSIsCmZ1bGxOYW1lID0gInByb2ovOS4yLjEiLApsb2FkT3JkZXIgPSAyNCwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcm9qLzkuMi4xIiwKd1YgPSAiMDAwMDAwMDA5LjAwMDAwMDAwMi4wMDAwMDAwMDEuKnpm', 'DO_GENESIS': True, 'BUFR_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_4', '__LMOD_REF_COUNT_DIAGUTIL_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/sys_check/sys_check.sh:1', 'COM_WAVE_GEMPAK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak', 'COM_ATMOS_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2', 'RUN': 'gfs', 'BUFR_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_8', 'COM_ICE_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon', 'py_pandas_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw', '__INTEL_POST_CFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64'], 'ARCDIR': '/work2/noaa/global/mterry/archive/C48_S2SW', '__LMOD_REF_COUNT_NLSPATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64/locale/%l_%t/%N:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin/locale/%l_%t/%N:1', '_ModuleTable069_': 'ZXIgPSAxMywKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMiwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzcWxpdGUvMy40My4yIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDA0My4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sClsic3RhY2staW50ZWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9Db3JlL3N0YWNrLWludGVsLzIwMjEuOS4wLmx1YSIsCmZ1bGxOYW1lID0gInN0YWNrLWludGVsLzIwMjEuOS4wIiwKbG9hZE9yZGVyID0gMywKcHJvcFQgPSB7fSwK', 'SENDECF': False, 'parallelio_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed', 'py_pytz_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s', 'ILPOST': 3, 'WGRIB2_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib/libwgrib2.a', 'parallel_netcdf_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3', 'CMPLR_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0', 'prod_util_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq', 'COMINukmet': '/work2/noaa/global/role-global/data/external_gempak/ukmet', 'libpng_ver': '1.6.37', 'COPYGB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/copygb2', 'NMV': '/bin/mv', 'NOSCRUB': '/work2/noaa/global/mterry', 'SCRATCH_hercules': '/scratch/hercules', 'W3EMC_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_d.a', 'SENDSDM': False, 'envir': 'prod', 'DO_METP': False, 'CRTM_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib/libcrtm.a', 'LOGSCRIPT': '', 'INTEL_ONEAPI_COMPILERS_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72', '_ModuleTable074_': 'MDAwMDAwMi4wMDAwMDAwMzguMDAwMDAwMDAxLip6ZmluYWwiLAp9LAp3M2VtYyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC93M2VtYy8yLjEwLjAubHVhIiwKZnVsbE5hbWUgPSAidzNlbWMvMi4xMC4wIiwKbG9hZE9yZGVyID0gMzksCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAidzNlbWMvMi4xMC4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKd2dyaWIy', 'py_jinja2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj', 'FPGA_VARS_ARGS': '', 'COMOUT_ICE_NETCDF': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ice/netcdf', 'INCVARS_ZERO_STRAT': ["'sphum_inc'", "'liq_wat_inc'", "'icmr_inc'", "'rwmr_inc'", "'snmr_inc'", "'grle_inc'"], 'COM_CONF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf', '_ModuleTable003_': 'YSIsCmZ1bGxOYW1lID0gImJhY2lvLzIuNC4xIiwKbG9hZE9yZGVyID0gMzgsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAzLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiYmFjaW8vMi40LjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDA0LjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKYnVmciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9idWZyLzExLjcuMC5sdWEiLApmdWxsTmFtZSA9ICJidWZyLzExLjcuMCIsCmxvYWRPcmRlciA9', 'HYDRA_BOOTSTRAP': 'slurm', '__LMOD_REF_COUNT___INTEL_POST_FFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1'], 'COM_CHEM_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon', 'MDATE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/mdate', 'WGRIB2': 'wgrib2', '_ModuleTable054_': 'LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktb3BlbnB5eGwvMy4xLjIubHVhIiwKZnVsbE5hbWUgPSAicHktb3BlbnB5eGwvMy4xLjIiLApsb2FkT3JkZXIgPSA2NiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1vcGVucHl4bC8zLjEuMiIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbInB5LXBhY2thZ2luZyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxs', 'EXP_WARM_START': False, 'MODULEPATH_ROOT': '/apps/other/modulefiles', 'COM_SNOW_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon', 'CHGRP_RSTPROD': True, 'py_pyxlsb_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge', 'BACIO_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/include_8', 'restart_interval_gfs': 12, 'DO_GOES': False, 'pgm': '', 'SLURM_TOPOLOGY_ADDR_PATTERN': 'node', 'DO_LAND_IAU': False, 'ntiles': 6, 'FHCYC': 24, 'ROTDIR': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW', 'FHOUT_AERO': 3, 'py_jinja2_ver': '3.1.2', 'BACIO_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/include_4', 'INCREMENTS_TO_ZERO': ["'liq_wat_inc'", "'icmr_inc'", "'rwmr_inc'", "'snmr_inc'", "'grle_inc'"], 'FHR_LIST': 102, 'USHprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/ush', 'BUFR_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_d', 'hdf5_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh', 'INTEL_ONEAPI_MKL_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci', 'SCRIPTSfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/scripts', '__LMOD_REF_COUNT_FI_PROVIDER_PATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib/prov:1;/usr/lib64/libfabric:1', 'FIXcice': '/work2/noaa/global/mterry/global-workflow_forked/fix/cice', 'IAUFHRS': [6], 'XDG_SESSION_CLASS': 'background', 'INTERVAL_GFS': 6, 'py_openpyxl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe', 'FHMAX_FITS': 120, 'COM_ATMOS_MASTER_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master', 'LMOD_PKG': '/apps/other/lmod/lmod', 'MPI_MEMMAP_OFF': 1, 'UUID_HERCULES_DTN': '869912fe-f6de-46c0-af10-b22efd84a022', '_ModuleTable079_': 'L3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAiCiwgIi9hcHBzL3NwYWNrLW1hbmFnZWQvbW9kdWxlZmlsZXMvbGludXgtcm9ja3k5LXg4Nl82NC9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLWE2NmVhaXAvb25lYXBpLzIwMjMuMS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wIgosICIvYXBwcy9zcGFjay1tYW5hZ2VkL21vZHVsZWZp', 'SDATE': datetime.datetime(2021, 3, 23, 12, 0), 'SLURM_SCRIPT_CONTEXT': 'prolog_task', 'MPI_CXX': 'mpiicpc', 'lwrite4danl': True, 'SLURM_MEM_PER_NODE': 98304, 'CASE_ENS': '{{ CASE_ENS }}', '_ModuleTable049_': 'aWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LWppbmphMi8zLjEuMi5sdWEiLApmdWxsTmFtZSA9ICJweS1qaW5qYTIvMy4xLjIiLApsb2FkT3JkZXIgPSA2MiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1qaW5qYTIvMy4xLjIiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDAxLjAwMDAwMDAwMi4qemZpbmFsIiwKfSwKWyJweS1tYXJrdXBzYWZlIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3Vs', 'YAML_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include', 'PYTHONPATH': '/apps/other/globus-cli-3.35.2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:/work2/noaa/global/mterry/global-workflow_forked/sorc/wxflow/src:/work2/noaa/global/mterry/global-workflow_forked/ush/python', '__LMOD_REF_COUNT_PYTHONPATH': '/apps/other/globus-cli-3.35.2:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:3;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:1', 'py_xarray_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4', 'nst_anl': True, 'FHOUT_GFS': 3, 'WORK': '/work/hercules', 'W3EMC_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_4.a', 'F77': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'BASE_DATA': '/work2/noaa/global/role-global/data', 'py_xlwt_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56', 'ACLOCAL_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/aclocal:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/aclocal:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/aclocal', 'QUEUE_SERVICE': 'batch', 'W3EMC_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_8.a', 'DATA': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ice_f102.449847', 'LESSOPEN': '||/usr/bin/lesspipe.sh %s', 'DO_VERFOZN': True, 'antlr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2', 'crtm_fix_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2', 'TOCGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib', 'USER': 'mterry', 'HPSS_PROJECT': 'emc-global', 'FHMAX_WAV_GFS': 120, 'FIXgdas': '/work2/noaa/global/mterry/global-workflow_forked/fix/gdas', 'launcher': 'srun -l --export=ALL --hint=nomultithread', 'NDATE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/ndate', 'LIBRARY_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib', 'gsl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z', 'SLURM_NODELIST': 'hercules-08-55', 'DO_FIT2OBS': True, 'IP_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_4', 'DOIAU_ENKF': True, 'COM_ATMOS_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos', 'py_xarray_ver': '2023.7.0', '__INTEL_POST_FFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64'], 'netcdf_diag': True, 'DOBNDPNT_WAVE': True, 'FETCHDIR': '/NCEPDEV/emc-global/1year/David.Grumm/test_data', 'ENVIRONMENT': 'BATCH', 'gsi_ncdiag_ver': '1.1.2', 'IP_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_8', 'py_bottleneck_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a', 'w3emc_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo', 'py_netcdf4_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i', 'wgrib2_ver': '2.0.8', 'TOCGRIB2SUPER': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib2super', 'COM_ICE_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2', 'KEEPDATA': False, 'udunits_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo', 'cdo_ver': '2.2.0', 'COM_OCEAN_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input', 'FNTSFA': ' ', 'LOADEDMODULES': 'contrib/0.1:intel-oneapi-compilers/2023.1.0:stack-intel/2021.9.0:intel-oneapi-mpi/2021.9.0:stack-intel-oneapi-mpi/2021.9.0:intel-oneapi-mkl/2023.1.0:zlib/1.2.13:pigz/2.7:zstd/1.5.2:tar/1.34:gettext/0.21.1:libxcrypt/4.4.35:sqlite/3.43.2:util-linux-uuid/2.38.1:python/3.11.6:libjpeg/2.1.0:jasper/2.0.32:libpng/1.6.37:openjpeg/2.3.1:eccodes/2.32.0:fftw/3.3.10:nghttp2/1.57.0:curl/8.4.0:proj/9.2.1:udunits/2.2.28:cdo/2.2.0:hdf5/1.14.0:snappy/1.1.10:c-blosc/1.21.5:netcdf-c/4.9.2:netcdf-fortran/4.6.1:parallel-netcdf/1.12.2:parallelio/2.5.10:esmf/8.6.0:antlr/2.7.7:gsl/2.7.1:nco/5.0.6:bacio/2.4.1:w3emc/2.10.0:prod_util/2.1.1:g2/3.4.5:sp/2.5.0:ip/4.3.0:grib-util/1.3.0:g2tmpl/1.13.0:gsi-ncdiag/1.1.2:crtm-fix/2.4.0.1_emc:git-lfs/3.1.2:crtm/2.4.0.1:openblas/0.3.24:py-setuptools/63.4.3:py-numpy/1.23.4:bufr/11.7.0:gmake/4.2.1:wgrib2/2.0.8:py-f90nml/1.4.3:py-cftime/1.0.3.4:py-netcdf4/1.5.8:libyaml/0.2.5:py-pyyaml/6.0:py-markupsafe/2.1.3:py-jinja2/3.1.2:py-bottleneck/1.3.7:py-numexpr/2.8.4:py-et-xmlfile/1.0.1:py-openpyxl/3.1.2:py-pytz/2023.3:py-pyxlsb/1.0.10:py-xlrd/2.0.1:py-xlsxwriter/3.1.7:py-xlwt/1.3.0:py-pandas/1.5.3:py-six/1.16.0:py-python-dateutil/2.8.2:g2c/1.8.0:netcdf-cxx4/4.3.1:met/9.1.3:metplus/3.1.1:py-packaging/23.1:py-xarray/2023.7.0:prepobs/1.1.0:fit2obs/1.1.7.1:globus-cli/3.35.2:module_base.hercules', 'SLURM_JOB_ACCOUNT': 'fv3-cpu', 'DO_OCN': True, 'SLURM_PRIO_PROCESS': 0, 'HOMEfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1', 'gmake_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq', 'FIXfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/fix', 'py_python_dateutil_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy', 'tasks_per_node': 1, 'SLURM_NPROCS': 1, 'COM_CHEM_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history', 'LMOD_ROOT': '/apps/other/lmod', 'DOHYBVAR': '{{ DOHYBVAR }}', 'GSL_ROOT_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z', 'SERIAL_F77': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'IP_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_d', 'SHLVL': 4, '_ModuleTable071_': 'e30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAic3RhY2staW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMCIsCndWID0gIjAwMDAwMjAyMS4wMDAwMDAwMDkuKnpmaW5hbCIsCn0sCnRhciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC90YXIvMS4zNC5sdWEiLApmdWxsTmFtZSA9ICJ0YXIvMS4zNCIsCmxvYWRPcmRlciA9IDEwLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIs', '_ModuleTable012_': 'c3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImZmdHcvMy4zLjEwIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAwMy4wMDAwMDAwMTAuKnpmaW5hbCIsCn0sCmZpdDJvYnMgPSB7CmZuID0gIi93b3JrMi9ub2FhL2dsb2JhbC9yb2xlLWdsb2JhbC9naXQvRml0Mk9icy92MS4xLjcuMS9tb2R1bGVmaWxlcy9maXQyb2JzLzEuMS43LjEubHVhIiwKZnVsbE5hbWUgPSAiZml0Mm9icy8xLjEuNy4xIiwKbG9hZE9yZGVyID0gODIsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZml0Mm9icy8xLjEuNy4xIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4wMDAwMDAwMDcuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApnMiA9', 'INCVARS_EFOLD': 5, 'SLURM_NNODES': 1, 'RESERVATION': '', 'FHMIN_GFS': 0, 'FHMAX_HF_WAV': 48, 'FIXugwd': '/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd', 'BASH_ENV': '/apps/other/lmod/lmod/init/bash', 'CDATE': datetime.datetime(2021, 3, 23, 12, 0), '_ModuleTable063_': 'aW5hbCIsCn0sClsicHkteGxyZCJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXhscmQvMi4wLjEubHVhIiwKZnVsbE5hbWUgPSAicHkteGxyZC8yLjAuMSIsCmxvYWRPcmRlciA9IDY5LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXhscmQvMi4wLjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJweS14bHN4d3JpdGVyIl0g', '_ModuleTable006_': 'Mi4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMi4qemZpbmFsIiwKfSwKY29udHJpYiA9IHsKZm4gPSAiL2FwcHMvb3RoZXIvbW9kdWxlZmlsZXMvY29udHJpYi8wLjEiLApmdWxsTmFtZSA9ICJjb250cmliLzAuMSIsCmxvYWRPcmRlciA9IDEsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAwLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY29udHJpYiIsCndWID0gIjAwMDAwMDAwMC4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sCmNydG0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50', 'FHOUT_OCN_GFS': 6, '_ModuleTable026_': 'cmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2phc3Blci8yLjAuMzIubHVhIiwKZnVsbE5hbWUgPSAiamFzcGVyLzIuMC4zMiIsCmxvYWRPcmRlciA9IDE3LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImphc3Blci8yLjAuMzIiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAzMi4qemZpbmFsIiwKfSwKbGlianBlZyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1z', 'DONST': True, '_ModuleTable051_': 'ay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL3B5LW5ldGNkZjQvMS41LjgubHVhIiwKZnVsbE5hbWUgPSAicHktbmV0Y2RmNC8xLjUuOCIsCmxvYWRPcmRlciA9IDU4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LW5ldGNkZjQvMS41LjgiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA1LjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKWyJweS1udW1leHByIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3Rh', 'LMOD_sys': 'Linux', '_ModuleTable033_': 'L3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL25jby81LjAuNi5sdWEiLApmdWxsTmFtZSA9ICJuY28vNS4wLjYiLApsb2FkT3JkZXIgPSAzNywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJuY28vNS4wLjYiLAp3ViA9ICIwMDAwMDAwMDUuMDAwMDAwMDAwLjAwMDAwMDAwNi4qemZpbmFsIiwKfSwKWyJuZXRjZGYtYyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1', 'FHMIN_WAV': 0, 'LMOD_FAMILY_METACOMPILER_VERSION': '2021.9.0', 'I_MPI_PMI_LIBRARY': '/opt/slurm/lib/libpmi2.so', 'machine': 'HERCULES', 'py_pyyaml_ver': 6.0, 'OCL_ICD_FILENAMES': 'libintelocl_emu.so:libalteracl.so:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64/libintelocl.so', '_ModuleTable017_': 'bG9hZE9yZGVyID0gNDgsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ2l0LWxmcy8zLjEuMiIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbImdsb2J1cy1jbGkiXSA9IHsKZm4gPSAiL2FwcHMvb3RoZXIvbW9kdWxlZmlsZXMvZ2xvYnVzLWNsaS8zLjM1LjIubHVhIiwKZnVsbE5hbWUgPSAiZ2xvYnVzLWNsaS8zLjM1LjIiLApsb2FkT3JkZXIgPSA4MywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJnbG9idXMtY2xpIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAzNS4wMDAwMDAwMDIu', 'DO_FETCH_HPSS': False, 'PROJ_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/proj', 'FHOUT_ICE': 3, 'COM_WAVE_STATION_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station', 'COMINecmwf': '/work2/noaa/global/role-global/data/external_gempak/ecmwf', 'COM_TOP_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}', 'WGRIB2_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include', 'mpmd_opt': '--multi-prog --output=mpmd.%j.%t.out', 'DO_PREP_SFC': False, 'LMOD_G2TMPLVIRT_NAME': 'g2tmpl', '_ModuleTable066_': 'bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5dGhvbi8zLjExLjYubHVhIiwKZnVsbE5hbWUgPSAicHl0aG9uLzMuMTEuNiIsCmxvYWRPcmRlciA9IDE1LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5dGhvbi8zLjExLjYiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDExLjAwMDAwMDAwNi4qemZpbmFsIiwKfSwKc25hcHB5ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAy', '_ModuleTable009_': 'MjEuOS4wL2N1cmwvOC40LjAubHVhIiwKZnVsbE5hbWUgPSAiY3VybC84LjQuMCIsCmxvYWRPcmRlciA9IDIzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImN1cmwvOC40LjAiLAp3ViA9ICIwMDAwMDAwMDguMDAwMDAwMDA0Lip6ZmluYWwiLAp9LAplY2NvZGVzID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2VjY29kZXMvMi4zMi4wLmx1YSIsCmZ1bGxOYW1lID0gImVjY29kZXMvMi4z', 'py_numexpr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq', '__LMOD_REF_COUNT_MANPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/share/man:2;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/man:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/documentation/en/man/common:1;/apps/other/lmod/lmod/share/man:1;/usr/share/man:1;/apps/share/man:1;/apps/man:1;/opt/slurm/share/man:1', 'DOLETKF_OCN': False, 'MPI_GROUP_MAX': 256, 'py_cftime_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i', 'XDG_SESSION_ID': 'c21', 'zstd_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4', 'CLUSTERS': '', '_ModuleTable034_': 'bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAvbmV0Y2RmLWMvNC45LjIubHVhIiwKZnVsbE5hbWUgPSAibmV0Y2RmLWMvNC45LjIiLApsb2FkT3JkZXIgPSAzMCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJuZXRjZGYtYy80LjkuMiIsCndWID0gIjAwMDAwMDAwNC4wMDAwMDAwMDkuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbIm5ldGNkZi1jeHg0Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2st', 'COM_OCEAN_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean', 'CLUSTERS_SERVICE': '', 'pgmout': 'OUTPUT.450170', 'EXECfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/exec', '_ModuleTable031_': 'aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJtZXQvOS4xLjMiLAp3ViA9ICIwMDAwMDAwMDkuMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKbWV0cGx1cyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL21ldHBsdXMvMy4xLjEubHVhIiwKZnVsbE5hbWUgPSAibWV0cGx1cy8zLjEuMSIsCmxvYWRPcmRlciA9IDc4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIs', 'COMOUT_ICE_GRIB': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ice/grib2', 'COM_ATMOS_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon', 'HOMEobsproc': '/work2/noaa/global/role-global/git/obsproc/v1.2.0', 'esmf_ver': '8.6.0', 'DO_ARCHCOM': False, 'SLURM_SUBMIT_HOST': 'hercules-login-1.hpc.msstate.edu', 'WRITE_DOPOST': True, 'EXECcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/exec', 'DUMP_SUFFIX': '', '_ModuleTable072_': 'CnVzZXJOYW1lID0gInRhci8xLjM0IiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAzNC4qemZpbmFsIiwKfSwKdWR1bml0cyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC91ZHVuaXRzLzIuMi4yOC5sdWEiLApmdWxsTmFtZSA9ICJ1ZHVuaXRzLzIuMi4yOCIsCmxvYWRPcmRlciA9IDI1LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInVkdW5pdHMvMi4yLjI4IiwKd1YgPSAiMDAwMDAwMDAy', 'CASE': 'C48', 'SENDDBN': False, 'gettext_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43', 'COM_ICE_BMATRIX_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice', 'git_lfs_ROOT': '/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l', '_ModuleTable001_': 'X01vZHVsZVRhYmxlXyA9IHsKTVR2ZXJzaW9uID0gMywKY19yZWJ1aWxkVGltZSA9IGZhbHNlLApjX3Nob3J0VGltZSA9IGZhbHNlLApkZXB0aFQgPSB7fSwKZmFtaWx5ID0gewpNZXRhQ29tcGlsZXIgPSAic3RhY2staW50ZWwiLApNZXRhTVBJID0gInN0YWNrLWludGVsLW9uZWFwaS1tcGkiLApjb21waWxlciA9ICJpbnRlbC1vbmVhcGktY29tcGlsZXJzIiwKZzJ0bXBsdmlydCA9ICJnMnRtcGwiLAptcGkgPSAiaW50ZWwtb25lYXBpLW1waSIsCn0sCm1UID0gewphbnRsciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9t', 'COM_ICE_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart', '_ModuleTable025_': 'LAp9LAppcCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9pcC80LjMuMC5sdWEiLApmdWxsTmFtZSA9ICJpcC80LjMuMCIsCmxvYWRPcmRlciA9IDQzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImlwLzQuMy4wIiwKd1YgPSAiMDAwMDAwMDA0LjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKamFzcGVyID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hl', 'COM_OCEAN_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart', 'ARCHCOM_TO': 'globus_hpss', 'COM_ATMOS_WMO_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo', 'ip_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf', 'CLASSPATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/mpi.jar', 'INTELFPGAOCLSDKROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga', 'LD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/host/linux64/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/lib:/usr/lib64:/usr/lib:/opt/slurm/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib', 'DO_CA': True, 'g2_ver': '3.4.5', 'LMOD_FAMILY_COMPILER': 'intel-oneapi-compilers', 'XDG_RUNTIME_DIR': '/run/user/9583', 'BASE_ENV': '/work2/noaa/global/mterry/global-workflow_forked/env', 'SLURM_JOB_ID': 5951739, '_ModuleTable014_': 'dWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nMmMvMS44LjAubHVhIiwKZnVsbE5hbWUgPSAiZzJjLzEuOC4wIiwKbG9hZE9yZGVyID0gNzUsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZzJjLzEuOC4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKZzJ0bXBsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9p', 'OCEANICEPRODUCTS_CONFIG': '/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml', 'NTHREADS_OCNICEPOST': 1, 'DO_JEDIOCNVAR': False, 'FIXreg2grb2': '/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2', 'SLURM_NODEID': 0, 'ip_ver': '4.3.0', 'KMP_AFFINITY': 'scatter', 'FHOUT_ICE_GFS': 6, '_ModuleTable065_': 'L3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS14bHd0LzEuMy4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXhsd3QvMS4zLjAiLApsb2FkT3JkZXIgPSA3MSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS14bHd0LzEuMy4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKcHl0aG9uID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1', 'DO_ATM': True, '_ModuleTable004_': 'IDUzLApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImJ1ZnIvMTEuNy4wIiwKd1YgPSAiMDAwMDAwMDExLjAwMDAwMDAwNy4qemZpbmFsIiwKfSwKWyJjLWJsb3NjIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvYy1ibG9zYy8xLjIxLjUubHVhIiwKZnVsbE5hbWUgPSAiYy1ibG9zYy8xLjIxLjUiLApsb2FkT3JkZXIgPSAyOSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1', 'esmf_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep', 'LMOD_FAMILY_MPI_VERSION': '2021.9.0', '_ModuleTable020_': 'c3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9nc2ktbmNkaWFnLzEuMS4yLmx1YSIsCmZ1bGxOYW1lID0gImdzaS1uY2RpYWcvMS4xLjIiLApsb2FkT3JkZXIgPSA0NiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJnc2ktbmNkaWFnLzEuMS4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCmdzbCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYu', 'DO_JEDIATMVAR': False, 'DO_AERO_ANL': False, 'NCLEN': '/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen', 'PS4': '+ $(basename ${BASH_SOURCE[0]:-${FUNCNAME[0]:-"Unknown"}})[${LINENO}]', 'COM_SNOW_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow', '_ModuleTable053_': 'YWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbnVtcHkvMS4yMy40Lmx1YSIsCmZ1bGxOYW1lID0gInB5LW51bXB5LzEuMjMuNCIsCmxvYWRPcmRlciA9IDUyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSA4LApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LW51bXB5LzEuMjMuNCIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMjMuMDAwMDAwMDA0Lip6ZmluYWwiLAp9LApbInB5LW9wZW5weXhsIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0', '__LMOD_REF_COUNT_ACL_BOARD_VENDOR_PATH': '/opt/Intel/OpenCLFPGA/oneAPI/Boards:1', 'LMOD_FAMILY_METAMPI_VERSION': '2021.9.0', '_ModuleTable059_': 'IDY4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXB5eGxzYi8xLjAuMTAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAwLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKWyJweS1weXlhbWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1weXlhbWwvNi4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXB5eWFtbC82LjAiLApsb2FkT3JkZXIgPSA2MCwKcHJvcFQgPSB7fSwKc3RhY2tE', 'DO_CALC_INCREMENT_ENKF_GFS': False, 'MKLROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0', 'DEBUGINFOD_URLS': 'https://debuginfod.centos.org/ ', 'sven_root_path': '/home/gfekete/sven', '_ModuleTable047_': 'LApbInB5LWV0LXhtbGZpbGUiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1ldC14bWxmaWxlLzEuMC4xLmx1YSIsCmZ1bGxOYW1lID0gInB5LWV0LXhtbGZpbGUvMS4wLjEiLApsb2FkT3JkZXIgPSA2NSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1ldC14bWxmaWxlLzEuMC4xIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMC4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsi', 'restart_interval_enkfgfs': 3, 'SENDDBN_NTC': False, 'FIXaer': '/work2/noaa/global/mterry/global-workflow_forked/fix/aer', 'FHMAX_HF_GFS': 48, 'COM_ICE_LETKF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf', 'LMOD_FAMILY_MPI': 'intel-oneapi-mpi', 'netcdf_fortran_ver': '4.6.1', 'FIXmom': '/work2/noaa/global/mterry/global-workflow_forked/fix/mom6', 'cyc': 12, 'BACIO_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib/libbacio_4.a', 'FC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'FHOUT_WAV_GFS': 3, 'SDATE_GFS': datetime.datetime(2021, 3, 23, 12, 0), 'BACIO_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib/libbacio_8.a', 'sp_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr', 'OPS_RES': 'C768', 'NMEM_ENS_GFS': 30, 'which_declare': 'declare -f', '_ModuleTable056_': 'OS4wL3B5LXBhbmRhcy8xLjUuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1wYW5kYXMvMS41LjMiLApsb2FkT3JkZXIgPSA3MiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1wYW5kYXMvMS41LjMiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA1LjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKWyJweS1weXRob24tZGF0ZXV0aWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1weXRob24tZGF0ZXV0aWwvMi44', 'COM_WAVE_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded', 'LMOD_FAMILY_G2TMPLVIRT': 'g2tmpl', 'QUEUE_DTN': 'batch', 'DO_GENESIS_FSU': False, 'prod_util_ver': '2.1.1', 'MPI_F90': 'mpiifort', 'LMOD_FAMILY_COMPILER_VERSION': '2023.1.0', '_ModuleTable042_': 'MDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcGlnei8yLjcubHVhIiwKZnVsbE5hbWUgPSAicGlnei8yLjciLApsb2FkT3JkZXIgPSA4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gNCwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInBpZ3ovMi43IiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4qemZpbmFsIiwKfSwKcHJlcG9icyA9IHsKZm4gPSAiL3dvcmsyL25vYWEvZ2xvYmFsL3JvbGUtZ2xvYmFsL2dpdC9wcmVwb2JzL3YxLjEuMC9tb2R1bGVmaWxlcy9wcmVwb2JzLzEuMS4wLmx1YSIsCmZ1bGxOYW1lID0gInByZXBvYnMvMS4xLjAiLApsb2FkT3JkZXIgPSA4MSwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0', 'openjpeg_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu', 'USHfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/ush', 'YAML_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib', 'DO_GSISOILDA': False, 'nemsio_ver': '2.5.4', 'DO_NPOESS': False, 'SLURM_MPI_TYPE': 'pmi2', 'OMP_STACKSIZE': 2048000, 'py_et_xmlfile_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq', 'FORECAST_HOUR': 102, 'ATARDIR': '/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW', 'EXECgfs': '/work2/noaa/global/mterry/global-workflow_forked/exec', 'py_python_dateutil_ver': '2.8.2', 'assim_freq': 6, 'DOIAU': True, 'FHOUT_WAV': 1, 'py_pyyaml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2', 'met_ver': '9.1.3', 'gsi_ncdiag_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe', 'DO_FETCH_LOCAL': False, 'NLSPATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64/locale/%l_%t/%N:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin/locale/%l_%t/%N', 'netcdf_fortran_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe', 'SP_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_4.a', 'COM_OCEAN_BMATRIX_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean', 'SLURM_CONF': '/var/spool/slurmd/conf-cache/slurm.conf', 'PATH': '/apps/other/globus-cli-3.35.2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r/ush:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/bin:/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/bin/intel64:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/bin:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/bin:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin:/usr/sbin:/usr/bin:/apps/sbin:/apps/bin:/opt/slurm/bin:/home/gfekete/sven/bin', 'SP_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_8.a', 'netcdf_c_ver': '4.9.2', 'SLURM_JOB_NAME': 'C48_S2SW_gfs_ice_prod_f102_12', 'COM_ICE_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history', 'NLN': '/bin/ln -sf', 'MODULEPATH': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles:/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0-a66eaip/g2tmpl/1.13.0/intel/2021.9.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/g2tmpl/1.13.0/intel/2021.9.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/intel-oneapi-mpi/2021.9.0-a66eaip/oneapi/2023.1.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core:/work2/noaa/global/mterry/global-workflow_forked/modulefiles:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core:/apps/other/modulefiles:/apps/containers/modulefiles:/apps/licensed/modulefiles:/apps/contrib/modulefiles', 'CC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icc', 'grib_util_ver': '1.3.0', 'py_f90nml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z', 'DO_MERGENSST': False, 'SLURM_NTASKS_PER_NODE': 1, '_LMFILES_': '/apps/other/modulefiles/contrib/0.1:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core/intel-oneapi-compilers/2023.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core/stack-intel/2021.9.0.lua:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0/intel-oneapi-mpi/2021.9.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/stack-intel-oneapi-mpi/2021.9.0.lua:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core/intel-oneapi-mkl/2023.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/zlib/1.2.13.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/pigz/2.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/zstd/1.5.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/tar/1.34.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gettext/0.21.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libxcrypt/4.4.35.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/sqlite/3.43.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/util-linux-uuid/2.38.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/python/3.11.6.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libjpeg/2.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/jasper/2.0.32.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libpng/1.6.37.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/openjpeg/2.3.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/eccodes/2.32.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/fftw/3.3.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/nghttp2/1.57.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/curl/8.4.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/proj/9.2.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/udunits/2.2.28.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/cdo/2.2.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/hdf5/1.14.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/snappy/1.1.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/c-blosc/1.21.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-c/4.9.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-fortran/4.6.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/parallel-netcdf/1.12.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/parallelio/2.5.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/esmf/8.6.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/antlr/2.7.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gsl/2.7.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/nco/5.0.6.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/bacio/2.4.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/w3emc/2.10.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/prod_util/2.1.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2/3.4.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/sp/2.5.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/ip/4.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/grib-util/1.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2tmpl/1.13.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/gsi-ncdiag/1.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/crtm-fix/2.4.0.1_emc.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/git-lfs/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/crtm/2.4.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/openblas/0.3.24.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-setuptools/63.4.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-numpy/1.23.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/bufr/11.7.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gmake/4.2.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/wgrib2/2.0.8.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-f90nml/1.4.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-cftime/1.0.3.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/py-netcdf4/1.5.8.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libyaml/0.2.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pyyaml/6.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-markupsafe/2.1.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-jinja2/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-bottleneck/1.3.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-numexpr/2.8.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-et-xmlfile/1.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-openpyxl/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pytz/2023.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pyxlsb/1.0.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlrd/2.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlsxwriter/3.1.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlwt/1.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pandas/1.5.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-six/1.16.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-python-dateutil/2.8.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2c/1.8.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-cxx4/4.3.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/met/9.1.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/metplus/3.1.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-packaging/23.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xarray/2023.7.0.lua:/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles/prepobs/1.1.0.lua:/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles/fit2obs/1.1.7.1.lua:/apps/other/modulefiles/globus-cli/3.35.2.lua:/work2/noaa/global/mterry/global-workflow_forked/modulefiles/module_base.hercules.lua', 'globus_cli_ver': 3.27, '_ModuleTable028_': 'L2xpYnBuZy8xLjYuMzcubHVhIiwKZnVsbE5hbWUgPSAibGlicG5nLzEuNi4zNyIsCmxvYWRPcmRlciA9IDE4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImxpYnBuZy8xLjYuMzciLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA2LjAwMDAwMDAzNy4qemZpbmFsIiwKfSwKbGlieGNyeXB0ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2xpYnhjcnlwdC80LjQuMzUubHVhIiwKZnVsbE5hbWUgPSAibGlieGNyeXB0', 'py_numpy_ver': '1.23.4', '__LMOD_REF_COUNT_HDF5_PLUGIN_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/plugins:1', 'bufr_ver': '11.7.0', 'COMINnam': '/work2/noaa/global/role-global/data/external_gempak/nam', 'DBUS_SESSION_BUS_ADDRESS': 'unix:path=/run/user/9583/bus', 'DO_AERO_FCST': False, 'py_pandas_ver': '1.5.3', '_ModuleTable039_': 'dGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9vcGVuanBlZy8yLjMuMS5sdWEiLApmdWxsTmFtZSA9ICJvcGVuanBlZy8yLjMuMSIsCmxvYWRPcmRlciA9IDE5LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm9wZW5qcGVnLzIuMy4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsicGFyYWxsZWwtbmV0Y2RmIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMv', 'COM_ICE_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice', 'UUID_ORION_DTN': '8a10dd4f-24ee-4794-a39d-9c313ab6a34b', 'MPI_F77': 'mpiifort', 'LMOD_CMD': '/apps/other/lmod/lmod/libexec/lmod', '_ModuleTable045_': 'aW5hbCIsCn0sClsicHktYm90dGxlbmVjayJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LWJvdHRsZW5lY2svMS4zLjcubHVhIiwKZnVsbE5hbWUgPSAicHktYm90dGxlbmVjay8xLjMuNyIsCmxvYWRPcmRlciA9IDYzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWJvdHRsZW5lY2svMS4zLjciLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAzLjAwMDAwMDAwNy4qemZpbmFs', 'LMOD_MPI_VERSION': '2021.9.0-a66eaip', 'LMOD_SYSTEM_NAME': 'hercules', 'RUN_ENVIR': 'emc', 'w3emc_ver': '2.10.0', '__LMOD_REF_COUNT_LD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/host/linux64/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/lib:1;/usr/lib64:1;/usr/lib:1;/opt/slurm/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:2', 'DOHYBVAR_OCN': False, 'NUM_SND_COLLECTIVES': 9, 'walltime': '00:15:00', 'netcdf_cxx4_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu', 'ens_tracker_ver': 'v1.2.0', 'LMOD_SYSTEM_DEFAULT_MODULES': 'contrib', 'OMPI_MCA_plm_slurm_args': '--external-launcher', '_ModuleTable078_': 'cyIKLCAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLWE2NmVhaXAvZzJ0bXBsLzEuMTMuMC9pbnRlbC8yMDIxLjkuMCIKLCAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9nMnRtcGwvMS4xMy4wL2ludGVsLzIwMjEuOS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVz', 'FIXcpl': '/work2/noaa/global/mterry/global-workflow_forked/fix/cpl', 'G2C_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64/libg2c.so', '_ModuleTable019_': 'cmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9ncmliLXV0aWwvMS4zLjAubHVhIiwKZnVsbE5hbWUgPSAiZ3JpYi11dGlsLzEuMy4wIiwKbG9hZE9yZGVyID0gNDQsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ3JpYi11dGlsLzEuMy4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKWyJnc2ktbmNkaWFnIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMv', 'SLURM_JOB_GID': 17000, 'SLURM_GET_USER_ENV': 1, 'IAUFHRS_ENKF': [3, 6, 9], 'CPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/include:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/include:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/include', 'G2TMPL_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include', 'DO_JEDISNOWDA': False, 'SP_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_d.a', 'pigz_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq', 'COM_ATMOS_BUFR_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr', 'COM_OCEAN_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2', 'CNVGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/cnvgrib', 'COM_ATMOS_TRACK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks', 'DO_VERFRAD': True, 'pgmerr': 'errfile', 'MPI_BUFS_PER_HOST': 2048, 'DO_PREP_OBS_AERO': False, 'LMOD_FAMILY_G2TMPLVIRT_VERSION': '1.13.0', 'PNG_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu', 'COM_ATMOS_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart', 'DEGRIB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/degrib2', 'DO_ANLSTAT': False, 'LMOD_MPI_NAME': 'intel-oneapi-mpi', 'py_f90nml_ver': '1.4.3', 'NMEM_ENS_GFS_OFFSET': 20, 'SMOOTH_ENKF': False, 'FHMAX_GOES': 120, 'OLDPWD': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', 'QUEUE': 'batch', 'COM_OBS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/obs', 'DO_VMINMON': True, 'metplus_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r', 'VERBOSE': True, 'SLURM_JOB_NODELIST': 'hercules-08-55', 'CLUSTER': 'hercules', 'DBNROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn', 'FHOUT_HF_WAV': 1, 'PARMgfs': '/work2/noaa/global/mterry/global-workflow_forked/parm', 'I_MPI_HYDRA_BOOTSTRAP': 'slurm', 'APRUN_OCNICEPOST': 'srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1', 'BASH_FUNC_ml%%': '() { eval "$($LMOD_DIR/ml_cmd "$@")"\n}', 'BASH_FUNC_which%%': '() { ( alias;\n eval ${which_declare} ) | /usr/bin/which --tty-only --read-alias --read-functions --show-tilde --show-dot $@\n}', 'BASH_FUNC_module%%': '() { if [ -z "${LMOD_SH_DBG_ON+x}" ]; then\n case "$-" in \n *v*x*)\n __lmod_sh_dbg=\'vx\'\n ;;\n *v*)\n __lmod_sh_dbg=\'v\'\n ;;\n *x*)\n __lmod_sh_dbg=\'x\'\n ;;\n esac;\n fi;\n if [ -n "${__lmod_sh_dbg:-}" ]; then\n set +$__lmod_sh_dbg;\n echo "Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for Lmod\'s output" 1>&2;\n fi;\n eval "$($LMOD_CMD $LMOD_SHELL_PRGM "$@")" && eval "$(${LMOD_SETTARG_CMD:-:} -s sh)";\n __lmod_my_status=$?;\n if [ -n "${__lmod_sh_dbg:-}" ]; then\n echo "Shell debugging restarted" 1>&2;\n set -$__lmod_sh_dbg;\n fi;\n unset __lmod_sh_dbg;\n return $__lmod_my_status\n}', 'BASH_FUNC_declare_from_tmpl%%': '() { if [[ ${DEBUG_WORKFLOW:-"NO"} == "NO" ]]; then\n set +x;\n fi;\n local opts="-g";\n local OPTIND=1;\n while getopts "rx" option; do\n opts="${opts}${option}";\n done;\n shift $((OPTIND-1));\n for input in "$@";\n do\n IFS=\':\' read -ra args <<< "${input}";\n local com_var="${args[0]}";\n local template;\n local value;\n if (( ${#args[@]} > 1 )); then\n template="${args[1]}";\n else\n template="${com_var}_TMPL";\n fi;\n if [[ ! -v "${template}" ]]; then\n echo "FATAL ERROR in declare_from_tmpl: Requested template ${template} not defined!";\n exit 2;\n fi;\n value=$(echo "${!template}" | envsubst);\n declare ${opts} "${com_var}"="${value}";\n echo "declare_from_tmpl :: ${com_var}=${value}";\n done;\n set_trace\n}', 'BASH_FUNC_err_exit%%': ['() { set +eux;\n msg1=${*:-Job ${jobid} failed};\n if [[ -n "${pgm}" ]]; then\n msg1+="', 'ERROR IN ${pgm}";\n fi;\n if [[ -n "${err}" ]]; then\n msg1+=" RETURN CODE ${err}";\n fi;\n msg2="\n -------------------------------------------------------------\n -- FATAL ERROR: ${msg1}\n -- ABNORMAL EXIT at $(date) on ${HOSTNAME}\n -------------------------------------------------------------\n ";\n echo "${msg2}" 1>&2;\n module list;\n echo "" 1>&2;\n echo "${msg1}" 1>&2;\n if [[ -n "${DATA}" ]]; then\n echo "${DATA}" 1>&2;\n ls -ltr "${DATA}" 1>&2;\n else\n echo "WARNING: DATA variable not defined" 1>&2;\n fi;\n if [[ -n "${pgmout}" ]]; then\n if [[ -s errfile ]]; then\n echo "----- contents of errfile -----" >> "${pgmout}";\n cat errfile >> "${pgmout}";\n fi;\n cat "${pgmout}" 1>&2;\n else\n if [[ -s errfile ]]; then\n cat errfile 1>&2;\n fi;\n fi;\n if [[ "${SENDECF}" == "YES" ]]; then\n timeout 30 ecflow_client --msg "${ECF_NAME}: ${msg1}";\n timeout 30 ssh "${ECF_HOST}" "echo \\"${msg}2\\" >> ${ECF_JOBOUT:?}";\n fi;\n if [[ "${SENDECF}" == "YES" ]]; then\n ecflow_client --kill="${ECF_NAME:?}";\n fi;\n if [[ -n "${PBS_JOBID}" ]]; then\n qdel "${PBS_JOBID}";\n else\n if [[ -n "${SLURM_JOB_ID}" ]]; then\n scancel "${SLURM_JOB_ID}";\n fi;\n fi\n}'], 'BASH_FUNC_wait_for_file%%': '() { set +x;\n local file_name=${1:?"wait_for_file() requires a file name"};\n local sleep_interval=${2:-60};\n local max_tries=${3:-100};\n for ((iter=0; iter' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.oceanice_products -+++ config.oceanice_products[5]echo 'BEGIN: config.oceanice_products' -BEGIN: config.oceanice_products -+++ config.oceanice_products[8]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources oceanice_products -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=oceanice_products -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[978]walltime=00:15:00 -++++ config.resources[979]ntasks=1 -++++ config.resources[980]tasks_per_node=1 -++++ config.resources[981]threads_per_task=1 -++++ config.resources[982]memory=96GB -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n '' ]] -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export memory -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.oceanice_products[11]export MAX_TASKS=25 -+++ config.oceanice_products[11]MAX_TASKS=25 -+++ config.oceanice_products[13]export OCEANICEPRODUCTS_CONFIG=/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml -+++ config.oceanice_products[13]OCEANICEPRODUCTS_CONFIG=/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml -+++ config.oceanice_products[16]export NFHRS_PER_GROUP=3 -+++ config.oceanice_products[16]NFHRS_PER_GROUP=3 -+++ config.oceanice_products[18]echo 'END: config.oceanice_products' -END: config.oceanice_products -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env oceanice_products -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=oceanice_products -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 100663296 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 1 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 1 ]] -+++ HERCULES.env[34]max_threads_per_task=80 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 80 ]] -+++ HERCULES.env[40][[ 1 -gt 80 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 1' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[286]export NTHREADS_OCNICEPOST=1 -+++ HERCULES.env[286]NTHREADS_OCNICEPOST=1 -+++ HERCULES.env[287]export 'APRUN_OCNICEPOST=srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1' -+++ HERCULES.env[287]APRUN_OCNICEPOST='srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1' -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_OCEANICE_PRODUCTS[11]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[11]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[11]declare_from_tmpl -rx COMIN_OCEAN_HISTORY:COM_OCEAN_HISTORY_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_OCEAN_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/ocean/history -+ JGLOBAL_OCEANICE_PRODUCTS[12]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[12]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[12]declare_from_tmpl -rx COMOUT_OCEAN_GRIB:COM_OCEAN_GRIB_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_OCEAN_GRIB=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ocean/grib2 -+ JGLOBAL_OCEANICE_PRODUCTS[13]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[13]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[13]declare_from_tmpl -rx COMOUT_OCEAN_NETCDF:COM_OCEAN_NETCDF_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_OCEAN_NETCDF=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ocean/netcdf -+ JGLOBAL_OCEANICE_PRODUCTS[17]/work2/noaa/global/mterry/global-workflow_forked/scripts/exglobal_oceanice_products.py -2025-07-28 21:29:51,939 - INFO - root : BEGIN: __main__.main -2025-07-28 21:29:51,939 - DEBUG - root : ( ) -2025-07-28 21:29:51,947 - INFO - oceanice_products: BEGIN: OceanIceProducts.__init__ -2025-07-28 21:29:51,947 - DEBUG - oceanice_products: ( , {'SHELL': '/bin/bash', 'PDY': datetime.datetime(2021, 3, 23, 0, 0), 'DO_JEDIATMENS': False, 'sqlite_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4', 'COPYGB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/copygb', 'WORK_hercules': '/work/hercules', 'TOCGRIB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib2', 'G2_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/include_d', 'PIO_TYPENAME_VALID_VALUES': ['netcdf', 'netcdf4p', 'netcdf4c', 'pnetcdf'], 'WGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/wgrib', 'SERIAL_CXX': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icpc', 'FHMAX_WAV': 9, 'FIXlut': '/work2/noaa/global/mterry/global-workflow_forked/fix/lut', 'fftw_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj', 'snappy_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx', '_ModuleTable016_': 'bGxOYW1lID0gImdldHRleHQvMC4yMS4xIiwKbG9hZE9yZGVyID0gMTEsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ2V0dGV4dC8wLjIxLjEiLAp3ViA9ICIwMDAwMDAwMDAuMDAwMDAwMDIxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJnaXQtbGZzIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvZ2l0LWxmcy8zLjEuMi5sdWEiLApmdWxsTmFtZSA9ICJnaXQtbGZzLzMuMS4yIiwK', 'MAKE_NSSTBUFR': False, 'cdo_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e', 'curl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops', 'PIO_VERSION_MAJOR': 2, 'threads_per_task': 1, 'SLURM_JOB_USER': 'mterry', '_ModuleTable067_': 'NC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3NuYXBweS8xLjEuMTAubHVhIiwKZnVsbE5hbWUgPSAic25hcHB5LzEuMS4xMCIsCmxvYWRPcmRlciA9IDI4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInNuYXBweS8xLjEuMTAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAxLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKc3AgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45', 'COM_ICE_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}', 'COM_OCEAN_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}', '_ModuleTable022_': 'ZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9oZGY1LzEuMTQuMC5sdWEiLApmdWxsTmFtZSA9ICJoZGY1LzEuMTQuMCIsCmxvYWRPcmRlciA9IDI3LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImhkZjUvMS4xNC4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAxNC4qemZpbmFsIiwKfSwKWyJpbnRlbC1vbmVhcGktY29tcGlsZXJzIl0gPSB7CmZuID0gIi9hcHBzL3NwYWNrLW1hbmFnZWQvbW9kdWxlZmlsZXMvbGludXgtcm9ja3k5LXg4Nl82NC9Db3JlL2ludGVsLW9uZWFwaS1jb21waWxlcnMvMjAyMy4xLjAubHVhIiwKZnVsbE5hbWUgPSAiaW50ZWwtb25lYXBpLWNvbXBpbGVy', 'SLURM_TASKS_PER_NODE': 1, 'COM_ATMOS_OZNMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon', 'COM_ATMOS_GEMPAK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}', 'SLURM_JOB_UID': 9583, 'HISTCONTROL': 'ignoredups', 'SLURM_EXPORT_ENV': 'NONE', 'g2c_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp', 'COM_OCEAN_NETCDF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf', 'GRB2INDEX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grb2index', 'l4densvar': True, 'FHOUT_HF_GFS': 1, 'I_MPI_EXTRA_FILESYSTEM': 1, 'SLURM_TASK_PID': 365333, 'EXECprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/exec', 'COMOUT_OCEAN_NETCDF': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ocean/netcdf', 'g2tmpl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel', 'MPI_BUFS_PER_PROC': 2048, 'IAU_DELTHRS_ENKF': 6, 'COM_WAVE_WMO_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo', '_ModuleTable027_': 'LTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9saWJqcGVnLzIuMS4wLmx1YSIsCmZ1bGxOYW1lID0gImxpYmpwZWcvMi4xLjAiLApsb2FkT3JkZXIgPSAxNiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJsaWJqcGVnLzIuMS4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKbGlicG5nID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4w', 'NMEM_ENS': 0, '__LMOD_REF_COUNT_CLASSPATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/mpi.jar:1', '_ModuleTable050_': 'ZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbWFya3Vwc2FmZS8yLjEuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1tYXJrdXBzYWZlLzIuMS4zIiwKbG9hZE9yZGVyID0gNjEsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktbWFya3Vwc2FmZS8yLjEuMyIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDEuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LW5ldGNkZjQiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFj', 'err': 0, '_ModuleTable032_': 'CnVzZXJOYW1lID0gIm1ldHBsdXMvMy4xLjEiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJtb2R1bGVfYmFzZS5oZXJjdWxlcyJdID0gewpmbiA9ICIvd29yazIvbm9hYS9nbG9iYWwvbXRlcnJ5L2dsb2JhbC13b3JrZmxvd19mb3JrZWQvbW9kdWxlZmlsZXMvbW9kdWxlX2Jhc2UuaGVyY3VsZXMubHVhIiwKZnVsbE5hbWUgPSAibW9kdWxlX2Jhc2UuaGVyY3VsZXMiLApsb2FkT3JkZXIgPSA4NCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDAsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJtb2R1bGVfYmFzZS5oZXJjdWxlcyIsCndWID0gIk0uKnpmaW5hbCIsCn0sCm5jbyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGlj', 'CHGRP_CMD': 'chgrp rstprod', 'PKG_CONFIG_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib/pkgconfig:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/pkgconfig:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/lib/pkgconfig', 'SLURM_LOCALID': 0, 'MAX_TASKS': 25, '_ModuleTable040_': 'aW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9wYXJhbGxlbC1uZXRjZGYvMS4xMi4yLmx1YSIsCmZ1bGxOYW1lID0gInBhcmFsbGVsLW5ldGNkZi8xLjEyLjIiLApsb2FkT3JkZXIgPSAzMiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwYXJhbGxlbC1uZXRjZGYvMS4xMi4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAxMi4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCnBhcmFsbGVsaW8gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0', 'SLURM_SUBMIT_DIR': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', '_ModuleTable077_': 'LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvenN0ZC8xLjUuMi5sdWEiLApmdWxsTmFtZSA9ICJ6c3RkLzEuNS4yIiwKbG9hZE9yZGVyID0gOSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMywKc3RhY2tEZXB0aCA9IDQsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJ6c3RkLzEuNS4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwNS4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCn0sCm1wYXRoQSA9IHsKIi93b3JrMi9ub2FhL2dsb2JhbC9yb2xlLWdsb2JhbC9naXQvRml0Mk9icy92MS4xLjcuMS9tb2R1bGVmaWxlcyIKLCAiL3dvcmsyL25vYWEvZ2xvYmFsL3JvbGUtZ2xvYmFsL2dpdC9wcmVwb2JzL3YxLjEuMC9tb2R1bGVmaWxl', 'HOSTNAME': 'hercules-02-53', 'HISTSIZE': 1000, 'wgrib2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln', 'FRAC_GRID': True, 'G2_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/include_4', 'intel_oneapi_mpi_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse', 'SLURMD_NODENAME': 'hercules-02-53', '_ModuleTable080_': 'bGVzL2xpbnV4LXJvY2t5OS14ODZfNjQvb25lYXBpLzIwMjMuMS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL0NvcmUiCiwgIi93b3JrMi9ub2FhL2dsb2JhbC9tdGVycnkvZ2xvYmFsLXdvcmtmbG93X2ZvcmtlZC9tb2R1bGVmaWxlcyIKLCAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L0NvcmUiCiwgIi9hcHBzL290aGVyL21vZHVsZWZpbGVzIiwgIi9hcHBzL2NvbnRhaW5lcnMvbW9kdWxlZmlsZXMiCiwgIi9hcHBzL2xpY2Vuc2VkL21vZHVsZWZpbGVzIiwg', '__LMOD_REF_COUNT_OCL_ICD_FILENAMES': 'libintelocl_emu.so:1;libalteracl.so:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64/libintelocl.so:1', 'c_blosc_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn', 'FHOUT_OCN': 3, 'COM_OBS_JEDI': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi', 'eccodes_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj', '_ModuleTable037_': 'ay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9uZ2h0dHAyLzEuNTcuMC5sdWEiLApmdWxsTmFtZSA9ICJuZ2h0dHAyLzEuNTcuMCIsCmxvYWRPcmRlciA9IDIyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gNCwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm5naHR0cDIvMS41Ny4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDA1Ny4qemZpbmFsIiwKfSwKb3BlbmJsYXMgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2kt', 'netcdf_c_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy', '_ModuleTable070_': 'c3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzdGFjay1pbnRlbC8yMDIxLjkuMCIsCndWID0gIjAwMDAwMjAyMS4wMDAwMDAwMDkuKnpmaW5hbCIsCn0sClsic3RhY2staW50ZWwtb25lYXBpLW1waSJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3N0YWNrLWludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAubHVhIiwKZnVsbE5hbWUgPSAic3RhY2staW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMCIsCmxvYWRPcmRlciA9IDUsCnByb3BUID0g', 'SLURM_JOB_START_TIME': 1753756157, '_ModuleTable011_': 'b3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZXNtZi84LjYuMCIsCndWID0gIjAwMDAwMDAwOC4wMDAwMDAwMDYuKnpmaW5hbCIsCn0sCmZmdHcgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9mZnR3LzMuMy4xMC5sdWEiLApmdWxsTmFtZSA9ICJmZnR3LzMuMy4xMCIsCmxvYWRPcmRlciA9IDIxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwK', 'DO_NEST': False, 'STMP': '/work2/noaa/stmp/mterry/HERCULES', 'HYDRA_LAUNCHER_EXTRA_ARGS': '--external-launcher', 'ACL_BOARD_VENDOR_PATH': '/opt/Intel/OpenCLFPGA/oneAPI/Boards', '_ModuleTable060_': 'ZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1weXlhbWwvNi4wIiwKd1YgPSAiMDAwMDAwMDA2Lip6ZmluYWwiLAp9LApbInB5LXNldHVwdG9vbHMiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1zZXR1cHRvb2xzLzYzLjQuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1zZXR1cHRvb2xzLzYzLjQuMyIsCmxvYWRPcmRlciA9IDUxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIs', '_ModuleTable007_': 'ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9jcnRtLzIuNC4wLjEubHVhIiwKZnVsbE5hbWUgPSAiY3J0bS8yLjQuMC4xIiwKbG9hZE9yZGVyID0gNDksCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY3J0bS8yLjQuMC4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNC4wMDAwMDAwMDAuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbImNydG0tZml4Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45', 'G2TMPL_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib/libg2tmpl.a', 'DO_WAVE': True, 'SCRgfs': '/work2/noaa/global/mterry/global-workflow_forked/scripts', 'SP_INC8': 'include_8', '_ModuleTable002_': 'b2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9hbnRsci8yLjcuNy5sdWEiLApmdWxsTmFtZSA9ICJhbnRsci8yLjcuNyIsCmxvYWRPcmRlciA9IDM1LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImFudGxyLzIuNy43IiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4wMDAwMDAwMDcuKnpmaW5hbCIsCn0sCmJhY2lvID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2JhY2lvLzIuNC4xLmx1', 'MPICC': 'mpiicc', 'HDF5_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh', '_ModuleTable055_': 'L21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXBhY2thZ2luZy8yMy4xLmx1YSIsCmZ1bGxOYW1lID0gInB5LXBhY2thZ2luZy8yMy4xIiwKbG9hZE9yZGVyID0gNzksCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcGFja2FnaW5nLzIzLjEiLAp3ViA9ICIwMDAwMDAwMjMuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbInB5LXBhbmRhcyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEu', 'SP_INC4': 'include_4', 'FPGA_VARS_DIR': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga', 'FHOUT': 3, '__LMOD_REF_COUNT_MODULEPATH': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles:1;/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0-a66eaip/g2tmpl/1.13.0/intel/2021.9.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/g2tmpl/1.13.0/intel/2021.9.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/intel-oneapi-mpi/2021.9.0-a66eaip/oneapi/2023.1.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core:1;/work2/noaa/global/mterry/global-workflow_forked/modulefiles:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core:1;/apps/other/modulefiles:1;/apps/containers/modulefiles:1;/apps/licensed/modulefiles:1;/apps/contrib/modulefiles:1', 'cycle': 't12z', '_ModuleTable043_': 'aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcmVwb2JzLzEuMS4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKcHJvZF91dGlsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3Byb2RfdXRpbC8yLjEuMS5sdWEiLApmdWxsTmFtZSA9ICJwcm9kX3V0aWwvMi4xLjEiLApsb2FkT3JkZXIgPSA0MCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcm9kX3V0aWwvMi4x', 'MPI_CC': 'mpiicc', 'PARTITION_DTN': '', 'sp_ver': '2.5.0', 'SHELLOPTS': 'braceexpand:errexit:hashall:interactive-comments:nounset:xtrace', 'DO_AWIPS': False, 'EDATE': datetime.datetime(2021, 3, 23, 12, 0), '__LMOD_REF_COUNT___INTEL_POST_CFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1'], 'I_MPI_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0', 'CRTM_FIX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2/fix', '_ModuleTable062_': 'MTYuMCIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMTYuKnpmaW5hbCIsCn0sClsicHkteGFycmF5Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHkteGFycmF5LzIwMjMuNy4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXhhcnJheS8yMDIzLjcuMCIsCmxvYWRPcmRlciA9IDgwLApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXhhcnJheS8yMDIzLjcuMCIsCndWID0gIjAwMDAwMjAyMy4wMDAwMDAwMDcuKnpm', 'proj_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez', 'FSYNC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file', '_ModuleTable005_': 'cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJjLWJsb3NjLzEuMjEuNSIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMjEuMDAwMDAwMDA1Lip6ZmluYWwiLAp9LApjZG8gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9jZG8vMi4yLjAubHVhIiwKZnVsbE5hbWUgPSAiY2RvLzIuMi4wIiwKbG9hZE9yZGVyID0gMjYsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY2RvLzIu', 'g2tmpl_ver': '1.13.0', 'FHOUT_GOES': 3, '_ModuleTable021_': 'MC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nc2wvMi43LjEubHVhIiwKZnVsbE5hbWUgPSAiZ3NsLzIuNy4xIiwKbG9hZE9yZGVyID0gMzYsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ3NsLzIuNy4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sCmhkZjUgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxl', 'EUPD_CYC': 'gdas', 'fms_ver': '2023.02.01', 'util_linux_uuid_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj', '_ModuleTable052_': 'Y2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbnVtZXhwci8yLjguNC5sdWEiLApmdWxsTmFtZSA9ICJweS1udW1leHByLzIuOC40IiwKbG9hZE9yZGVyID0gNjQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktbnVtZXhwci8yLjguNCIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDguMDAwMDAwMDA0Lip6ZmluYWwiLAp9LApbInB5LW51bXB5Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2kt', '_ModuleTable030_': 'cFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJsaWJ5YW1sLzAuMi41IiwKd1YgPSAiMDAwMDAwMDAwLjAwMDAwMDAwMi4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCm1ldCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL21ldC85LjEuMy5sdWEiLApmdWxsTmFtZSA9ICJtZXQvOS4xLjMiLApsb2FkT3JkZXIgPSA3NywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0', 'COM_ATMOS_IMAGERY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery', 'I_MPI_EXTRA_FILESYSTEM_LIST': 'lustre', 'COM_WAVE_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history', 'SLURM_CLUSTER_NAME': 'hercules', '__LMOD_REF_COUNT_ACLOCAL_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/aclocal:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/aclocal:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/aclocal:2', 'SERIAL_CC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icc', 'COM_OBSPROC_TMPL': '${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos', 'SLURM_JOB_END_TIME': 1753757057, '_ModuleTable075_': 'ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3dncmliMi8yLjAuOC5sdWEiLApmdWxsTmFtZSA9ICJ3Z3JpYjIvMi4wLjgiLApsb2FkT3JkZXIgPSA1NSwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJ3Z3JpYjIvMi4wLjgiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKemxpYiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJj', 'FHMAX_GFS': 120, 'SENDAWIP': False, 'G2_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64/libg2_d.a', 'spack_mod_path': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core', 'OCNRES': 500, 'COM_MED_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart', 'SLURM_CPUS_ON_NODE': 1, 'LMOD_FAMILY_METAMPI': 'stack-intel-oneapi-mpi', 'LEVS': 128, 'FIXgsi': '/work2/noaa/global/mterry/global-workflow_forked/fix/gsi', 'NFHRS_PER_GROUP': 3, 'COM_WAVE_PREP_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep', 'FI_PROVIDER_PATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib/prov:/usr/lib64/libfabric', 'python_ver': '3.11.6', 'IAU_OFFSET': 0, 'py_packaging_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6', 'DIAGUTIL_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/sys_check/sys_check.sh', 'python_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2', 'SCRIPTScfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/scripts', 'py_markupsafe_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7', 'G2C_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include', 'py_xlrd_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f', 'LMOD_G2TMPLVIRT_VERSION': '1.13.0', 'pid': 365753, 'OUTPUT_GRID': 'gaussian_grid', 'jobid': 'oceanice_products_ocean_f048.365390', 'SLURM_JOB_CPUS_PER_NODE': 1, 'spack_env': 'gsi-addon-dev-fms-2024.01', 'INTEL_ONEAPI_MPI_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse', 'LMOD_FAMILY_METACOMPILER': 'stack-intel', 'SP_INCd': 'include_d', 'py_six_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o', 'COM_ATMOS_GOES_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim', 'MPICXX': 'mpiicpc', 'restart_interval_gdas': 3, 'fit2obs_ver': '1.1.7.1', 'LMOD_DIR': '/apps/other/lmod/lmod/libexec', '_ModuleTable068_': 'LjAvc3AvMi41LjAubHVhIiwKZnVsbE5hbWUgPSAic3AvMi41LjAiLApsb2FkT3JkZXIgPSA0MiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMiwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzcC8yLjUuMCIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCnNxbGl0ZSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9zcWxpdGUvMy40My4yLmx1YSIsCmZ1bGxOYW1lID0gInNxbGl0ZS8zLjQzLjIiLApsb2FkT3Jk', 'REPLAY_ICS': False, 'COM_ICE_NETCDF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf', 'G2_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64/libg2_4.a', '_ModuleTable058_': 'InB5LXB5dHovMjAyMy4zIiwKbG9hZE9yZGVyID0gNjcsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcHl0ei8yMDIzLjMiLAp3ViA9ICIwMDAwMDIwMjMuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LXB5eGxzYiJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXB5eGxzYi8xLjAuMTAubHVhIiwKZnVsbE5hbWUgPSAicHktcHl4bHNiLzEuMC4xMCIsCmxvYWRPcmRlciA9', 'py_setuptools_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr', 'FHMIN': 0, 'COM_ICE_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input', '_ModuleTable048_': 'cHktZjkwbm1sIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktZjkwbm1sLzEuNC4zLmx1YSIsCmZ1bGxOYW1lID0gInB5LWY5MG5tbC8xLjQuMyIsCmxvYWRPcmRlciA9IDU2LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWY5MG5tbC8xLjQuMyIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMDQuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LWppbmphMiJdID0gewpmbiA9ICIvd29yay9ub2FhL2Vw', 'intel_mkl_ver': '2023.1.0', 'PRTE_MCA_plm_slurm_args': '--external-launcher', 'PWD': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ocean_f048.365390', 'SLURM_GTIDS': 0, 'LOGNAME': 'mterry', 'MAKE_ACFTBUFR': False, 'W3EMC_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_8', 'HOMEDIR': '/work2/noaa/global/mterry', 'IAU_DELTHRS': 6, 'W3EMC_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_4', 'XDG_SESSION_TYPE': 'unspecified', 'ESMFMKFILE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib/esmf.mk', 'FHMAX': 9, 'WRITE_NSFLIP': True, 'SLURM_JOB_PARTITION': 'hercules', 'YAML_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x', 'PARTITION_SERVICE': 'service', 'MODULESHOME': '/apps/other/lmod/lmod', '__LMOD_REF_COUNT_DYLD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:1', 'COM_WAVE_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart', 'BUFR_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_4.so', 'BUFR_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_8.so', 'HOMEcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1', 'COM_OCEAN_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon', 'DEBUG_POSTSCRIPT': False, 'MANPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/share/man:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/man:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/documentation/en/man/common:/apps/other/lmod/lmod/share/man:/usr/share/man:/apps/share/man:/apps/man:/opt/slurm/share/man::', 'libyaml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x', 'jasper_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt', 'py_xlsxwriter_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok', 'SLURM_TRES_PER_TASK': 'cpu=1', 'nco_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq', 'stack_intel_ver': '2021.9.0', 'SLURM_OOM_KILL_STEP': 0, 'BUFR_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m', 'bufr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m', 'CLUSTERS_DTN': '', 'grib_util_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35', 'nghttp2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky', 'DO_TEST_MODE': True, 'PACKAGEROOT': '/work2/noaa/global/role-global/nwpara', 'lobsdiag_forenkf': True, 'SLURM_JOB_NUM_NODES': 1, 'py_netcdf4_ver': '1.5.8', 'ICERES': 500, 'CXX': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icpc', 'GRIB2GRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grib2grib2', 'DMPDIR': '/work/noaa/rstprod/dump', 'LSOIL_INCR': 2, 'metplus_ver': '3.1.1', '_ModuleTable008_': 'LjAvY3J0bS1maXgvMi40LjAuMV9lbWMubHVhIiwKZnVsbE5hbWUgPSAiY3J0bS1maXgvMi40LjAuMV9lbWMiLApsb2FkT3JkZXIgPSA0NywKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJjcnRtLWZpeC8yLjQuMC4xX2VtYyIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDQuMDAwMDAwMDAwLjAwMDAwMDAwMS4qXy4qZW1jLip6ZmluYWwiLAp9LApjdXJsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIw', 'COM_OCEAN_LETKF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf', 'zlib_ver': '1.2.13', 'BUFR_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_d.so', 'ENKF_SPREAD': True, 'zlib_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j', '__LMOD_REF_COUNT_CMAKE_PREFIX_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r:2;/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/IntelDPCPP:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72:1', '_ModuleTable035_': 'c3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9uZXRjZGYtY3h4NC80LjMuMS5sdWEiLApmdWxsTmFtZSA9ICJuZXRjZGYtY3h4NC80LjMuMSIsCmxvYWRPcmRlciA9IDc2LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm5ldGNkZi1jeHg0LzQuMy4xIiwKd1YgPSAiMDAwMDAwMDA0LjAwMDAwMDAwMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsibmV0Y2RmLWZvcnRyYW4iXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFj', 'FIXgfs': '/work2/noaa/global/mterry/global-workflow_forked/fix', 'DO_COUPLED': True, 'SLURM_JOBID': 5951672, 'SERIAL_FC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'COM_ATMOS_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input', 'W3EMC_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_d', '_ModuleTable013_': 'IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nMi8zLjQuNS5sdWEiLApmdWxsTmFtZSA9ICJnMi8zLjQuNSIsCmxvYWRPcmRlciA9IDQxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImcyLzMuNC41IiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAwNC4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCmcyYyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJj', 'COM_ATMOS_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history', 'FIXam': '/work2/noaa/global/mterry/global-workflow_forked/fix/am', 'hdf5_ver': '1.14.0', 'restart_interval_enkfgdas': 3, 'COM_ATMOS_RADMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon', 'DO_CALC_INCREMENT': False, 'DO_STARTMEM_FROM_JEDIICE': False, 'FIXprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/fix', 'SLURM_JOB_QOS': 'batch', 'I_MPI_HYDRA_BOOTSTRAP_EXEC_EXTRA_ARGS': '--external-launcher', 'obsproc_run_ver': '1.2.0', 'COM_ATMOS_MINMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon', 'crtm_ver': '2.4.0.1', 'FIXorog': '/work2/noaa/global/mterry/global-workflow_forked/fix/orog', 'COM_OCEAN_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history', '__LMOD_REF_COUNT_PATH': '/apps/other/globus-cli-3.35.2/bin:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r/ush:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/bin:2;/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/bin:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/bin/intel64:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/bin:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/bin:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/bin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/bin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin:1;/usr/sbin:1;/usr/bin:1;/apps/sbin:1;/apps/bin:1;/opt/slurm/bin:1;/home/gfekete/sven/bin:1', '_ModuleTable018_': 'KnpmaW5hbCIsCn0sCmdtYWtlID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2dtYWtlLzQuMi4xLmx1YSIsCmZ1bGxOYW1lID0gImdtYWtlLzQuMi4xIiwKbG9hZE9yZGVyID0gNTQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ21ha2UvNC4yLjEiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDAyLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJncmliLXV0aWwiXSA9IHsKZm4gPSAiL3dv', 'COMROOT': '/work2/noaa/global/mterry/RUNTESTS/COMROOT', 'HOME': '/home/mterry', 'FI_PROVIDER': 'mlx', 'met_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b', 'COM_RTOFS_TMPL': '${DMPDIR}', '_ModuleTable015_': 'bnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2cydG1wbC8xLjEzLjAubHVhIiwKZnVsbE5hbWUgPSAiZzJ0bXBsLzEuMTMuMCIsCmxvYWRPcmRlciA9IDQ1LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImcydG1wbC8xLjEzLjAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDEzLip6ZmluYWwiLAp9LApnZXR0ZXh0ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2dldHRleHQvMC4yMS4xLmx1YSIsCmZ1', '_ModuleTable_Sz_': 81, 'LANG': 'C.UTF-8', '__LMOD_REF_COUNT_LIBRARY_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:1', 'DO_BUFRSND': False, 'memory': '96GB', '_ModuleTable064_': 'PSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHkteGxzeHdyaXRlci8zLjEuNy5sdWEiLApmdWxsTmFtZSA9ICJweS14bHN4d3JpdGVyLzMuMS43IiwKbG9hZE9yZGVyID0gNzAsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHkteGxzeHdyaXRlci8zLjEuNyIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDA3Lip6ZmluYWwiLAp9LApbInB5LXhsd3QiXSA9IHsKZm4gPSAi', 'DYLD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib', 'PARTITION_BATCH': 'hercules', '_ModuleTable023_': 'cy8yMDIzLjEuMCIsCmxvYWRPcmRlciA9IDIsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiaW50ZWwtb25lYXBpLWNvbXBpbGVycy8yMDIzLjEuMCIsCndWID0gIjAwMDAwMjAyMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsiaW50ZWwtb25lYXBpLW1rbCJdID0gewpmbiA9ICIvYXBwcy9zcGFjay1tYW5hZ2VkL21vZHVsZWZpbGVzL2xpbnV4LXJvY2t5OS14ODZfNjQvQ29yZS9pbnRlbC1vbmVhcGktbWtsLzIwMjMuMS4wLmx1YSIsCmZ1bGxOYW1lID0gImludGVsLW9uZWFwaS1ta2wvMjAyMy4xLjAiLApsb2FkT3JkZXIgPSA2LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVz', '_ModuleTable081_': 'Ii9hcHBzL2NvbnRyaWIvbW9kdWxlZmlsZXMiLAp9LApzeXN0ZW1CYXNlTVBBVEggPSAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L0NvcmU6L2FwcHMvb3RoZXIvbW9kdWxlZmlsZXM6L2FwcHMvY29udGFpbmVycy9tb2R1bGVmaWxlczovYXBwcy9saWNlbnNlZC9tb2R1bGVmaWxlcyIsCn0K', 'REDERR': '2>', 'PSLOT': 'C48_S2SW', '__LMOD_REF_COUNT_PKG_CONFIG_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib/pkgconfig:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/pkgconfig:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib/pkgconfig:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/pkgconfig:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib/pkgconfig:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/lib/pkgconfig:1', '_ModuleTable036_': 'ay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL25ldGNkZi1mb3J0cmFuLzQuNi4xLmx1YSIsCmZ1bGxOYW1lID0gIm5ldGNkZi1mb3J0cmFuLzQuNi4xIiwKbG9hZE9yZGVyID0gMzEsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAibmV0Y2RmLWZvcnRyYW4vNC42LjEiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDA2LjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKbmdodHRwMiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFj', 'COM_ATMOS_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}', 'USHcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/ush', 'DO_GEMPAK': False, 'SLURM_PROCID': 0, 'USHgfs': '/work2/noaa/global/mterry/global-workflow_forked/ush', 'DOIBP_WAV': False, 'prepobs_run_ver': '1.1.0', 'EXPDIR': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', 'job': 'oceanice_products', 'imp_physics': 8, 'LMOD_SHELL_PRGM': 'bash', 'IAU_FHROT': 0, 'MPIF90': 'mpiifort', 'IP_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_4.a', 'waveGRD': 'uglo_100km', 'COM_CHEM_BMAT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix', 'jasper_ver': '2.0.32', 'SCRATCH': '/scratch/hercules', 'IP_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_8.a', 'CLIENT_GLOBUS_UUID': '869912fe-f6de-46c0-af10-b22efd84a022', 'ACCOUNT': 'fv3-cpu', 'HOMEgfs': '/work2/noaa/global/mterry/global-workflow_forked', 'NHOUR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/nhour', 'binary_diag': False, 'COM_WAVE_GRID_RES_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}', 'SCRIPTSprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/scripts', 'MP_STDOUTMODE': 'ORDERED', 'sigio_ver': '2.3.2', '__LMOD_REF_COUNT_CPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/include:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/include:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/include:1', 'LMOD_SETTARG_FULL_SUPPORT': False, 'COMINsyn': '/work2/noaa/global/role-global/com/gfs/prod/syndat', 'OFFSET_START_HOUR': 0, 'REDOUT': '1>', 'PTMP': '/work2/noaa/stmp/mterry/HERCULES', '_ModuleTable076_': 'dWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC96bGliLzEuMi4xMy5sdWEiLApmdWxsTmFtZSA9ICJ6bGliLzEuMi4xMyIsCmxvYWRPcmRlciA9IDcsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEwLApzdGFja0RlcHRoID0gNSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInpsaWIvMS4yLjEzIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMi4wMDAwMDAwMTMuKnpmaW5hbCIsCn0sCnpzdGQgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2', 'g2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato', 'MODE': 'forecast-only', 'DATAROOT': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312', 'bacio_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj', 'APP': 'S2SW', '_ModuleTable029_': 'LzQuNC4zNSIsCmxvYWRPcmRlciA9IDEyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImxpYnhjcnlwdC80LjQuMzUiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDA0LjAwMDAwMDAzNS4qemZpbmFsIiwKfSwKbGlieWFtbCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9saWJ5YW1sLzAuMi41Lmx1YSIsCmZ1bGxOYW1lID0gImxpYnlhbWwvMC4yLjUiLApsb2FkT3JkZXIgPSA1OSwKcHJv', 'openblas_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f', 'stack_impi_ver': '2021.9.0', 'COM_CHEM_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem', 'TMPDIR': '/local/scratch/mterry/5951672', 'HDF5_PLUGIN_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/plugins', 'DO_TRACKER': True, 'CMAKE_PREFIX_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r:/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/IntelDPCPP:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72', 'crtm_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r', '_ModuleTable038_': 'YWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvb3BlbmJsYXMvMC4zLjI0Lmx1YSIsCmZ1bGxOYW1lID0gIm9wZW5ibGFzLzAuMy4yNCIsCmxvYWRPcmRlciA9IDUwLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm9wZW5ibGFzLzAuMy4yNCIsCndWID0gIjAwMDAwMDAwMC4wMDAwMDAwMDMuMDAwMDAwMDI0Lip6ZmluYWwiLAp9LApvcGVuanBlZyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5z', 'nco_ver': '5.0.6', '_ModuleTable046_': 'IiwKfSwKWyJweS1jZnRpbWUiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1jZnRpbWUvMS4wLjMuNC5sdWEiLApmdWxsTmFtZSA9ICJweS1jZnRpbWUvMS4wLjMuNCIsCmxvYWRPcmRlciA9IDU3LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWNmdGltZS8xLjAuMy40IiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMC4wMDAwMDAwMDMuMDAwMDAwMDA0Lip6ZmluYWwiLAp9', 'GRBINDEX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grbindex', 'libjpeg_turbo_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy', 'MPIF77': 'mpiifort', '_ModuleTable073_': 'LjAwMDAwMDAwMi4wMDAwMDAwMjguKnpmaW5hbCIsCn0sClsidXRpbC1saW51eC11dWlkIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvdXRpbC1saW51eC11dWlkLzIuMzguMS5sdWEiLApmdWxsTmFtZSA9ICJ1dGlsLWxpbnV4LXV1aWQvMi4zOC4xIiwKbG9hZE9yZGVyID0gMTQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAidXRpbC1saW51eC11dWlkLzIuMzguMSIsCndWID0gIjAw', '_ModuleTable010_': 'Mi4wIiwKbG9hZE9yZGVyID0gMjAsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZWNjb2Rlcy8yLjMyLjAiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDMyLip6ZmluYWwiLAp9LAplc21mID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAvZXNtZi84LjYuMC5sdWEiLApmdWxsTmFtZSA9ICJlc21mLzguNi4wIiwKbG9hZE9yZGVyID0gMzQsCnBy', 'NTHSTACK': 1024000000, 'SLURM_CPUS_PER_TASK': 1, 'FIXcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/fix', 'SLURM_NTASKS': 1, 'DO_ICE': True, 'NET': 'gfs', 'COM_ATMOS_GENESIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital', 'HOMEpost': '/work2/noaa/global/mterry/global-workflow_forked', '_ModuleTable061_': 'CnVzZXJOYW1lID0gInB5LXNldHVwdG9vbHMvNjMuNC4zIiwKd1YgPSAiMDAwMDAwMDYzLjAwMDAwMDAwNC4wMDAwMDAwMDMuKnpmaW5hbCIsCn0sClsicHktc2l4Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktc2l4LzEuMTYuMC5sdWEiLApmdWxsTmFtZSA9ICJweS1zaXgvMS4xNi4wIiwKbG9hZE9yZGVyID0gNzMsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktc2l4LzEu', 'bacio_ver': '2.4.1', 'cmake_ver': '3.23.1', 'UTILROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq', '_ModuleTable024_': 'ZXJOYW1lID0gImludGVsLW9uZWFwaS1ta2wvMjAyMy4xLjAiLAp3ViA9ICIwMDAwMDIwMjMuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbImludGVsLW9uZWFwaS1tcGkiXSA9IHsKZm4gPSAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L29uZWFwaS8yMDIzLjEuMC9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLmx1YSIsCmZ1bGxOYW1lID0gImludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAiLApsb2FkT3JkZXIgPSA0LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAiLAp3ViA9ICIwMDAwMDIwMjEuMDAwMDAwMDA5Lip6ZmluYWwi', 'tar_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths', 'ntasks': 1, 'max_tasks_per_node': 80, 'NCDUMP': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump', '_ModuleTable057_': 'LjIubHVhIiwKZnVsbE5hbWUgPSAicHktcHl0aG9uLWRhdGV1dGlsLzIuOC4yIiwKbG9hZE9yZGVyID0gNzQsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcHl0aG9uLWRhdGV1dGlsLzIuOC4yIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwOC4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sClsicHktcHl0eiJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXB5dHovMjAyMy4zLmx1YSIsCmZ1bGxOYW1lID0g', 'py_numpy_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne', 'QUILTING': True, 'IP_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_d.a', 'SLURM_TOPOLOGY_ADDR': 'hercules-02-53', 'libxcrypt_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr', 'CRTM_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include', 'COMIN_OCEAN_HISTORY': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/ocean/history', 'LMOD_VERSION': '8.7.14', 'spack_stack_ver': '1.6.0', '_ModuleTable041_': 'LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9wYXJhbGxlbGlvLzIuNS4xMC5sdWEiLApmdWxsTmFtZSA9ICJwYXJhbGxlbGlvLzIuNS4xMCIsCmxvYWRPcmRlciA9IDMzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInBhcmFsbGVsaW8vMi41LjEwIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNS4wMDAwMDAwMTAuKnpmaW5hbCIsCn0sCnBpZ3ogPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0y', 'HOMEprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0', 'COMPONENT': 'ocean', 'BASE_GIT': '/work2/noaa/global/role-global/git', 'FCST_SEGMENTS': [0, 120], 'libpng_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu', '_ModuleTable044_': 'LjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKcHJvaiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9wcm9qLzkuMi4xLmx1YSIsCmZ1bGxOYW1lID0gInByb2ovOS4yLjEiLApsb2FkT3JkZXIgPSAyNCwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcm9qLzkuMi4xIiwKd1YgPSAiMDAwMDAwMDA5LjAwMDAwMDAwMi4wMDAwMDAwMDEuKnpm', 'DO_GENESIS': True, 'BUFR_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_4', '__LMOD_REF_COUNT_DIAGUTIL_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/sys_check/sys_check.sh:1', 'COM_WAVE_GEMPAK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak', 'COM_ATMOS_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2', 'RUN': 'gfs', 'BUFR_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_8', 'COM_ICE_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon', 'py_pandas_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw', '__INTEL_POST_CFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64'], 'ARCDIR': '/work2/noaa/global/mterry/archive/C48_S2SW', '__LMOD_REF_COUNT_NLSPATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64/locale/%l_%t/%N:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin/locale/%l_%t/%N:1', '_ModuleTable069_': 'ZXIgPSAxMywKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMiwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzcWxpdGUvMy40My4yIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDA0My4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sClsic3RhY2staW50ZWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9Db3JlL3N0YWNrLWludGVsLzIwMjEuOS4wLmx1YSIsCmZ1bGxOYW1lID0gInN0YWNrLWludGVsLzIwMjEuOS4wIiwKbG9hZE9yZGVyID0gMywKcHJvcFQgPSB7fSwK', 'SENDECF': False, 'parallelio_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed', 'py_pytz_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s', 'ILPOST': 3, 'WGRIB2_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib/libwgrib2.a', 'parallel_netcdf_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3', 'CMPLR_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0', 'prod_util_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq', 'COMINukmet': '/work2/noaa/global/role-global/data/external_gempak/ukmet', 'libpng_ver': '1.6.37', 'COPYGB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/copygb2', 'NMV': '/bin/mv', 'NOSCRUB': '/work2/noaa/global/mterry', 'SCRATCH_hercules': '/scratch/hercules', 'W3EMC_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_d.a', 'SENDSDM': False, 'envir': 'prod', 'DO_METP': False, 'CRTM_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib/libcrtm.a', 'LOGSCRIPT': '', 'INTEL_ONEAPI_COMPILERS_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72', '_ModuleTable074_': 'MDAwMDAwMi4wMDAwMDAwMzguMDAwMDAwMDAxLip6ZmluYWwiLAp9LAp3M2VtYyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC93M2VtYy8yLjEwLjAubHVhIiwKZnVsbE5hbWUgPSAidzNlbWMvMi4xMC4wIiwKbG9hZE9yZGVyID0gMzksCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAidzNlbWMvMi4xMC4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKd2dyaWIy', 'py_jinja2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj', 'FPGA_VARS_ARGS': '', 'INCVARS_ZERO_STRAT': ["'sphum_inc'", "'liq_wat_inc'", "'icmr_inc'", "'rwmr_inc'", "'snmr_inc'", "'grle_inc'"], 'COM_CONF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf', '_ModuleTable003_': 'YSIsCmZ1bGxOYW1lID0gImJhY2lvLzIuNC4xIiwKbG9hZE9yZGVyID0gMzgsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAzLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiYmFjaW8vMi40LjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDA0LjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKYnVmciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9idWZyLzExLjcuMC5sdWEiLApmdWxsTmFtZSA9ICJidWZyLzExLjcuMCIsCmxvYWRPcmRlciA9', 'HYDRA_BOOTSTRAP': 'slurm', '__LMOD_REF_COUNT___INTEL_POST_FFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1'], 'COM_CHEM_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon', 'MDATE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/mdate', 'WGRIB2': 'wgrib2', '_ModuleTable054_': 'LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktb3BlbnB5eGwvMy4xLjIubHVhIiwKZnVsbE5hbWUgPSAicHktb3BlbnB5eGwvMy4xLjIiLApsb2FkT3JkZXIgPSA2NiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1vcGVucHl4bC8zLjEuMiIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbInB5LXBhY2thZ2luZyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxs', 'EXP_WARM_START': False, 'MODULEPATH_ROOT': '/apps/other/modulefiles', 'COM_SNOW_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon', 'CHGRP_RSTPROD': True, 'py_pyxlsb_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge', 'BACIO_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/include_8', 'restart_interval_gfs': 12, 'DO_GOES': False, 'pgm': '', 'SLURM_TOPOLOGY_ADDR_PATTERN': 'node', 'DO_LAND_IAU': False, 'ntiles': 6, 'FHCYC': 24, 'ROTDIR': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW', 'FHOUT_AERO': 3, 'py_jinja2_ver': '3.1.2', 'BACIO_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/include_4', 'INCREMENTS_TO_ZERO': ["'liq_wat_inc'", "'icmr_inc'", "'rwmr_inc'", "'snmr_inc'", "'grle_inc'"], 'FHR_LIST': 48, 'USHprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/ush', 'BUFR_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_d', 'hdf5_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh', 'INTEL_ONEAPI_MKL_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci', 'SCRIPTSfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/scripts', '__LMOD_REF_COUNT_FI_PROVIDER_PATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib/prov:1;/usr/lib64/libfabric:1', 'FIXcice': '/work2/noaa/global/mterry/global-workflow_forked/fix/cice', 'IAUFHRS': [6], 'XDG_SESSION_CLASS': 'background', 'INTERVAL_GFS': 6, 'py_openpyxl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe', 'FHMAX_FITS': 120, 'COM_ATMOS_MASTER_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master', 'LMOD_PKG': '/apps/other/lmod/lmod', 'MPI_MEMMAP_OFF': 1, 'UUID_HERCULES_DTN': '869912fe-f6de-46c0-af10-b22efd84a022', '_ModuleTable079_': 'L3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAiCiwgIi9hcHBzL3NwYWNrLW1hbmFnZWQvbW9kdWxlZmlsZXMvbGludXgtcm9ja3k5LXg4Nl82NC9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLWE2NmVhaXAvb25lYXBpLzIwMjMuMS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wIgosICIvYXBwcy9zcGFjay1tYW5hZ2VkL21vZHVsZWZp', 'SDATE': datetime.datetime(2021, 3, 23, 12, 0), 'SLURM_SCRIPT_CONTEXT': 'prolog_task', 'MPI_CXX': 'mpiicpc', 'lwrite4danl': True, 'SLURM_MEM_PER_NODE': 98304, 'CASE_ENS': '{{ CASE_ENS }}', '_ModuleTable049_': 'aWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LWppbmphMi8zLjEuMi5sdWEiLApmdWxsTmFtZSA9ICJweS1qaW5qYTIvMy4xLjIiLApsb2FkT3JkZXIgPSA2MiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1qaW5qYTIvMy4xLjIiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDAxLjAwMDAwMDAwMi4qemZpbmFsIiwKfSwKWyJweS1tYXJrdXBzYWZlIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3Vs', 'YAML_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include', 'PYTHONPATH': '/apps/other/globus-cli-3.35.2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:/work2/noaa/global/mterry/global-workflow_forked/sorc/wxflow/src:/work2/noaa/global/mterry/global-workflow_forked/ush/python', '__LMOD_REF_COUNT_PYTHONPATH': '/apps/other/globus-cli-3.35.2:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:3;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:1', 'py_xarray_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4', 'nst_anl': True, 'FHOUT_GFS': 3, 'WORK': '/work/hercules', 'W3EMC_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_4.a', 'F77': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'BASE_DATA': '/work2/noaa/global/role-global/data', 'py_xlwt_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56', 'ACLOCAL_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/aclocal:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/aclocal:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/aclocal', 'QUEUE_SERVICE': 'batch', 'W3EMC_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_8.a', 'DATA': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ocean_f048.365390', 'LESSOPEN': '||/usr/bin/lesspipe.sh %s', 'DO_VERFOZN': True, 'antlr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2', 'crtm_fix_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2', 'TOCGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib', 'USER': 'mterry', 'HPSS_PROJECT': 'emc-global', 'FHMAX_WAV_GFS': 120, 'FIXgdas': '/work2/noaa/global/mterry/global-workflow_forked/fix/gdas', 'launcher': 'srun -l --export=ALL --hint=nomultithread', 'NDATE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/ndate', 'LIBRARY_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib', 'gsl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z', 'SLURM_NODELIST': 'hercules-02-53', 'DO_FIT2OBS': True, 'IP_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_4', 'DOIAU_ENKF': True, 'COM_ATMOS_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos', 'py_xarray_ver': '2023.7.0', '__INTEL_POST_FFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64'], 'netcdf_diag': True, 'DOBNDPNT_WAVE': True, 'FETCHDIR': '/NCEPDEV/emc-global/1year/David.Grumm/test_data', 'ENVIRONMENT': 'BATCH', 'gsi_ncdiag_ver': '1.1.2', 'IP_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_8', 'py_bottleneck_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a', 'w3emc_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo', 'py_netcdf4_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i', 'wgrib2_ver': '2.0.8', 'TOCGRIB2SUPER': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib2super', 'COM_ICE_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2', 'KEEPDATA': False, 'udunits_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo', 'cdo_ver': '2.2.0', 'COM_OCEAN_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input', 'FNTSFA': ' ', 'LOADEDMODULES': 'contrib/0.1:intel-oneapi-compilers/2023.1.0:stack-intel/2021.9.0:intel-oneapi-mpi/2021.9.0:stack-intel-oneapi-mpi/2021.9.0:intel-oneapi-mkl/2023.1.0:zlib/1.2.13:pigz/2.7:zstd/1.5.2:tar/1.34:gettext/0.21.1:libxcrypt/4.4.35:sqlite/3.43.2:util-linux-uuid/2.38.1:python/3.11.6:libjpeg/2.1.0:jasper/2.0.32:libpng/1.6.37:openjpeg/2.3.1:eccodes/2.32.0:fftw/3.3.10:nghttp2/1.57.0:curl/8.4.0:proj/9.2.1:udunits/2.2.28:cdo/2.2.0:hdf5/1.14.0:snappy/1.1.10:c-blosc/1.21.5:netcdf-c/4.9.2:netcdf-fortran/4.6.1:parallel-netcdf/1.12.2:parallelio/2.5.10:esmf/8.6.0:antlr/2.7.7:gsl/2.7.1:nco/5.0.6:bacio/2.4.1:w3emc/2.10.0:prod_util/2.1.1:g2/3.4.5:sp/2.5.0:ip/4.3.0:grib-util/1.3.0:g2tmpl/1.13.0:gsi-ncdiag/1.1.2:crtm-fix/2.4.0.1_emc:git-lfs/3.1.2:crtm/2.4.0.1:openblas/0.3.24:py-setuptools/63.4.3:py-numpy/1.23.4:bufr/11.7.0:gmake/4.2.1:wgrib2/2.0.8:py-f90nml/1.4.3:py-cftime/1.0.3.4:py-netcdf4/1.5.8:libyaml/0.2.5:py-pyyaml/6.0:py-markupsafe/2.1.3:py-jinja2/3.1.2:py-bottleneck/1.3.7:py-numexpr/2.8.4:py-et-xmlfile/1.0.1:py-openpyxl/3.1.2:py-pytz/2023.3:py-pyxlsb/1.0.10:py-xlrd/2.0.1:py-xlsxwriter/3.1.7:py-xlwt/1.3.0:py-pandas/1.5.3:py-six/1.16.0:py-python-dateutil/2.8.2:g2c/1.8.0:netcdf-cxx4/4.3.1:met/9.1.3:metplus/3.1.1:py-packaging/23.1:py-xarray/2023.7.0:prepobs/1.1.0:fit2obs/1.1.7.1:globus-cli/3.35.2:module_base.hercules', 'SLURM_JOB_ACCOUNT': 'fv3-cpu', 'DO_OCN': True, 'SLURM_PRIO_PROCESS': 0, 'HOMEfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1', 'gmake_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq', 'FIXfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/fix', 'py_python_dateutil_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy', 'tasks_per_node': 1, 'SLURM_NPROCS': 1, 'COM_CHEM_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history', 'LMOD_ROOT': '/apps/other/lmod', 'DOHYBVAR': '{{ DOHYBVAR }}', 'GSL_ROOT_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z', 'SERIAL_F77': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'IP_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_d', 'SHLVL': 4, '_ModuleTable071_': 'e30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAic3RhY2staW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMCIsCndWID0gIjAwMDAwMjAyMS4wMDAwMDAwMDkuKnpmaW5hbCIsCn0sCnRhciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC90YXIvMS4zNC5sdWEiLApmdWxsTmFtZSA9ICJ0YXIvMS4zNCIsCmxvYWRPcmRlciA9IDEwLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIs', '_ModuleTable012_': 'c3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImZmdHcvMy4zLjEwIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAwMy4wMDAwMDAwMTAuKnpmaW5hbCIsCn0sCmZpdDJvYnMgPSB7CmZuID0gIi93b3JrMi9ub2FhL2dsb2JhbC9yb2xlLWdsb2JhbC9naXQvRml0Mk9icy92MS4xLjcuMS9tb2R1bGVmaWxlcy9maXQyb2JzLzEuMS43LjEubHVhIiwKZnVsbE5hbWUgPSAiZml0Mm9icy8xLjEuNy4xIiwKbG9hZE9yZGVyID0gODIsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZml0Mm9icy8xLjEuNy4xIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4wMDAwMDAwMDcuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApnMiA9', 'INCVARS_EFOLD': 5, 'SLURM_NNODES': 1, 'RESERVATION': '', 'FHMIN_GFS': 0, 'FHMAX_HF_WAV': 48, 'FIXugwd': '/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd', 'BASH_ENV': '/apps/other/lmod/lmod/init/bash', 'CDATE': datetime.datetime(2021, 3, 23, 12, 0), '_ModuleTable063_': 'aW5hbCIsCn0sClsicHkteGxyZCJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXhscmQvMi4wLjEubHVhIiwKZnVsbE5hbWUgPSAicHkteGxyZC8yLjAuMSIsCmxvYWRPcmRlciA9IDY5LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXhscmQvMi4wLjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJweS14bHN4d3JpdGVyIl0g', '_ModuleTable006_': 'Mi4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMi4qemZpbmFsIiwKfSwKY29udHJpYiA9IHsKZm4gPSAiL2FwcHMvb3RoZXIvbW9kdWxlZmlsZXMvY29udHJpYi8wLjEiLApmdWxsTmFtZSA9ICJjb250cmliLzAuMSIsCmxvYWRPcmRlciA9IDEsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAwLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY29udHJpYiIsCndWID0gIjAwMDAwMDAwMC4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sCmNydG0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50', 'FHOUT_OCN_GFS': 6, '_ModuleTable026_': 'cmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2phc3Blci8yLjAuMzIubHVhIiwKZnVsbE5hbWUgPSAiamFzcGVyLzIuMC4zMiIsCmxvYWRPcmRlciA9IDE3LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImphc3Blci8yLjAuMzIiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAzMi4qemZpbmFsIiwKfSwKbGlianBlZyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1z', 'DONST': True, '_ModuleTable051_': 'ay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL3B5LW5ldGNkZjQvMS41LjgubHVhIiwKZnVsbE5hbWUgPSAicHktbmV0Y2RmNC8xLjUuOCIsCmxvYWRPcmRlciA9IDU4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LW5ldGNkZjQvMS41LjgiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA1LjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKWyJweS1udW1leHByIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3Rh', 'LMOD_sys': 'Linux', '_ModuleTable033_': 'L3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL25jby81LjAuNi5sdWEiLApmdWxsTmFtZSA9ICJuY28vNS4wLjYiLApsb2FkT3JkZXIgPSAzNywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJuY28vNS4wLjYiLAp3ViA9ICIwMDAwMDAwMDUuMDAwMDAwMDAwLjAwMDAwMDAwNi4qemZpbmFsIiwKfSwKWyJuZXRjZGYtYyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1', 'FHMIN_WAV': 0, 'LMOD_FAMILY_METACOMPILER_VERSION': '2021.9.0', 'I_MPI_PMI_LIBRARY': '/opt/slurm/lib/libpmi2.so', 'machine': 'HERCULES', 'py_pyyaml_ver': 6.0, 'OCL_ICD_FILENAMES': 'libintelocl_emu.so:libalteracl.so:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64/libintelocl.so', '_ModuleTable017_': 'bG9hZE9yZGVyID0gNDgsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ2l0LWxmcy8zLjEuMiIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbImdsb2J1cy1jbGkiXSA9IHsKZm4gPSAiL2FwcHMvb3RoZXIvbW9kdWxlZmlsZXMvZ2xvYnVzLWNsaS8zLjM1LjIubHVhIiwKZnVsbE5hbWUgPSAiZ2xvYnVzLWNsaS8zLjM1LjIiLApsb2FkT3JkZXIgPSA4MywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJnbG9idXMtY2xpIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAzNS4wMDAwMDAwMDIu', 'DO_FETCH_HPSS': False, 'PROJ_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/proj', 'FHOUT_ICE': 3, 'COM_WAVE_STATION_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station', 'COMINecmwf': '/work2/noaa/global/role-global/data/external_gempak/ecmwf', 'COM_TOP_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}', 'WGRIB2_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include', 'mpmd_opt': '--multi-prog --output=mpmd.%j.%t.out', 'DO_PREP_SFC': False, 'LMOD_G2TMPLVIRT_NAME': 'g2tmpl', '_ModuleTable066_': 'bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5dGhvbi8zLjExLjYubHVhIiwKZnVsbE5hbWUgPSAicHl0aG9uLzMuMTEuNiIsCmxvYWRPcmRlciA9IDE1LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5dGhvbi8zLjExLjYiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDExLjAwMDAwMDAwNi4qemZpbmFsIiwKfSwKc25hcHB5ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAy', '_ModuleTable009_': 'MjEuOS4wL2N1cmwvOC40LjAubHVhIiwKZnVsbE5hbWUgPSAiY3VybC84LjQuMCIsCmxvYWRPcmRlciA9IDIzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImN1cmwvOC40LjAiLAp3ViA9ICIwMDAwMDAwMDguMDAwMDAwMDA0Lip6ZmluYWwiLAp9LAplY2NvZGVzID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2VjY29kZXMvMi4zMi4wLmx1YSIsCmZ1bGxOYW1lID0gImVjY29kZXMvMi4z', 'py_numexpr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq', '__LMOD_REF_COUNT_MANPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/share/man:2;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/man:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/documentation/en/man/common:1;/apps/other/lmod/lmod/share/man:1;/usr/share/man:1;/apps/share/man:1;/apps/man:1;/opt/slurm/share/man:1', 'DOLETKF_OCN': False, 'MPI_GROUP_MAX': 256, 'py_cftime_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i', 'XDG_SESSION_ID': 'c20', 'zstd_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4', 'CLUSTERS': '', '_ModuleTable034_': 'bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAvbmV0Y2RmLWMvNC45LjIubHVhIiwKZnVsbE5hbWUgPSAibmV0Y2RmLWMvNC45LjIiLApsb2FkT3JkZXIgPSAzMCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJuZXRjZGYtYy80LjkuMiIsCndWID0gIjAwMDAwMDAwNC4wMDAwMDAwMDkuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbIm5ldGNkZi1jeHg0Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2st', 'COM_OCEAN_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean', 'CLUSTERS_SERVICE': '', 'pgmout': 'OUTPUT.365753', 'EXECfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/exec', '_ModuleTable031_': 'aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJtZXQvOS4xLjMiLAp3ViA9ICIwMDAwMDAwMDkuMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKbWV0cGx1cyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL21ldHBsdXMvMy4xLjEubHVhIiwKZnVsbE5hbWUgPSAibWV0cGx1cy8zLjEuMSIsCmxvYWRPcmRlciA9IDc4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIs', 'COM_ATMOS_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon', 'HOMEobsproc': '/work2/noaa/global/role-global/git/obsproc/v1.2.0', 'esmf_ver': '8.6.0', 'DO_ARCHCOM': False, 'SLURM_SUBMIT_HOST': 'hercules-login-1.hpc.msstate.edu', 'WRITE_DOPOST': True, 'EXECcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/exec', 'DUMP_SUFFIX': '', '_ModuleTable072_': 'CnVzZXJOYW1lID0gInRhci8xLjM0IiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAzNC4qemZpbmFsIiwKfSwKdWR1bml0cyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC91ZHVuaXRzLzIuMi4yOC5sdWEiLApmdWxsTmFtZSA9ICJ1ZHVuaXRzLzIuMi4yOCIsCmxvYWRPcmRlciA9IDI1LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInVkdW5pdHMvMi4yLjI4IiwKd1YgPSAiMDAwMDAwMDAy', 'COMOUT_OCEAN_GRIB': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ocean/grib2', 'CASE': 'C48', 'SENDDBN': False, 'gettext_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43', 'COM_ICE_BMATRIX_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice', 'git_lfs_ROOT': '/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l', '_ModuleTable001_': 'X01vZHVsZVRhYmxlXyA9IHsKTVR2ZXJzaW9uID0gMywKY19yZWJ1aWxkVGltZSA9IGZhbHNlLApjX3Nob3J0VGltZSA9IGZhbHNlLApkZXB0aFQgPSB7fSwKZmFtaWx5ID0gewpNZXRhQ29tcGlsZXIgPSAic3RhY2staW50ZWwiLApNZXRhTVBJID0gInN0YWNrLWludGVsLW9uZWFwaS1tcGkiLApjb21waWxlciA9ICJpbnRlbC1vbmVhcGktY29tcGlsZXJzIiwKZzJ0bXBsdmlydCA9ICJnMnRtcGwiLAptcGkgPSAiaW50ZWwtb25lYXBpLW1waSIsCn0sCm1UID0gewphbnRsciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9t', 'COM_ICE_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart', '_ModuleTable025_': 'LAp9LAppcCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9pcC80LjMuMC5sdWEiLApmdWxsTmFtZSA9ICJpcC80LjMuMCIsCmxvYWRPcmRlciA9IDQzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImlwLzQuMy4wIiwKd1YgPSAiMDAwMDAwMDA0LjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKamFzcGVyID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hl', 'COM_OCEAN_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart', 'ARCHCOM_TO': 'globus_hpss', 'COM_ATMOS_WMO_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo', 'ip_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf', 'CLASSPATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/mpi.jar', 'INTELFPGAOCLSDKROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga', 'LD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/host/linux64/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/lib:/usr/lib64:/usr/lib:/opt/slurm/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib', 'DO_CA': True, 'g2_ver': '3.4.5', 'LMOD_FAMILY_COMPILER': 'intel-oneapi-compilers', 'XDG_RUNTIME_DIR': '/run/user/9583', 'BASE_ENV': '/work2/noaa/global/mterry/global-workflow_forked/env', 'SLURM_JOB_ID': 5951672, '_ModuleTable014_': 'dWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nMmMvMS44LjAubHVhIiwKZnVsbE5hbWUgPSAiZzJjLzEuOC4wIiwKbG9hZE9yZGVyID0gNzUsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZzJjLzEuOC4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKZzJ0bXBsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9p', 'OCEANICEPRODUCTS_CONFIG': '/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml', 'NTHREADS_OCNICEPOST': 1, 'DO_JEDIOCNVAR': False, 'FIXreg2grb2': '/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2', 'SLURM_NODEID': 0, 'ip_ver': '4.3.0', 'KMP_AFFINITY': 'scatter', 'FHOUT_ICE_GFS': 6, '_ModuleTable065_': 'L3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS14bHd0LzEuMy4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXhsd3QvMS4zLjAiLApsb2FkT3JkZXIgPSA3MSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS14bHd0LzEuMy4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKcHl0aG9uID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1', 'DO_ATM': True, '_ModuleTable004_': 'IDUzLApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImJ1ZnIvMTEuNy4wIiwKd1YgPSAiMDAwMDAwMDExLjAwMDAwMDAwNy4qemZpbmFsIiwKfSwKWyJjLWJsb3NjIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvYy1ibG9zYy8xLjIxLjUubHVhIiwKZnVsbE5hbWUgPSAiYy1ibG9zYy8xLjIxLjUiLApsb2FkT3JkZXIgPSAyOSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1', 'esmf_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep', 'LMOD_FAMILY_MPI_VERSION': '2021.9.0', '_ModuleTable020_': 'c3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9nc2ktbmNkaWFnLzEuMS4yLmx1YSIsCmZ1bGxOYW1lID0gImdzaS1uY2RpYWcvMS4xLjIiLApsb2FkT3JkZXIgPSA0NiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJnc2ktbmNkaWFnLzEuMS4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCmdzbCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYu', 'DO_JEDIATMVAR': False, 'DO_AERO_ANL': False, 'NCLEN': '/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen', 'PS4': '+ $(basename ${BASH_SOURCE[0]:-${FUNCNAME[0]:-"Unknown"}})[${LINENO}]', 'COM_SNOW_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow', '_ModuleTable053_': 'YWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbnVtcHkvMS4yMy40Lmx1YSIsCmZ1bGxOYW1lID0gInB5LW51bXB5LzEuMjMuNCIsCmxvYWRPcmRlciA9IDUyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSA4LApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LW51bXB5LzEuMjMuNCIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMjMuMDAwMDAwMDA0Lip6ZmluYWwiLAp9LApbInB5LW9wZW5weXhsIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0', '__LMOD_REF_COUNT_ACL_BOARD_VENDOR_PATH': '/opt/Intel/OpenCLFPGA/oneAPI/Boards:1', 'LMOD_FAMILY_METAMPI_VERSION': '2021.9.0', '_ModuleTable059_': 'IDY4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXB5eGxzYi8xLjAuMTAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAwLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKWyJweS1weXlhbWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1weXlhbWwvNi4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXB5eWFtbC82LjAiLApsb2FkT3JkZXIgPSA2MCwKcHJvcFQgPSB7fSwKc3RhY2tE', 'DO_CALC_INCREMENT_ENKF_GFS': False, 'MKLROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0', 'DEBUGINFOD_URLS': 'https://debuginfod.centos.org/ ', 'sven_root_path': '/home/gfekete/sven', '_ModuleTable047_': 'LApbInB5LWV0LXhtbGZpbGUiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1ldC14bWxmaWxlLzEuMC4xLmx1YSIsCmZ1bGxOYW1lID0gInB5LWV0LXhtbGZpbGUvMS4wLjEiLApsb2FkT3JkZXIgPSA2NSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1ldC14bWxmaWxlLzEuMC4xIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMC4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsi', 'restart_interval_enkfgfs': 3, 'SENDDBN_NTC': False, 'FIXaer': '/work2/noaa/global/mterry/global-workflow_forked/fix/aer', 'FHMAX_HF_GFS': 48, 'COM_ICE_LETKF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf', 'LMOD_FAMILY_MPI': 'intel-oneapi-mpi', 'netcdf_fortran_ver': '4.6.1', 'FIXmom': '/work2/noaa/global/mterry/global-workflow_forked/fix/mom6', 'cyc': 12, 'BACIO_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib/libbacio_4.a', 'FC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'FHOUT_WAV_GFS': 3, 'SDATE_GFS': datetime.datetime(2021, 3, 23, 12, 0), 'BACIO_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib/libbacio_8.a', 'sp_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr', 'OPS_RES': 'C768', 'NMEM_ENS_GFS': 30, 'which_declare': 'declare -f', '_ModuleTable056_': 'OS4wL3B5LXBhbmRhcy8xLjUuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1wYW5kYXMvMS41LjMiLApsb2FkT3JkZXIgPSA3MiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1wYW5kYXMvMS41LjMiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA1LjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKWyJweS1weXRob24tZGF0ZXV0aWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1weXRob24tZGF0ZXV0aWwvMi44', 'COM_WAVE_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded', 'LMOD_FAMILY_G2TMPLVIRT': 'g2tmpl', 'QUEUE_DTN': 'batch', 'DO_GENESIS_FSU': False, 'prod_util_ver': '2.1.1', 'MPI_F90': 'mpiifort', 'LMOD_FAMILY_COMPILER_VERSION': '2023.1.0', '_ModuleTable042_': 'MDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcGlnei8yLjcubHVhIiwKZnVsbE5hbWUgPSAicGlnei8yLjciLApsb2FkT3JkZXIgPSA4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gNCwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInBpZ3ovMi43IiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4qemZpbmFsIiwKfSwKcHJlcG9icyA9IHsKZm4gPSAiL3dvcmsyL25vYWEvZ2xvYmFsL3JvbGUtZ2xvYmFsL2dpdC9wcmVwb2JzL3YxLjEuMC9tb2R1bGVmaWxlcy9wcmVwb2JzLzEuMS4wLmx1YSIsCmZ1bGxOYW1lID0gInByZXBvYnMvMS4xLjAiLApsb2FkT3JkZXIgPSA4MSwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0', 'openjpeg_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu', 'USHfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/ush', 'YAML_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib', 'DO_GSISOILDA': False, 'nemsio_ver': '2.5.4', 'DO_NPOESS': False, 'SLURM_MPI_TYPE': 'pmi2', 'OMP_STACKSIZE': 2048000, 'py_et_xmlfile_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq', 'FORECAST_HOUR': 48, 'ATARDIR': '/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW', 'EXECgfs': '/work2/noaa/global/mterry/global-workflow_forked/exec', 'py_python_dateutil_ver': '2.8.2', 'assim_freq': 6, 'DOIAU': True, 'FHOUT_WAV': 1, 'py_pyyaml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2', 'met_ver': '9.1.3', 'gsi_ncdiag_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe', 'DO_FETCH_LOCAL': False, 'NLSPATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64/locale/%l_%t/%N:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin/locale/%l_%t/%N', 'netcdf_fortran_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe', 'SP_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_4.a', 'COM_OCEAN_BMATRIX_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean', 'SLURM_CONF': '/var/spool/slurmd/conf-cache/slurm.conf', 'PATH': '/apps/other/globus-cli-3.35.2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r/ush:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/bin:/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/bin/intel64:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/bin:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/bin:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin:/usr/sbin:/usr/bin:/apps/sbin:/apps/bin:/opt/slurm/bin:/home/gfekete/sven/bin', 'SP_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_8.a', 'netcdf_c_ver': '4.9.2', 'SLURM_JOB_NAME': 'C48_S2SW_gfs_ocean_prod_f048_12', 'COM_ICE_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history', 'NLN': '/bin/ln -sf', 'MODULEPATH': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles:/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0-a66eaip/g2tmpl/1.13.0/intel/2021.9.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/g2tmpl/1.13.0/intel/2021.9.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/intel-oneapi-mpi/2021.9.0-a66eaip/oneapi/2023.1.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core:/work2/noaa/global/mterry/global-workflow_forked/modulefiles:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core:/apps/other/modulefiles:/apps/containers/modulefiles:/apps/licensed/modulefiles:/apps/contrib/modulefiles', 'CC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icc', 'grib_util_ver': '1.3.0', 'py_f90nml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z', 'DO_MERGENSST': False, 'SLURM_NTASKS_PER_NODE': 1, '_LMFILES_': '/apps/other/modulefiles/contrib/0.1:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core/intel-oneapi-compilers/2023.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core/stack-intel/2021.9.0.lua:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0/intel-oneapi-mpi/2021.9.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/stack-intel-oneapi-mpi/2021.9.0.lua:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core/intel-oneapi-mkl/2023.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/zlib/1.2.13.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/pigz/2.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/zstd/1.5.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/tar/1.34.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gettext/0.21.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libxcrypt/4.4.35.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/sqlite/3.43.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/util-linux-uuid/2.38.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/python/3.11.6.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libjpeg/2.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/jasper/2.0.32.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libpng/1.6.37.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/openjpeg/2.3.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/eccodes/2.32.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/fftw/3.3.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/nghttp2/1.57.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/curl/8.4.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/proj/9.2.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/udunits/2.2.28.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/cdo/2.2.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/hdf5/1.14.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/snappy/1.1.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/c-blosc/1.21.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-c/4.9.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-fortran/4.6.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/parallel-netcdf/1.12.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/parallelio/2.5.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/esmf/8.6.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/antlr/2.7.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gsl/2.7.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/nco/5.0.6.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/bacio/2.4.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/w3emc/2.10.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/prod_util/2.1.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2/3.4.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/sp/2.5.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/ip/4.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/grib-util/1.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2tmpl/1.13.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/gsi-ncdiag/1.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/crtm-fix/2.4.0.1_emc.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/git-lfs/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/crtm/2.4.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/openblas/0.3.24.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-setuptools/63.4.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-numpy/1.23.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/bufr/11.7.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gmake/4.2.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/wgrib2/2.0.8.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-f90nml/1.4.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-cftime/1.0.3.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/py-netcdf4/1.5.8.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libyaml/0.2.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pyyaml/6.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-markupsafe/2.1.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-jinja2/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-bottleneck/1.3.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-numexpr/2.8.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-et-xmlfile/1.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-openpyxl/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pytz/2023.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pyxlsb/1.0.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlrd/2.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlsxwriter/3.1.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlwt/1.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pandas/1.5.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-six/1.16.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-python-dateutil/2.8.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2c/1.8.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-cxx4/4.3.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/met/9.1.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/metplus/3.1.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-packaging/23.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xarray/2023.7.0.lua:/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles/prepobs/1.1.0.lua:/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles/fit2obs/1.1.7.1.lua:/apps/other/modulefiles/globus-cli/3.35.2.lua:/work2/noaa/global/mterry/global-workflow_forked/modulefiles/module_base.hercules.lua', 'globus_cli_ver': 3.27, '_ModuleTable028_': 'L2xpYnBuZy8xLjYuMzcubHVhIiwKZnVsbE5hbWUgPSAibGlicG5nLzEuNi4zNyIsCmxvYWRPcmRlciA9IDE4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImxpYnBuZy8xLjYuMzciLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA2LjAwMDAwMDAzNy4qemZpbmFsIiwKfSwKbGlieGNyeXB0ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2xpYnhjcnlwdC80LjQuMzUubHVhIiwKZnVsbE5hbWUgPSAibGlieGNyeXB0', 'py_numpy_ver': '1.23.4', '__LMOD_REF_COUNT_HDF5_PLUGIN_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/plugins:1', 'bufr_ver': '11.7.0', 'COMINnam': '/work2/noaa/global/role-global/data/external_gempak/nam', 'DBUS_SESSION_BUS_ADDRESS': 'unix:path=/run/user/9583/bus', 'DO_AERO_FCST': False, 'py_pandas_ver': '1.5.3', '_ModuleTable039_': 'dGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9vcGVuanBlZy8yLjMuMS5sdWEiLApmdWxsTmFtZSA9ICJvcGVuanBlZy8yLjMuMSIsCmxvYWRPcmRlciA9IDE5LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm9wZW5qcGVnLzIuMy4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsicGFyYWxsZWwtbmV0Y2RmIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMv', 'COM_ICE_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice', 'UUID_ORION_DTN': '8a10dd4f-24ee-4794-a39d-9c313ab6a34b', 'MPI_F77': 'mpiifort', 'LMOD_CMD': '/apps/other/lmod/lmod/libexec/lmod', '_ModuleTable045_': 'aW5hbCIsCn0sClsicHktYm90dGxlbmVjayJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LWJvdHRsZW5lY2svMS4zLjcubHVhIiwKZnVsbE5hbWUgPSAicHktYm90dGxlbmVjay8xLjMuNyIsCmxvYWRPcmRlciA9IDYzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWJvdHRsZW5lY2svMS4zLjciLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAzLjAwMDAwMDAwNy4qemZpbmFs', 'LMOD_MPI_VERSION': '2021.9.0-a66eaip', 'LMOD_SYSTEM_NAME': 'hercules', 'RUN_ENVIR': 'emc', 'w3emc_ver': '2.10.0', '__LMOD_REF_COUNT_LD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/host/linux64/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/lib:1;/usr/lib64:1;/usr/lib:1;/opt/slurm/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:2', 'DOHYBVAR_OCN': False, 'NUM_SND_COLLECTIVES': 9, 'walltime': '00:15:00', 'netcdf_cxx4_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu', 'ens_tracker_ver': 'v1.2.0', 'LMOD_SYSTEM_DEFAULT_MODULES': 'contrib', 'OMPI_MCA_plm_slurm_args': '--external-launcher', '_ModuleTable078_': 'cyIKLCAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLWE2NmVhaXAvZzJ0bXBsLzEuMTMuMC9pbnRlbC8yMDIxLjkuMCIKLCAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9nMnRtcGwvMS4xMy4wL2ludGVsLzIwMjEuOS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVz', 'FIXcpl': '/work2/noaa/global/mterry/global-workflow_forked/fix/cpl', 'G2C_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64/libg2c.so', '_ModuleTable019_': 'cmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9ncmliLXV0aWwvMS4zLjAubHVhIiwKZnVsbE5hbWUgPSAiZ3JpYi11dGlsLzEuMy4wIiwKbG9hZE9yZGVyID0gNDQsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ3JpYi11dGlsLzEuMy4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKWyJnc2ktbmNkaWFnIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMv', 'SLURM_JOB_GID': 17000, 'SLURM_GET_USER_ENV': 1, 'IAUFHRS_ENKF': [3, 6, 9], 'CPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/include:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/include:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/include', 'G2TMPL_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include', 'DO_JEDISNOWDA': False, 'SP_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_d.a', 'pigz_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq', 'COM_ATMOS_BUFR_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr', 'COM_OCEAN_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2', 'CNVGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/cnvgrib', 'COM_ATMOS_TRACK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks', 'DO_VERFRAD': True, 'pgmerr': 'errfile', 'MPI_BUFS_PER_HOST': 2048, 'DO_PREP_OBS_AERO': False, 'LMOD_FAMILY_G2TMPLVIRT_VERSION': '1.13.0', 'PNG_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu', 'COM_ATMOS_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart', 'DEGRIB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/degrib2', 'DO_ANLSTAT': False, 'LMOD_MPI_NAME': 'intel-oneapi-mpi', 'py_f90nml_ver': '1.4.3', 'NMEM_ENS_GFS_OFFSET': 20, 'SMOOTH_ENKF': False, 'FHMAX_GOES': 120, 'OLDPWD': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', 'QUEUE': 'batch', 'COM_OBS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/obs', 'DO_VMINMON': True, 'metplus_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r', 'VERBOSE': True, 'SLURM_JOB_NODELIST': 'hercules-02-53', 'CLUSTER': 'hercules', 'DBNROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn', 'FHOUT_HF_WAV': 1, 'PARMgfs': '/work2/noaa/global/mterry/global-workflow_forked/parm', 'I_MPI_HYDRA_BOOTSTRAP': 'slurm', 'APRUN_OCNICEPOST': 'srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1', 'BASH_FUNC_ml%%': '() { eval "$($LMOD_DIR/ml_cmd "$@")"\n}', 'BASH_FUNC_which%%': '() { ( alias;\n eval ${which_declare} ) | /usr/bin/which --tty-only --read-alias --read-functions --show-tilde --show-dot $@\n}', 'BASH_FUNC_module%%': '() { if [ -z "${LMOD_SH_DBG_ON+x}" ]; then\n case "$-" in \n *v*x*)\n __lmod_sh_dbg=\'vx\'\n ;;\n *v*)\n __lmod_sh_dbg=\'v\'\n ;;\n *x*)\n __lmod_sh_dbg=\'x\'\n ;;\n esac;\n fi;\n if [ -n "${__lmod_sh_dbg:-}" ]; then\n set +$__lmod_sh_dbg;\n echo "Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for Lmod\'s output" 1>&2;\n fi;\n eval "$($LMOD_CMD $LMOD_SHELL_PRGM "$@")" && eval "$(${LMOD_SETTARG_CMD:-:} -s sh)";\n __lmod_my_status=$?;\n if [ -n "${__lmod_sh_dbg:-}" ]; then\n echo "Shell debugging restarted" 1>&2;\n set -$__lmod_sh_dbg;\n fi;\n unset __lmod_sh_dbg;\n return $__lmod_my_status\n}', 'BASH_FUNC_declare_from_tmpl%%': '() { if [[ ${DEBUG_WORKFLOW:-"NO"} == "NO" ]]; then\n set +x;\n fi;\n local opts="-g";\n local OPTIND=1;\n while getopts "rx" option; do\n opts="${opts}${option}";\n done;\n shift $((OPTIND-1));\n for input in "$@";\n do\n IFS=\':\' read -ra args <<< "${input}";\n local com_var="${args[0]}";\n local template;\n local value;\n if (( ${#args[@]} > 1 )); then\n template="${args[1]}";\n else\n template="${com_var}_TMPL";\n fi;\n if [[ ! -v "${template}" ]]; then\n echo "FATAL ERROR in declare_from_tmpl: Requested template ${template} not defined!";\n exit 2;\n fi;\n value=$(echo "${!template}" | envsubst);\n declare ${opts} "${com_var}"="${value}";\n echo "declare_from_tmpl :: ${com_var}=${value}";\n done;\n set_trace\n}', 'BASH_FUNC_err_exit%%': ['() { set +eux;\n msg1=${*:-Job ${jobid} failed};\n if [[ -n "${pgm}" ]]; then\n msg1+="', 'ERROR IN ${pgm}";\n fi;\n if [[ -n "${err}" ]]; then\n msg1+=" RETURN CODE ${err}";\n fi;\n msg2="\n -------------------------------------------------------------\n -- FATAL ERROR: ${msg1}\n -- ABNORMAL EXIT at $(date) on ${HOSTNAME}\n -------------------------------------------------------------\n ";\n echo "${msg2}" 1>&2;\n module list;\n echo "" 1>&2;\n echo "${msg1}" 1>&2;\n if [[ -n "${DATA}" ]]; then\n echo "${DATA}" 1>&2;\n ls -ltr "${DATA}" 1>&2;\n else\n echo "WARNING: DATA variable not defined" 1>&2;\n fi;\n if [[ -n "${pgmout}" ]]; then\n if [[ -s errfile ]]; then\n echo "----- contents of errfile -----" >> "${pgmout}";\n cat errfile >> "${pgmout}";\n fi;\n cat "${pgmout}" 1>&2;\n else\n if [[ -s errfile ]]; then\n cat errfile 1>&2;\n fi;\n fi;\n if [[ "${SENDECF}" == "YES" ]]; then\n timeout 30 ecflow_client --msg "${ECF_NAME}: ${msg1}";\n timeout 30 ssh "${ECF_HOST}" "echo \\"${msg}2\\" >> ${ECF_JOBOUT:?}";\n fi;\n if [[ "${SENDECF}" == "YES" ]]; then\n ecflow_client --kill="${ECF_NAME:?}";\n fi;\n if [[ -n "${PBS_JOBID}" ]]; then\n qdel "${PBS_JOBID}";\n else\n if [[ -n "${SLURM_JOB_ID}" ]]; then\n scancel "${SLURM_JOB_ID}";\n fi;\n fi\n}'], 'BASH_FUNC_wait_for_file%%': '() { set +x;\n local file_name=${1:?"wait_for_file() requires a file name"};\n local sleep_interval=${2:-60};\n local max_tries=${3:-100};\n for ((iter=0; iter' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.oceanice_products -+++ config.oceanice_products[5]echo 'BEGIN: config.oceanice_products' -BEGIN: config.oceanice_products -+++ config.oceanice_products[8]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources oceanice_products -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=oceanice_products -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[978]walltime=00:15:00 -++++ config.resources[979]ntasks=1 -++++ config.resources[980]tasks_per_node=1 -++++ config.resources[981]threads_per_task=1 -++++ config.resources[982]memory=96GB -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n '' ]] -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export memory -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.oceanice_products[11]export MAX_TASKS=25 -+++ config.oceanice_products[11]MAX_TASKS=25 -+++ config.oceanice_products[13]export OCEANICEPRODUCTS_CONFIG=/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml -+++ config.oceanice_products[13]OCEANICEPRODUCTS_CONFIG=/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml -+++ config.oceanice_products[16]export NFHRS_PER_GROUP=3 -+++ config.oceanice_products[16]NFHRS_PER_GROUP=3 -+++ config.oceanice_products[18]echo 'END: config.oceanice_products' -END: config.oceanice_products -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env oceanice_products -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=oceanice_products -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 100663296 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 1 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 1 ]] -+++ HERCULES.env[34]max_threads_per_task=80 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 80 ]] -+++ HERCULES.env[40][[ 1 -gt 80 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 1' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[286]export NTHREADS_OCNICEPOST=1 -+++ HERCULES.env[286]NTHREADS_OCNICEPOST=1 -+++ HERCULES.env[287]export 'APRUN_OCNICEPOST=srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1' -+++ HERCULES.env[287]APRUN_OCNICEPOST='srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1' -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_OCEANICE_PRODUCTS[11]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[11]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[11]declare_from_tmpl -rx COMIN_OCEAN_HISTORY:COM_OCEAN_HISTORY_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_OCEAN_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/ocean/history -+ JGLOBAL_OCEANICE_PRODUCTS[12]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[12]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[12]declare_from_tmpl -rx COMOUT_OCEAN_GRIB:COM_OCEAN_GRIB_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_OCEAN_GRIB=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ocean/grib2 -+ JGLOBAL_OCEANICE_PRODUCTS[13]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[13]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[13]declare_from_tmpl -rx COMOUT_OCEAN_NETCDF:COM_OCEAN_NETCDF_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_OCEAN_NETCDF=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ocean/netcdf -+ JGLOBAL_OCEANICE_PRODUCTS[17]/work2/noaa/global/mterry/global-workflow_forked/scripts/exglobal_oceanice_products.py -2025-07-28 21:29:51,939 - INFO - root : BEGIN: __main__.main -2025-07-28 21:29:51,939 - DEBUG - root : ( ) -2025-07-28 21:29:51,947 - INFO - oceanice_products: BEGIN: OceanIceProducts.__init__ -2025-07-28 21:29:51,947 - DEBUG - oceanice_products: ( , {'SHELL': '/bin/bash', 'PDY': datetime.datetime(2021, 3, 23, 0, 0), 'DO_JEDIATMENS': False, 'sqlite_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4', 'COPYGB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/copygb', 'WORK_hercules': '/work/hercules', 'TOCGRIB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib2', 'G2_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/include_d', 'PIO_TYPENAME_VALID_VALUES': ['netcdf', 'netcdf4p', 'netcdf4c', 'pnetcdf'], 'WGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/wgrib', 'SERIAL_CXX': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icpc', 'FHMAX_WAV': 9, 'FIXlut': '/work2/noaa/global/mterry/global-workflow_forked/fix/lut', 'fftw_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj', 'snappy_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx', '_ModuleTable016_': 'bGxOYW1lID0gImdldHRleHQvMC4yMS4xIiwKbG9hZE9yZGVyID0gMTEsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ2V0dGV4dC8wLjIxLjEiLAp3ViA9ICIwMDAwMDAwMDAuMDAwMDAwMDIxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJnaXQtbGZzIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvZ2l0LWxmcy8zLjEuMi5sdWEiLApmdWxsTmFtZSA9ICJnaXQtbGZzLzMuMS4yIiwK', 'MAKE_NSSTBUFR': False, 'cdo_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e', 'curl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops', 'PIO_VERSION_MAJOR': 2, 'threads_per_task': 1, 'SLURM_JOB_USER': 'mterry', '_ModuleTable067_': 'NC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3NuYXBweS8xLjEuMTAubHVhIiwKZnVsbE5hbWUgPSAic25hcHB5LzEuMS4xMCIsCmxvYWRPcmRlciA9IDI4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInNuYXBweS8xLjEuMTAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAxLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKc3AgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45', 'COM_ICE_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}', 'COM_OCEAN_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}', '_ModuleTable022_': 'ZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9oZGY1LzEuMTQuMC5sdWEiLApmdWxsTmFtZSA9ICJoZGY1LzEuMTQuMCIsCmxvYWRPcmRlciA9IDI3LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImhkZjUvMS4xNC4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAxNC4qemZpbmFsIiwKfSwKWyJpbnRlbC1vbmVhcGktY29tcGlsZXJzIl0gPSB7CmZuID0gIi9hcHBzL3NwYWNrLW1hbmFnZWQvbW9kdWxlZmlsZXMvbGludXgtcm9ja3k5LXg4Nl82NC9Db3JlL2ludGVsLW9uZWFwaS1jb21waWxlcnMvMjAyMy4xLjAubHVhIiwKZnVsbE5hbWUgPSAiaW50ZWwtb25lYXBpLWNvbXBpbGVy', 'SLURM_TASKS_PER_NODE': 1, 'COM_ATMOS_OZNMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon', 'COM_ATMOS_GEMPAK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}', 'SLURM_JOB_UID': 9583, 'HISTCONTROL': 'ignoredups', 'SLURM_EXPORT_ENV': 'NONE', 'g2c_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp', 'COM_OCEAN_NETCDF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf', 'GRB2INDEX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grb2index', 'l4densvar': True, 'FHOUT_HF_GFS': 1, 'I_MPI_EXTRA_FILESYSTEM': 1, 'SLURM_TASK_PID': 365330, 'EXECprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/exec', 'COMOUT_OCEAN_NETCDF': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ocean/netcdf', 'g2tmpl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel', 'MPI_BUFS_PER_PROC': 2048, 'IAU_DELTHRS_ENKF': 6, 'COM_WAVE_WMO_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo', '_ModuleTable027_': 'LTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9saWJqcGVnLzIuMS4wLmx1YSIsCmZ1bGxOYW1lID0gImxpYmpwZWcvMi4xLjAiLApsb2FkT3JkZXIgPSAxNiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJsaWJqcGVnLzIuMS4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKbGlicG5nID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4w', 'NMEM_ENS': 0, '__LMOD_REF_COUNT_CLASSPATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/mpi.jar:1', '_ModuleTable050_': 'ZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbWFya3Vwc2FmZS8yLjEuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1tYXJrdXBzYWZlLzIuMS4zIiwKbG9hZE9yZGVyID0gNjEsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktbWFya3Vwc2FmZS8yLjEuMyIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDEuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LW5ldGNkZjQiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFj', 'err': 0, '_ModuleTable032_': 'CnVzZXJOYW1lID0gIm1ldHBsdXMvMy4xLjEiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJtb2R1bGVfYmFzZS5oZXJjdWxlcyJdID0gewpmbiA9ICIvd29yazIvbm9hYS9nbG9iYWwvbXRlcnJ5L2dsb2JhbC13b3JrZmxvd19mb3JrZWQvbW9kdWxlZmlsZXMvbW9kdWxlX2Jhc2UuaGVyY3VsZXMubHVhIiwKZnVsbE5hbWUgPSAibW9kdWxlX2Jhc2UuaGVyY3VsZXMiLApsb2FkT3JkZXIgPSA4NCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDAsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJtb2R1bGVfYmFzZS5oZXJjdWxlcyIsCndWID0gIk0uKnpmaW5hbCIsCn0sCm5jbyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGlj', 'CHGRP_CMD': 'chgrp rstprod', 'PKG_CONFIG_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib/pkgconfig:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/pkgconfig:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/lib/pkgconfig', 'SLURM_LOCALID': 0, 'MAX_TASKS': 25, '_ModuleTable040_': 'aW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9wYXJhbGxlbC1uZXRjZGYvMS4xMi4yLmx1YSIsCmZ1bGxOYW1lID0gInBhcmFsbGVsLW5ldGNkZi8xLjEyLjIiLApsb2FkT3JkZXIgPSAzMiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwYXJhbGxlbC1uZXRjZGYvMS4xMi4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAxMi4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCnBhcmFsbGVsaW8gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0', 'SLURM_SUBMIT_DIR': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', '_ModuleTable077_': 'LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvenN0ZC8xLjUuMi5sdWEiLApmdWxsTmFtZSA9ICJ6c3RkLzEuNS4yIiwKbG9hZE9yZGVyID0gOSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMywKc3RhY2tEZXB0aCA9IDQsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJ6c3RkLzEuNS4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwNS4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCn0sCm1wYXRoQSA9IHsKIi93b3JrMi9ub2FhL2dsb2JhbC9yb2xlLWdsb2JhbC9naXQvRml0Mk9icy92MS4xLjcuMS9tb2R1bGVmaWxlcyIKLCAiL3dvcmsyL25vYWEvZ2xvYmFsL3JvbGUtZ2xvYmFsL2dpdC9wcmVwb2JzL3YxLjEuMC9tb2R1bGVmaWxl', 'HOSTNAME': 'hercules-02-53', 'HISTSIZE': 1000, 'wgrib2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln', 'FRAC_GRID': True, 'G2_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/include_4', 'intel_oneapi_mpi_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse', 'SLURMD_NODENAME': 'hercules-02-53', '_ModuleTable080_': 'bGVzL2xpbnV4LXJvY2t5OS14ODZfNjQvb25lYXBpLzIwMjMuMS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL0NvcmUiCiwgIi93b3JrMi9ub2FhL2dsb2JhbC9tdGVycnkvZ2xvYmFsLXdvcmtmbG93X2ZvcmtlZC9tb2R1bGVmaWxlcyIKLCAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L0NvcmUiCiwgIi9hcHBzL290aGVyL21vZHVsZWZpbGVzIiwgIi9hcHBzL2NvbnRhaW5lcnMvbW9kdWxlZmlsZXMiCiwgIi9hcHBzL2xpY2Vuc2VkL21vZHVsZWZpbGVzIiwg', '__LMOD_REF_COUNT_OCL_ICD_FILENAMES': 'libintelocl_emu.so:1;libalteracl.so:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64/libintelocl.so:1', 'c_blosc_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn', 'FHOUT_OCN': 3, 'COM_OBS_JEDI': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi', 'eccodes_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj', '_ModuleTable037_': 'ay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9uZ2h0dHAyLzEuNTcuMC5sdWEiLApmdWxsTmFtZSA9ICJuZ2h0dHAyLzEuNTcuMCIsCmxvYWRPcmRlciA9IDIyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gNCwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm5naHR0cDIvMS41Ny4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDA1Ny4qemZpbmFsIiwKfSwKb3BlbmJsYXMgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2kt', 'netcdf_c_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy', '_ModuleTable070_': 'c3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzdGFjay1pbnRlbC8yMDIxLjkuMCIsCndWID0gIjAwMDAwMjAyMS4wMDAwMDAwMDkuKnpmaW5hbCIsCn0sClsic3RhY2staW50ZWwtb25lYXBpLW1waSJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3N0YWNrLWludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAubHVhIiwKZnVsbE5hbWUgPSAic3RhY2staW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMCIsCmxvYWRPcmRlciA9IDUsCnByb3BUID0g', 'SLURM_JOB_START_TIME': 1753756157, '_ModuleTable011_': 'b3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZXNtZi84LjYuMCIsCndWID0gIjAwMDAwMDAwOC4wMDAwMDAwMDYuKnpmaW5hbCIsCn0sCmZmdHcgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9mZnR3LzMuMy4xMC5sdWEiLApmdWxsTmFtZSA9ICJmZnR3LzMuMy4xMCIsCmxvYWRPcmRlciA9IDIxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwK', 'DO_NEST': False, 'STMP': '/work2/noaa/stmp/mterry/HERCULES', 'HYDRA_LAUNCHER_EXTRA_ARGS': '--external-launcher', 'ACL_BOARD_VENDOR_PATH': '/opt/Intel/OpenCLFPGA/oneAPI/Boards', '_ModuleTable060_': 'ZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1weXlhbWwvNi4wIiwKd1YgPSAiMDAwMDAwMDA2Lip6ZmluYWwiLAp9LApbInB5LXNldHVwdG9vbHMiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1zZXR1cHRvb2xzLzYzLjQuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1zZXR1cHRvb2xzLzYzLjQuMyIsCmxvYWRPcmRlciA9IDUxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIs', '_ModuleTable007_': 'ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9jcnRtLzIuNC4wLjEubHVhIiwKZnVsbE5hbWUgPSAiY3J0bS8yLjQuMC4xIiwKbG9hZE9yZGVyID0gNDksCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY3J0bS8yLjQuMC4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNC4wMDAwMDAwMDAuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbImNydG0tZml4Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45', 'G2TMPL_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib/libg2tmpl.a', 'DO_WAVE': True, 'SCRgfs': '/work2/noaa/global/mterry/global-workflow_forked/scripts', 'SP_INC8': 'include_8', '_ModuleTable002_': 'b2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9hbnRsci8yLjcuNy5sdWEiLApmdWxsTmFtZSA9ICJhbnRsci8yLjcuNyIsCmxvYWRPcmRlciA9IDM1LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImFudGxyLzIuNy43IiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4wMDAwMDAwMDcuKnpmaW5hbCIsCn0sCmJhY2lvID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2JhY2lvLzIuNC4xLmx1', 'MPICC': 'mpiicc', 'HDF5_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh', '_ModuleTable055_': 'L21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXBhY2thZ2luZy8yMy4xLmx1YSIsCmZ1bGxOYW1lID0gInB5LXBhY2thZ2luZy8yMy4xIiwKbG9hZE9yZGVyID0gNzksCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcGFja2FnaW5nLzIzLjEiLAp3ViA9ICIwMDAwMDAwMjMuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbInB5LXBhbmRhcyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEu', 'SP_INC4': 'include_4', 'FPGA_VARS_DIR': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga', 'FHOUT': 3, '__LMOD_REF_COUNT_MODULEPATH': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles:1;/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0-a66eaip/g2tmpl/1.13.0/intel/2021.9.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/g2tmpl/1.13.0/intel/2021.9.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/intel-oneapi-mpi/2021.9.0-a66eaip/oneapi/2023.1.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core:1;/work2/noaa/global/mterry/global-workflow_forked/modulefiles:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core:1;/apps/other/modulefiles:1;/apps/containers/modulefiles:1;/apps/licensed/modulefiles:1;/apps/contrib/modulefiles:1', 'cycle': 't12z', '_ModuleTable043_': 'aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcmVwb2JzLzEuMS4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKcHJvZF91dGlsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3Byb2RfdXRpbC8yLjEuMS5sdWEiLApmdWxsTmFtZSA9ICJwcm9kX3V0aWwvMi4xLjEiLApsb2FkT3JkZXIgPSA0MCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcm9kX3V0aWwvMi4x', 'MPI_CC': 'mpiicc', 'PARTITION_DTN': '', 'sp_ver': '2.5.0', 'SHELLOPTS': 'braceexpand:errexit:hashall:interactive-comments:nounset:xtrace', 'DO_AWIPS': False, 'EDATE': datetime.datetime(2021, 3, 23, 12, 0), '__LMOD_REF_COUNT___INTEL_POST_CFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1'], 'I_MPI_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0', 'CRTM_FIX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2/fix', '_ModuleTable062_': 'MTYuMCIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMTYuKnpmaW5hbCIsCn0sClsicHkteGFycmF5Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHkteGFycmF5LzIwMjMuNy4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXhhcnJheS8yMDIzLjcuMCIsCmxvYWRPcmRlciA9IDgwLApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXhhcnJheS8yMDIzLjcuMCIsCndWID0gIjAwMDAwMjAyMy4wMDAwMDAwMDcuKnpm', 'proj_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez', 'FSYNC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file', '_ModuleTable005_': 'cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJjLWJsb3NjLzEuMjEuNSIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMjEuMDAwMDAwMDA1Lip6ZmluYWwiLAp9LApjZG8gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9jZG8vMi4yLjAubHVhIiwKZnVsbE5hbWUgPSAiY2RvLzIuMi4wIiwKbG9hZE9yZGVyID0gMjYsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY2RvLzIu', 'g2tmpl_ver': '1.13.0', 'FHOUT_GOES': 3, '_ModuleTable021_': 'MC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nc2wvMi43LjEubHVhIiwKZnVsbE5hbWUgPSAiZ3NsLzIuNy4xIiwKbG9hZE9yZGVyID0gMzYsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ3NsLzIuNy4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sCmhkZjUgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxl', 'EUPD_CYC': 'gdas', 'fms_ver': '2023.02.01', 'util_linux_uuid_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj', '_ModuleTable052_': 'Y2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbnVtZXhwci8yLjguNC5sdWEiLApmdWxsTmFtZSA9ICJweS1udW1leHByLzIuOC40IiwKbG9hZE9yZGVyID0gNjQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktbnVtZXhwci8yLjguNCIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDguMDAwMDAwMDA0Lip6ZmluYWwiLAp9LApbInB5LW51bXB5Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2kt', '_ModuleTable030_': 'cFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJsaWJ5YW1sLzAuMi41IiwKd1YgPSAiMDAwMDAwMDAwLjAwMDAwMDAwMi4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCm1ldCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL21ldC85LjEuMy5sdWEiLApmdWxsTmFtZSA9ICJtZXQvOS4xLjMiLApsb2FkT3JkZXIgPSA3NywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0', 'COM_ATMOS_IMAGERY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery', 'I_MPI_EXTRA_FILESYSTEM_LIST': 'lustre', 'COM_WAVE_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history', 'SLURM_CLUSTER_NAME': 'hercules', '__LMOD_REF_COUNT_ACLOCAL_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/aclocal:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/aclocal:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/aclocal:2', 'SERIAL_CC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icc', 'COM_OBSPROC_TMPL': '${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos', 'SLURM_JOB_END_TIME': 1753757057, '_ModuleTable075_': 'ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3dncmliMi8yLjAuOC5sdWEiLApmdWxsTmFtZSA9ICJ3Z3JpYjIvMi4wLjgiLApsb2FkT3JkZXIgPSA1NSwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJ3Z3JpYjIvMi4wLjgiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKemxpYiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJj', 'FHMAX_GFS': 120, 'SENDAWIP': False, 'G2_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64/libg2_d.a', 'spack_mod_path': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core', 'OCNRES': 500, 'COM_MED_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart', 'SLURM_CPUS_ON_NODE': 1, 'LMOD_FAMILY_METAMPI': 'stack-intel-oneapi-mpi', 'LEVS': 128, 'FIXgsi': '/work2/noaa/global/mterry/global-workflow_forked/fix/gsi', 'NFHRS_PER_GROUP': 3, 'COM_WAVE_PREP_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep', 'FI_PROVIDER_PATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib/prov:/usr/lib64/libfabric', 'python_ver': '3.11.6', 'IAU_OFFSET': 0, 'py_packaging_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6', 'DIAGUTIL_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/sys_check/sys_check.sh', 'python_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2', 'SCRIPTScfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/scripts', 'py_markupsafe_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7', 'G2C_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include', 'py_xlrd_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f', 'LMOD_G2TMPLVIRT_VERSION': '1.13.0', 'pid': 365754, 'OUTPUT_GRID': 'gaussian_grid', 'jobid': 'oceanice_products_ocean_f072.365351', 'SLURM_JOB_CPUS_PER_NODE': 1, 'spack_env': 'gsi-addon-dev-fms-2024.01', 'INTEL_ONEAPI_MPI_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse', 'LMOD_FAMILY_METACOMPILER': 'stack-intel', 'SP_INCd': 'include_d', 'py_six_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o', 'COM_ATMOS_GOES_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim', 'MPICXX': 'mpiicpc', 'restart_interval_gdas': 3, 'fit2obs_ver': '1.1.7.1', 'LMOD_DIR': '/apps/other/lmod/lmod/libexec', '_ModuleTable068_': 'LjAvc3AvMi41LjAubHVhIiwKZnVsbE5hbWUgPSAic3AvMi41LjAiLApsb2FkT3JkZXIgPSA0MiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMiwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzcC8yLjUuMCIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCnNxbGl0ZSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9zcWxpdGUvMy40My4yLmx1YSIsCmZ1bGxOYW1lID0gInNxbGl0ZS8zLjQzLjIiLApsb2FkT3Jk', 'REPLAY_ICS': False, 'COM_ICE_NETCDF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf', 'G2_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64/libg2_4.a', '_ModuleTable058_': 'InB5LXB5dHovMjAyMy4zIiwKbG9hZE9yZGVyID0gNjcsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcHl0ei8yMDIzLjMiLAp3ViA9ICIwMDAwMDIwMjMuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LXB5eGxzYiJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXB5eGxzYi8xLjAuMTAubHVhIiwKZnVsbE5hbWUgPSAicHktcHl4bHNiLzEuMC4xMCIsCmxvYWRPcmRlciA9', 'py_setuptools_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr', 'FHMIN': 0, 'COM_ICE_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input', '_ModuleTable048_': 'cHktZjkwbm1sIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktZjkwbm1sLzEuNC4zLmx1YSIsCmZ1bGxOYW1lID0gInB5LWY5MG5tbC8xLjQuMyIsCmxvYWRPcmRlciA9IDU2LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWY5MG5tbC8xLjQuMyIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMDQuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LWppbmphMiJdID0gewpmbiA9ICIvd29yay9ub2FhL2Vw', 'intel_mkl_ver': '2023.1.0', 'PRTE_MCA_plm_slurm_args': '--external-launcher', 'PWD': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ocean_f072.365351', 'SLURM_GTIDS': 0, 'LOGNAME': 'mterry', 'MAKE_ACFTBUFR': False, 'W3EMC_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_8', 'HOMEDIR': '/work2/noaa/global/mterry', 'IAU_DELTHRS': 6, 'W3EMC_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_4', 'XDG_SESSION_TYPE': 'unspecified', 'ESMFMKFILE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib/esmf.mk', 'FHMAX': 9, 'WRITE_NSFLIP': True, 'SLURM_JOB_PARTITION': 'hercules', 'YAML_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x', 'PARTITION_SERVICE': 'service', 'MODULESHOME': '/apps/other/lmod/lmod', '__LMOD_REF_COUNT_DYLD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:1', 'COM_WAVE_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart', 'BUFR_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_4.so', 'BUFR_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_8.so', 'HOMEcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1', 'COM_OCEAN_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon', 'DEBUG_POSTSCRIPT': False, 'MANPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/share/man:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/man:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/documentation/en/man/common:/apps/other/lmod/lmod/share/man:/usr/share/man:/apps/share/man:/apps/man:/opt/slurm/share/man::', 'libyaml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x', 'jasper_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt', 'py_xlsxwriter_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok', 'SLURM_TRES_PER_TASK': 'cpu=1', 'nco_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq', 'stack_intel_ver': '2021.9.0', 'SLURM_OOM_KILL_STEP': 0, 'BUFR_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m', 'bufr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m', 'CLUSTERS_DTN': '', 'grib_util_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35', 'nghttp2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky', 'DO_TEST_MODE': True, 'PACKAGEROOT': '/work2/noaa/global/role-global/nwpara', 'lobsdiag_forenkf': True, 'SLURM_JOB_NUM_NODES': 1, 'py_netcdf4_ver': '1.5.8', 'ICERES': 500, 'CXX': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icpc', 'GRIB2GRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grib2grib2', 'DMPDIR': '/work/noaa/rstprod/dump', 'LSOIL_INCR': 2, 'metplus_ver': '3.1.1', '_ModuleTable008_': 'LjAvY3J0bS1maXgvMi40LjAuMV9lbWMubHVhIiwKZnVsbE5hbWUgPSAiY3J0bS1maXgvMi40LjAuMV9lbWMiLApsb2FkT3JkZXIgPSA0NywKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJjcnRtLWZpeC8yLjQuMC4xX2VtYyIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDQuMDAwMDAwMDAwLjAwMDAwMDAwMS4qXy4qZW1jLip6ZmluYWwiLAp9LApjdXJsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIw', 'COM_OCEAN_LETKF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf', 'zlib_ver': '1.2.13', 'BUFR_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_d.so', 'ENKF_SPREAD': True, 'zlib_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j', '__LMOD_REF_COUNT_CMAKE_PREFIX_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r:2;/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/IntelDPCPP:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72:1', '_ModuleTable035_': 'c3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9uZXRjZGYtY3h4NC80LjMuMS5sdWEiLApmdWxsTmFtZSA9ICJuZXRjZGYtY3h4NC80LjMuMSIsCmxvYWRPcmRlciA9IDc2LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm5ldGNkZi1jeHg0LzQuMy4xIiwKd1YgPSAiMDAwMDAwMDA0LjAwMDAwMDAwMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsibmV0Y2RmLWZvcnRyYW4iXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFj', 'FIXgfs': '/work2/noaa/global/mterry/global-workflow_forked/fix', 'DO_COUPLED': True, 'SLURM_JOBID': 5951676, 'SERIAL_FC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'COM_ATMOS_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input', 'W3EMC_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_d', '_ModuleTable013_': 'IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nMi8zLjQuNS5sdWEiLApmdWxsTmFtZSA9ICJnMi8zLjQuNSIsCmxvYWRPcmRlciA9IDQxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImcyLzMuNC41IiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAwNC4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCmcyYyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJj', 'COM_ATMOS_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history', 'FIXam': '/work2/noaa/global/mterry/global-workflow_forked/fix/am', 'hdf5_ver': '1.14.0', 'restart_interval_enkfgdas': 3, 'COM_ATMOS_RADMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon', 'DO_CALC_INCREMENT': False, 'DO_STARTMEM_FROM_JEDIICE': False, 'FIXprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/fix', 'SLURM_JOB_QOS': 'batch', 'I_MPI_HYDRA_BOOTSTRAP_EXEC_EXTRA_ARGS': '--external-launcher', 'obsproc_run_ver': '1.2.0', 'COM_ATMOS_MINMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon', 'crtm_ver': '2.4.0.1', 'FIXorog': '/work2/noaa/global/mterry/global-workflow_forked/fix/orog', 'COM_OCEAN_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history', '__LMOD_REF_COUNT_PATH': '/apps/other/globus-cli-3.35.2/bin:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r/ush:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/bin:2;/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/bin:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/bin/intel64:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/bin:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/bin:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/bin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/bin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin:1;/usr/sbin:1;/usr/bin:1;/apps/sbin:1;/apps/bin:1;/opt/slurm/bin:1;/home/gfekete/sven/bin:1', '_ModuleTable018_': 'KnpmaW5hbCIsCn0sCmdtYWtlID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2dtYWtlLzQuMi4xLmx1YSIsCmZ1bGxOYW1lID0gImdtYWtlLzQuMi4xIiwKbG9hZE9yZGVyID0gNTQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ21ha2UvNC4yLjEiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDAyLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJncmliLXV0aWwiXSA9IHsKZm4gPSAiL3dv', 'COMROOT': '/work2/noaa/global/mterry/RUNTESTS/COMROOT', 'HOME': '/home/mterry', 'FI_PROVIDER': 'mlx', 'met_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b', 'COM_RTOFS_TMPL': '${DMPDIR}', '_ModuleTable015_': 'bnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2cydG1wbC8xLjEzLjAubHVhIiwKZnVsbE5hbWUgPSAiZzJ0bXBsLzEuMTMuMCIsCmxvYWRPcmRlciA9IDQ1LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImcydG1wbC8xLjEzLjAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDEzLip6ZmluYWwiLAp9LApnZXR0ZXh0ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2dldHRleHQvMC4yMS4xLmx1YSIsCmZ1', '_ModuleTable_Sz_': 81, 'LANG': 'C.UTF-8', '__LMOD_REF_COUNT_LIBRARY_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:1', 'DO_BUFRSND': False, 'memory': '96GB', '_ModuleTable064_': 'PSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHkteGxzeHdyaXRlci8zLjEuNy5sdWEiLApmdWxsTmFtZSA9ICJweS14bHN4d3JpdGVyLzMuMS43IiwKbG9hZE9yZGVyID0gNzAsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHkteGxzeHdyaXRlci8zLjEuNyIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDA3Lip6ZmluYWwiLAp9LApbInB5LXhsd3QiXSA9IHsKZm4gPSAi', 'DYLD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib', 'PARTITION_BATCH': 'hercules', '_ModuleTable023_': 'cy8yMDIzLjEuMCIsCmxvYWRPcmRlciA9IDIsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiaW50ZWwtb25lYXBpLWNvbXBpbGVycy8yMDIzLjEuMCIsCndWID0gIjAwMDAwMjAyMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsiaW50ZWwtb25lYXBpLW1rbCJdID0gewpmbiA9ICIvYXBwcy9zcGFjay1tYW5hZ2VkL21vZHVsZWZpbGVzL2xpbnV4LXJvY2t5OS14ODZfNjQvQ29yZS9pbnRlbC1vbmVhcGktbWtsLzIwMjMuMS4wLmx1YSIsCmZ1bGxOYW1lID0gImludGVsLW9uZWFwaS1ta2wvMjAyMy4xLjAiLApsb2FkT3JkZXIgPSA2LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVz', '_ModuleTable081_': 'Ii9hcHBzL2NvbnRyaWIvbW9kdWxlZmlsZXMiLAp9LApzeXN0ZW1CYXNlTVBBVEggPSAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L0NvcmU6L2FwcHMvb3RoZXIvbW9kdWxlZmlsZXM6L2FwcHMvY29udGFpbmVycy9tb2R1bGVmaWxlczovYXBwcy9saWNlbnNlZC9tb2R1bGVmaWxlcyIsCn0K', 'REDERR': '2>', 'PSLOT': 'C48_S2SW', '__LMOD_REF_COUNT_PKG_CONFIG_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib/pkgconfig:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/pkgconfig:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib/pkgconfig:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/pkgconfig:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib/pkgconfig:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/lib/pkgconfig:1', '_ModuleTable036_': 'ay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL25ldGNkZi1mb3J0cmFuLzQuNi4xLmx1YSIsCmZ1bGxOYW1lID0gIm5ldGNkZi1mb3J0cmFuLzQuNi4xIiwKbG9hZE9yZGVyID0gMzEsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAibmV0Y2RmLWZvcnRyYW4vNC42LjEiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDA2LjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKbmdodHRwMiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFj', 'COM_ATMOS_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}', 'USHcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/ush', 'DO_GEMPAK': False, 'SLURM_PROCID': 0, 'USHgfs': '/work2/noaa/global/mterry/global-workflow_forked/ush', 'DOIBP_WAV': False, 'prepobs_run_ver': '1.1.0', 'EXPDIR': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', 'job': 'oceanice_products', 'imp_physics': 8, 'LMOD_SHELL_PRGM': 'bash', 'IAU_FHROT': 0, 'MPIF90': 'mpiifort', 'IP_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_4.a', 'waveGRD': 'uglo_100km', 'COM_CHEM_BMAT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix', 'jasper_ver': '2.0.32', 'SCRATCH': '/scratch/hercules', 'IP_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_8.a', 'CLIENT_GLOBUS_UUID': '869912fe-f6de-46c0-af10-b22efd84a022', 'ACCOUNT': 'fv3-cpu', 'HOMEgfs': '/work2/noaa/global/mterry/global-workflow_forked', 'NHOUR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/nhour', 'binary_diag': False, 'COM_WAVE_GRID_RES_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}', 'SCRIPTSprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/scripts', 'MP_STDOUTMODE': 'ORDERED', 'sigio_ver': '2.3.2', '__LMOD_REF_COUNT_CPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/include:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/include:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/include:1', 'LMOD_SETTARG_FULL_SUPPORT': False, 'COMINsyn': '/work2/noaa/global/role-global/com/gfs/prod/syndat', 'OFFSET_START_HOUR': 0, 'REDOUT': '1>', 'PTMP': '/work2/noaa/stmp/mterry/HERCULES', '_ModuleTable076_': 'dWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC96bGliLzEuMi4xMy5sdWEiLApmdWxsTmFtZSA9ICJ6bGliLzEuMi4xMyIsCmxvYWRPcmRlciA9IDcsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEwLApzdGFja0RlcHRoID0gNSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInpsaWIvMS4yLjEzIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMi4wMDAwMDAwMTMuKnpmaW5hbCIsCn0sCnpzdGQgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2', 'g2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato', 'MODE': 'forecast-only', 'DATAROOT': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312', 'bacio_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj', 'APP': 'S2SW', '_ModuleTable029_': 'LzQuNC4zNSIsCmxvYWRPcmRlciA9IDEyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImxpYnhjcnlwdC80LjQuMzUiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDA0LjAwMDAwMDAzNS4qemZpbmFsIiwKfSwKbGlieWFtbCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9saWJ5YW1sLzAuMi41Lmx1YSIsCmZ1bGxOYW1lID0gImxpYnlhbWwvMC4yLjUiLApsb2FkT3JkZXIgPSA1OSwKcHJv', 'openblas_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f', 'stack_impi_ver': '2021.9.0', 'COM_CHEM_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem', 'TMPDIR': '/local/scratch/mterry/5951676', 'HDF5_PLUGIN_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/plugins', 'DO_TRACKER': True, 'CMAKE_PREFIX_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r:/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/IntelDPCPP:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72', 'crtm_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r', '_ModuleTable038_': 'YWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvb3BlbmJsYXMvMC4zLjI0Lmx1YSIsCmZ1bGxOYW1lID0gIm9wZW5ibGFzLzAuMy4yNCIsCmxvYWRPcmRlciA9IDUwLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm9wZW5ibGFzLzAuMy4yNCIsCndWID0gIjAwMDAwMDAwMC4wMDAwMDAwMDMuMDAwMDAwMDI0Lip6ZmluYWwiLAp9LApvcGVuanBlZyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5z', 'nco_ver': '5.0.6', '_ModuleTable046_': 'IiwKfSwKWyJweS1jZnRpbWUiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1jZnRpbWUvMS4wLjMuNC5sdWEiLApmdWxsTmFtZSA9ICJweS1jZnRpbWUvMS4wLjMuNCIsCmxvYWRPcmRlciA9IDU3LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWNmdGltZS8xLjAuMy40IiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMC4wMDAwMDAwMDMuMDAwMDAwMDA0Lip6ZmluYWwiLAp9', 'GRBINDEX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grbindex', 'libjpeg_turbo_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy', 'MPIF77': 'mpiifort', '_ModuleTable073_': 'LjAwMDAwMDAwMi4wMDAwMDAwMjguKnpmaW5hbCIsCn0sClsidXRpbC1saW51eC11dWlkIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvdXRpbC1saW51eC11dWlkLzIuMzguMS5sdWEiLApmdWxsTmFtZSA9ICJ1dGlsLWxpbnV4LXV1aWQvMi4zOC4xIiwKbG9hZE9yZGVyID0gMTQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAidXRpbC1saW51eC11dWlkLzIuMzguMSIsCndWID0gIjAw', '_ModuleTable010_': 'Mi4wIiwKbG9hZE9yZGVyID0gMjAsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZWNjb2Rlcy8yLjMyLjAiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDMyLip6ZmluYWwiLAp9LAplc21mID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAvZXNtZi84LjYuMC5sdWEiLApmdWxsTmFtZSA9ICJlc21mLzguNi4wIiwKbG9hZE9yZGVyID0gMzQsCnBy', 'NTHSTACK': 1024000000, 'SLURM_CPUS_PER_TASK': 1, 'FIXcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/fix', 'SLURM_NTASKS': 1, 'DO_ICE': True, 'NET': 'gfs', 'COM_ATMOS_GENESIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital', 'HOMEpost': '/work2/noaa/global/mterry/global-workflow_forked', '_ModuleTable061_': 'CnVzZXJOYW1lID0gInB5LXNldHVwdG9vbHMvNjMuNC4zIiwKd1YgPSAiMDAwMDAwMDYzLjAwMDAwMDAwNC4wMDAwMDAwMDMuKnpmaW5hbCIsCn0sClsicHktc2l4Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktc2l4LzEuMTYuMC5sdWEiLApmdWxsTmFtZSA9ICJweS1zaXgvMS4xNi4wIiwKbG9hZE9yZGVyID0gNzMsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktc2l4LzEu', 'bacio_ver': '2.4.1', 'cmake_ver': '3.23.1', 'UTILROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq', '_ModuleTable024_': 'ZXJOYW1lID0gImludGVsLW9uZWFwaS1ta2wvMjAyMy4xLjAiLAp3ViA9ICIwMDAwMDIwMjMuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbImludGVsLW9uZWFwaS1tcGkiXSA9IHsKZm4gPSAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L29uZWFwaS8yMDIzLjEuMC9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLmx1YSIsCmZ1bGxOYW1lID0gImludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAiLApsb2FkT3JkZXIgPSA0LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAiLAp3ViA9ICIwMDAwMDIwMjEuMDAwMDAwMDA5Lip6ZmluYWwi', 'tar_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths', 'ntasks': 1, 'max_tasks_per_node': 80, 'NCDUMP': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump', '_ModuleTable057_': 'LjIubHVhIiwKZnVsbE5hbWUgPSAicHktcHl0aG9uLWRhdGV1dGlsLzIuOC4yIiwKbG9hZE9yZGVyID0gNzQsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcHl0aG9uLWRhdGV1dGlsLzIuOC4yIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwOC4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sClsicHktcHl0eiJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXB5dHovMjAyMy4zLmx1YSIsCmZ1bGxOYW1lID0g', 'py_numpy_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne', 'QUILTING': True, 'IP_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_d.a', 'SLURM_TOPOLOGY_ADDR': 'hercules-02-53', 'libxcrypt_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr', 'CRTM_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include', 'COMIN_OCEAN_HISTORY': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/ocean/history', 'LMOD_VERSION': '8.7.14', 'spack_stack_ver': '1.6.0', '_ModuleTable041_': 'LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9wYXJhbGxlbGlvLzIuNS4xMC5sdWEiLApmdWxsTmFtZSA9ICJwYXJhbGxlbGlvLzIuNS4xMCIsCmxvYWRPcmRlciA9IDMzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInBhcmFsbGVsaW8vMi41LjEwIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNS4wMDAwMDAwMTAuKnpmaW5hbCIsCn0sCnBpZ3ogPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0y', 'HOMEprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0', 'COMPONENT': 'ocean', 'BASE_GIT': '/work2/noaa/global/role-global/git', 'FCST_SEGMENTS': [0, 120], 'libpng_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu', '_ModuleTable044_': 'LjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKcHJvaiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9wcm9qLzkuMi4xLmx1YSIsCmZ1bGxOYW1lID0gInByb2ovOS4yLjEiLApsb2FkT3JkZXIgPSAyNCwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcm9qLzkuMi4xIiwKd1YgPSAiMDAwMDAwMDA5LjAwMDAwMDAwMi4wMDAwMDAwMDEuKnpm', 'DO_GENESIS': True, 'BUFR_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_4', '__LMOD_REF_COUNT_DIAGUTIL_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/sys_check/sys_check.sh:1', 'COM_WAVE_GEMPAK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak', 'COM_ATMOS_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2', 'RUN': 'gfs', 'BUFR_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_8', 'COM_ICE_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon', 'py_pandas_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw', '__INTEL_POST_CFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64'], 'ARCDIR': '/work2/noaa/global/mterry/archive/C48_S2SW', '__LMOD_REF_COUNT_NLSPATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64/locale/%l_%t/%N:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin/locale/%l_%t/%N:1', '_ModuleTable069_': 'ZXIgPSAxMywKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMiwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzcWxpdGUvMy40My4yIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDA0My4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sClsic3RhY2staW50ZWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9Db3JlL3N0YWNrLWludGVsLzIwMjEuOS4wLmx1YSIsCmZ1bGxOYW1lID0gInN0YWNrLWludGVsLzIwMjEuOS4wIiwKbG9hZE9yZGVyID0gMywKcHJvcFQgPSB7fSwK', 'SENDECF': False, 'parallelio_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed', 'py_pytz_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s', 'ILPOST': 3, 'WGRIB2_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib/libwgrib2.a', 'parallel_netcdf_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3', 'CMPLR_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0', 'prod_util_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq', 'COMINukmet': '/work2/noaa/global/role-global/data/external_gempak/ukmet', 'libpng_ver': '1.6.37', 'COPYGB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/copygb2', 'NMV': '/bin/mv', 'NOSCRUB': '/work2/noaa/global/mterry', 'SCRATCH_hercules': '/scratch/hercules', 'W3EMC_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_d.a', 'SENDSDM': False, 'envir': 'prod', 'DO_METP': False, 'CRTM_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib/libcrtm.a', 'LOGSCRIPT': '', 'INTEL_ONEAPI_COMPILERS_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72', '_ModuleTable074_': 'MDAwMDAwMi4wMDAwMDAwMzguMDAwMDAwMDAxLip6ZmluYWwiLAp9LAp3M2VtYyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC93M2VtYy8yLjEwLjAubHVhIiwKZnVsbE5hbWUgPSAidzNlbWMvMi4xMC4wIiwKbG9hZE9yZGVyID0gMzksCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAidzNlbWMvMi4xMC4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKd2dyaWIy', 'py_jinja2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj', 'FPGA_VARS_ARGS': '', 'INCVARS_ZERO_STRAT': ["'sphum_inc'", "'liq_wat_inc'", "'icmr_inc'", "'rwmr_inc'", "'snmr_inc'", "'grle_inc'"], 'COM_CONF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf', '_ModuleTable003_': 'YSIsCmZ1bGxOYW1lID0gImJhY2lvLzIuNC4xIiwKbG9hZE9yZGVyID0gMzgsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAzLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiYmFjaW8vMi40LjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDA0LjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKYnVmciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9idWZyLzExLjcuMC5sdWEiLApmdWxsTmFtZSA9ICJidWZyLzExLjcuMCIsCmxvYWRPcmRlciA9', 'HYDRA_BOOTSTRAP': 'slurm', '__LMOD_REF_COUNT___INTEL_POST_FFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1'], 'COM_CHEM_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon', 'MDATE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/mdate', 'WGRIB2': 'wgrib2', '_ModuleTable054_': 'LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktb3BlbnB5eGwvMy4xLjIubHVhIiwKZnVsbE5hbWUgPSAicHktb3BlbnB5eGwvMy4xLjIiLApsb2FkT3JkZXIgPSA2NiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1vcGVucHl4bC8zLjEuMiIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbInB5LXBhY2thZ2luZyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxs', 'EXP_WARM_START': False, 'MODULEPATH_ROOT': '/apps/other/modulefiles', 'COM_SNOW_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon', 'CHGRP_RSTPROD': True, 'py_pyxlsb_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge', 'BACIO_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/include_8', 'restart_interval_gfs': 12, 'DO_GOES': False, 'pgm': '', 'SLURM_TOPOLOGY_ADDR_PATTERN': 'node', 'DO_LAND_IAU': False, 'ntiles': 6, 'FHCYC': 24, 'ROTDIR': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW', 'FHOUT_AERO': 3, 'py_jinja2_ver': '3.1.2', 'BACIO_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/include_4', 'INCREMENTS_TO_ZERO': ["'liq_wat_inc'", "'icmr_inc'", "'rwmr_inc'", "'snmr_inc'", "'grle_inc'"], 'FHR_LIST': 72, 'USHprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/ush', 'BUFR_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_d', 'hdf5_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh', 'INTEL_ONEAPI_MKL_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci', 'SCRIPTSfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/scripts', '__LMOD_REF_COUNT_FI_PROVIDER_PATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib/prov:1;/usr/lib64/libfabric:1', 'FIXcice': '/work2/noaa/global/mterry/global-workflow_forked/fix/cice', 'IAUFHRS': [6], 'XDG_SESSION_CLASS': 'background', 'INTERVAL_GFS': 6, 'py_openpyxl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe', 'FHMAX_FITS': 120, 'COM_ATMOS_MASTER_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master', 'LMOD_PKG': '/apps/other/lmod/lmod', 'MPI_MEMMAP_OFF': 1, 'UUID_HERCULES_DTN': '869912fe-f6de-46c0-af10-b22efd84a022', '_ModuleTable079_': 'L3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAiCiwgIi9hcHBzL3NwYWNrLW1hbmFnZWQvbW9kdWxlZmlsZXMvbGludXgtcm9ja3k5LXg4Nl82NC9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLWE2NmVhaXAvb25lYXBpLzIwMjMuMS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wIgosICIvYXBwcy9zcGFjay1tYW5hZ2VkL21vZHVsZWZp', 'SDATE': datetime.datetime(2021, 3, 23, 12, 0), 'SLURM_SCRIPT_CONTEXT': 'prolog_task', 'MPI_CXX': 'mpiicpc', 'lwrite4danl': True, 'SLURM_MEM_PER_NODE': 98304, 'CASE_ENS': '{{ CASE_ENS }}', '_ModuleTable049_': 'aWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LWppbmphMi8zLjEuMi5sdWEiLApmdWxsTmFtZSA9ICJweS1qaW5qYTIvMy4xLjIiLApsb2FkT3JkZXIgPSA2MiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1qaW5qYTIvMy4xLjIiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDAxLjAwMDAwMDAwMi4qemZpbmFsIiwKfSwKWyJweS1tYXJrdXBzYWZlIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3Vs', 'YAML_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include', 'PYTHONPATH': '/apps/other/globus-cli-3.35.2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:/work2/noaa/global/mterry/global-workflow_forked/sorc/wxflow/src:/work2/noaa/global/mterry/global-workflow_forked/ush/python', '__LMOD_REF_COUNT_PYTHONPATH': '/apps/other/globus-cli-3.35.2:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:3;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:1', 'py_xarray_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4', 'nst_anl': True, 'FHOUT_GFS': 3, 'WORK': '/work/hercules', 'W3EMC_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_4.a', 'F77': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'BASE_DATA': '/work2/noaa/global/role-global/data', 'py_xlwt_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56', 'ACLOCAL_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/aclocal:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/aclocal:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/aclocal', 'QUEUE_SERVICE': 'batch', 'W3EMC_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_8.a', 'DATA': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ocean_f072.365351', 'LESSOPEN': '||/usr/bin/lesspipe.sh %s', 'DO_VERFOZN': True, 'antlr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2', 'crtm_fix_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2', 'TOCGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib', 'USER': 'mterry', 'HPSS_PROJECT': 'emc-global', 'FHMAX_WAV_GFS': 120, 'FIXgdas': '/work2/noaa/global/mterry/global-workflow_forked/fix/gdas', 'launcher': 'srun -l --export=ALL --hint=nomultithread', 'NDATE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/ndate', 'LIBRARY_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib', 'gsl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z', 'SLURM_NODELIST': 'hercules-02-53', 'DO_FIT2OBS': True, 'IP_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_4', 'DOIAU_ENKF': True, 'COM_ATMOS_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos', 'py_xarray_ver': '2023.7.0', '__INTEL_POST_FFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64'], 'netcdf_diag': True, 'DOBNDPNT_WAVE': True, 'FETCHDIR': '/NCEPDEV/emc-global/1year/David.Grumm/test_data', 'ENVIRONMENT': 'BATCH', 'gsi_ncdiag_ver': '1.1.2', 'IP_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_8', 'py_bottleneck_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a', 'w3emc_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo', 'py_netcdf4_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i', 'wgrib2_ver': '2.0.8', 'TOCGRIB2SUPER': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib2super', 'COM_ICE_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2', 'KEEPDATA': False, 'udunits_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo', 'cdo_ver': '2.2.0', 'COM_OCEAN_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input', 'FNTSFA': ' ', 'LOADEDMODULES': 'contrib/0.1:intel-oneapi-compilers/2023.1.0:stack-intel/2021.9.0:intel-oneapi-mpi/2021.9.0:stack-intel-oneapi-mpi/2021.9.0:intel-oneapi-mkl/2023.1.0:zlib/1.2.13:pigz/2.7:zstd/1.5.2:tar/1.34:gettext/0.21.1:libxcrypt/4.4.35:sqlite/3.43.2:util-linux-uuid/2.38.1:python/3.11.6:libjpeg/2.1.0:jasper/2.0.32:libpng/1.6.37:openjpeg/2.3.1:eccodes/2.32.0:fftw/3.3.10:nghttp2/1.57.0:curl/8.4.0:proj/9.2.1:udunits/2.2.28:cdo/2.2.0:hdf5/1.14.0:snappy/1.1.10:c-blosc/1.21.5:netcdf-c/4.9.2:netcdf-fortran/4.6.1:parallel-netcdf/1.12.2:parallelio/2.5.10:esmf/8.6.0:antlr/2.7.7:gsl/2.7.1:nco/5.0.6:bacio/2.4.1:w3emc/2.10.0:prod_util/2.1.1:g2/3.4.5:sp/2.5.0:ip/4.3.0:grib-util/1.3.0:g2tmpl/1.13.0:gsi-ncdiag/1.1.2:crtm-fix/2.4.0.1_emc:git-lfs/3.1.2:crtm/2.4.0.1:openblas/0.3.24:py-setuptools/63.4.3:py-numpy/1.23.4:bufr/11.7.0:gmake/4.2.1:wgrib2/2.0.8:py-f90nml/1.4.3:py-cftime/1.0.3.4:py-netcdf4/1.5.8:libyaml/0.2.5:py-pyyaml/6.0:py-markupsafe/2.1.3:py-jinja2/3.1.2:py-bottleneck/1.3.7:py-numexpr/2.8.4:py-et-xmlfile/1.0.1:py-openpyxl/3.1.2:py-pytz/2023.3:py-pyxlsb/1.0.10:py-xlrd/2.0.1:py-xlsxwriter/3.1.7:py-xlwt/1.3.0:py-pandas/1.5.3:py-six/1.16.0:py-python-dateutil/2.8.2:g2c/1.8.0:netcdf-cxx4/4.3.1:met/9.1.3:metplus/3.1.1:py-packaging/23.1:py-xarray/2023.7.0:prepobs/1.1.0:fit2obs/1.1.7.1:globus-cli/3.35.2:module_base.hercules', 'SLURM_JOB_ACCOUNT': 'fv3-cpu', 'DO_OCN': True, 'SLURM_PRIO_PROCESS': 0, 'HOMEfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1', 'gmake_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq', 'FIXfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/fix', 'py_python_dateutil_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy', 'tasks_per_node': 1, 'SLURM_NPROCS': 1, 'COM_CHEM_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history', 'LMOD_ROOT': '/apps/other/lmod', 'DOHYBVAR': '{{ DOHYBVAR }}', 'GSL_ROOT_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z', 'SERIAL_F77': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'IP_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_d', 'SHLVL': 4, '_ModuleTable071_': 'e30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAic3RhY2staW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMCIsCndWID0gIjAwMDAwMjAyMS4wMDAwMDAwMDkuKnpmaW5hbCIsCn0sCnRhciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC90YXIvMS4zNC5sdWEiLApmdWxsTmFtZSA9ICJ0YXIvMS4zNCIsCmxvYWRPcmRlciA9IDEwLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIs', '_ModuleTable012_': 'c3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImZmdHcvMy4zLjEwIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAwMy4wMDAwMDAwMTAuKnpmaW5hbCIsCn0sCmZpdDJvYnMgPSB7CmZuID0gIi93b3JrMi9ub2FhL2dsb2JhbC9yb2xlLWdsb2JhbC9naXQvRml0Mk9icy92MS4xLjcuMS9tb2R1bGVmaWxlcy9maXQyb2JzLzEuMS43LjEubHVhIiwKZnVsbE5hbWUgPSAiZml0Mm9icy8xLjEuNy4xIiwKbG9hZE9yZGVyID0gODIsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZml0Mm9icy8xLjEuNy4xIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4wMDAwMDAwMDcuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApnMiA9', 'INCVARS_EFOLD': 5, 'SLURM_NNODES': 1, 'RESERVATION': '', 'FHMIN_GFS': 0, 'FHMAX_HF_WAV': 48, 'FIXugwd': '/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd', 'BASH_ENV': '/apps/other/lmod/lmod/init/bash', 'CDATE': datetime.datetime(2021, 3, 23, 12, 0), '_ModuleTable063_': 'aW5hbCIsCn0sClsicHkteGxyZCJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXhscmQvMi4wLjEubHVhIiwKZnVsbE5hbWUgPSAicHkteGxyZC8yLjAuMSIsCmxvYWRPcmRlciA9IDY5LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXhscmQvMi4wLjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJweS14bHN4d3JpdGVyIl0g', '_ModuleTable006_': 'Mi4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMi4qemZpbmFsIiwKfSwKY29udHJpYiA9IHsKZm4gPSAiL2FwcHMvb3RoZXIvbW9kdWxlZmlsZXMvY29udHJpYi8wLjEiLApmdWxsTmFtZSA9ICJjb250cmliLzAuMSIsCmxvYWRPcmRlciA9IDEsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAwLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY29udHJpYiIsCndWID0gIjAwMDAwMDAwMC4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sCmNydG0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50', 'FHOUT_OCN_GFS': 6, '_ModuleTable026_': 'cmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2phc3Blci8yLjAuMzIubHVhIiwKZnVsbE5hbWUgPSAiamFzcGVyLzIuMC4zMiIsCmxvYWRPcmRlciA9IDE3LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImphc3Blci8yLjAuMzIiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAzMi4qemZpbmFsIiwKfSwKbGlianBlZyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1z', 'DONST': True, '_ModuleTable051_': 'ay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL3B5LW5ldGNkZjQvMS41LjgubHVhIiwKZnVsbE5hbWUgPSAicHktbmV0Y2RmNC8xLjUuOCIsCmxvYWRPcmRlciA9IDU4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LW5ldGNkZjQvMS41LjgiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA1LjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKWyJweS1udW1leHByIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3Rh', 'LMOD_sys': 'Linux', '_ModuleTable033_': 'L3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL25jby81LjAuNi5sdWEiLApmdWxsTmFtZSA9ICJuY28vNS4wLjYiLApsb2FkT3JkZXIgPSAzNywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJuY28vNS4wLjYiLAp3ViA9ICIwMDAwMDAwMDUuMDAwMDAwMDAwLjAwMDAwMDAwNi4qemZpbmFsIiwKfSwKWyJuZXRjZGYtYyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1', 'FHMIN_WAV': 0, 'LMOD_FAMILY_METACOMPILER_VERSION': '2021.9.0', 'I_MPI_PMI_LIBRARY': '/opt/slurm/lib/libpmi2.so', 'machine': 'HERCULES', 'py_pyyaml_ver': 6.0, 'OCL_ICD_FILENAMES': 'libintelocl_emu.so:libalteracl.so:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64/libintelocl.so', '_ModuleTable017_': 'bG9hZE9yZGVyID0gNDgsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ2l0LWxmcy8zLjEuMiIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbImdsb2J1cy1jbGkiXSA9IHsKZm4gPSAiL2FwcHMvb3RoZXIvbW9kdWxlZmlsZXMvZ2xvYnVzLWNsaS8zLjM1LjIubHVhIiwKZnVsbE5hbWUgPSAiZ2xvYnVzLWNsaS8zLjM1LjIiLApsb2FkT3JkZXIgPSA4MywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJnbG9idXMtY2xpIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAzNS4wMDAwMDAwMDIu', 'DO_FETCH_HPSS': False, 'PROJ_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/proj', 'FHOUT_ICE': 3, 'COM_WAVE_STATION_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station', 'COMINecmwf': '/work2/noaa/global/role-global/data/external_gempak/ecmwf', 'COM_TOP_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}', 'WGRIB2_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include', 'mpmd_opt': '--multi-prog --output=mpmd.%j.%t.out', 'DO_PREP_SFC': False, 'LMOD_G2TMPLVIRT_NAME': 'g2tmpl', '_ModuleTable066_': 'bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5dGhvbi8zLjExLjYubHVhIiwKZnVsbE5hbWUgPSAicHl0aG9uLzMuMTEuNiIsCmxvYWRPcmRlciA9IDE1LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5dGhvbi8zLjExLjYiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDExLjAwMDAwMDAwNi4qemZpbmFsIiwKfSwKc25hcHB5ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAy', '_ModuleTable009_': 'MjEuOS4wL2N1cmwvOC40LjAubHVhIiwKZnVsbE5hbWUgPSAiY3VybC84LjQuMCIsCmxvYWRPcmRlciA9IDIzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImN1cmwvOC40LjAiLAp3ViA9ICIwMDAwMDAwMDguMDAwMDAwMDA0Lip6ZmluYWwiLAp9LAplY2NvZGVzID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2VjY29kZXMvMi4zMi4wLmx1YSIsCmZ1bGxOYW1lID0gImVjY29kZXMvMi4z', 'py_numexpr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq', '__LMOD_REF_COUNT_MANPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/share/man:2;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/man:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/documentation/en/man/common:1;/apps/other/lmod/lmod/share/man:1;/usr/share/man:1;/apps/share/man:1;/apps/man:1;/opt/slurm/share/man:1', 'DOLETKF_OCN': False, 'MPI_GROUP_MAX': 256, 'py_cftime_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i', 'XDG_SESSION_ID': 'c21', 'zstd_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4', 'CLUSTERS': '', '_ModuleTable034_': 'bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAvbmV0Y2RmLWMvNC45LjIubHVhIiwKZnVsbE5hbWUgPSAibmV0Y2RmLWMvNC45LjIiLApsb2FkT3JkZXIgPSAzMCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJuZXRjZGYtYy80LjkuMiIsCndWID0gIjAwMDAwMDAwNC4wMDAwMDAwMDkuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbIm5ldGNkZi1jeHg0Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2st', 'COM_OCEAN_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean', 'CLUSTERS_SERVICE': '', 'pgmout': 'OUTPUT.365754', 'EXECfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/exec', '_ModuleTable031_': 'aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJtZXQvOS4xLjMiLAp3ViA9ICIwMDAwMDAwMDkuMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKbWV0cGx1cyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL21ldHBsdXMvMy4xLjEubHVhIiwKZnVsbE5hbWUgPSAibWV0cGx1cy8zLjEuMSIsCmxvYWRPcmRlciA9IDc4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIs', 'COM_ATMOS_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon', 'HOMEobsproc': '/work2/noaa/global/role-global/git/obsproc/v1.2.0', 'esmf_ver': '8.6.0', 'DO_ARCHCOM': False, 'SLURM_SUBMIT_HOST': 'hercules-login-1.hpc.msstate.edu', 'WRITE_DOPOST': True, 'EXECcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/exec', 'DUMP_SUFFIX': '', '_ModuleTable072_': 'CnVzZXJOYW1lID0gInRhci8xLjM0IiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAzNC4qemZpbmFsIiwKfSwKdWR1bml0cyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC91ZHVuaXRzLzIuMi4yOC5sdWEiLApmdWxsTmFtZSA9ICJ1ZHVuaXRzLzIuMi4yOCIsCmxvYWRPcmRlciA9IDI1LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInVkdW5pdHMvMi4yLjI4IiwKd1YgPSAiMDAwMDAwMDAy', 'COMOUT_OCEAN_GRIB': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ocean/grib2', 'CASE': 'C48', 'SENDDBN': False, 'gettext_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43', 'COM_ICE_BMATRIX_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice', 'git_lfs_ROOT': '/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l', '_ModuleTable001_': 'X01vZHVsZVRhYmxlXyA9IHsKTVR2ZXJzaW9uID0gMywKY19yZWJ1aWxkVGltZSA9IGZhbHNlLApjX3Nob3J0VGltZSA9IGZhbHNlLApkZXB0aFQgPSB7fSwKZmFtaWx5ID0gewpNZXRhQ29tcGlsZXIgPSAic3RhY2staW50ZWwiLApNZXRhTVBJID0gInN0YWNrLWludGVsLW9uZWFwaS1tcGkiLApjb21waWxlciA9ICJpbnRlbC1vbmVhcGktY29tcGlsZXJzIiwKZzJ0bXBsdmlydCA9ICJnMnRtcGwiLAptcGkgPSAiaW50ZWwtb25lYXBpLW1waSIsCn0sCm1UID0gewphbnRsciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9t', 'COM_ICE_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart', '_ModuleTable025_': 'LAp9LAppcCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9pcC80LjMuMC5sdWEiLApmdWxsTmFtZSA9ICJpcC80LjMuMCIsCmxvYWRPcmRlciA9IDQzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImlwLzQuMy4wIiwKd1YgPSAiMDAwMDAwMDA0LjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKamFzcGVyID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hl', 'COM_OCEAN_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart', 'ARCHCOM_TO': 'globus_hpss', 'COM_ATMOS_WMO_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo', 'ip_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf', 'CLASSPATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/mpi.jar', 'INTELFPGAOCLSDKROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga', 'LD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/host/linux64/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/lib:/usr/lib64:/usr/lib:/opt/slurm/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib', 'DO_CA': True, 'g2_ver': '3.4.5', 'LMOD_FAMILY_COMPILER': 'intel-oneapi-compilers', 'XDG_RUNTIME_DIR': '/run/user/9583', 'BASE_ENV': '/work2/noaa/global/mterry/global-workflow_forked/env', 'SLURM_JOB_ID': 5951676, '_ModuleTable014_': 'dWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nMmMvMS44LjAubHVhIiwKZnVsbE5hbWUgPSAiZzJjLzEuOC4wIiwKbG9hZE9yZGVyID0gNzUsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZzJjLzEuOC4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKZzJ0bXBsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9p', 'OCEANICEPRODUCTS_CONFIG': '/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml', 'NTHREADS_OCNICEPOST': 1, 'DO_JEDIOCNVAR': False, 'FIXreg2grb2': '/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2', 'SLURM_NODEID': 0, 'ip_ver': '4.3.0', 'KMP_AFFINITY': 'scatter', 'FHOUT_ICE_GFS': 6, '_ModuleTable065_': 'L3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS14bHd0LzEuMy4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXhsd3QvMS4zLjAiLApsb2FkT3JkZXIgPSA3MSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS14bHd0LzEuMy4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKcHl0aG9uID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1', 'DO_ATM': True, '_ModuleTable004_': 'IDUzLApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImJ1ZnIvMTEuNy4wIiwKd1YgPSAiMDAwMDAwMDExLjAwMDAwMDAwNy4qemZpbmFsIiwKfSwKWyJjLWJsb3NjIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvYy1ibG9zYy8xLjIxLjUubHVhIiwKZnVsbE5hbWUgPSAiYy1ibG9zYy8xLjIxLjUiLApsb2FkT3JkZXIgPSAyOSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1', 'esmf_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep', 'LMOD_FAMILY_MPI_VERSION': '2021.9.0', '_ModuleTable020_': 'c3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9nc2ktbmNkaWFnLzEuMS4yLmx1YSIsCmZ1bGxOYW1lID0gImdzaS1uY2RpYWcvMS4xLjIiLApsb2FkT3JkZXIgPSA0NiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJnc2ktbmNkaWFnLzEuMS4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCmdzbCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYu', 'DO_JEDIATMVAR': False, 'DO_AERO_ANL': False, 'NCLEN': '/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen', 'PS4': '+ $(basename ${BASH_SOURCE[0]:-${FUNCNAME[0]:-"Unknown"}})[${LINENO}]', 'COM_SNOW_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow', '_ModuleTable053_': 'YWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbnVtcHkvMS4yMy40Lmx1YSIsCmZ1bGxOYW1lID0gInB5LW51bXB5LzEuMjMuNCIsCmxvYWRPcmRlciA9IDUyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSA4LApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LW51bXB5LzEuMjMuNCIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMjMuMDAwMDAwMDA0Lip6ZmluYWwiLAp9LApbInB5LW9wZW5weXhsIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0', '__LMOD_REF_COUNT_ACL_BOARD_VENDOR_PATH': '/opt/Intel/OpenCLFPGA/oneAPI/Boards:1', 'LMOD_FAMILY_METAMPI_VERSION': '2021.9.0', '_ModuleTable059_': 'IDY4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXB5eGxzYi8xLjAuMTAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAwLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKWyJweS1weXlhbWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1weXlhbWwvNi4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXB5eWFtbC82LjAiLApsb2FkT3JkZXIgPSA2MCwKcHJvcFQgPSB7fSwKc3RhY2tE', 'DO_CALC_INCREMENT_ENKF_GFS': False, 'MKLROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0', 'DEBUGINFOD_URLS': 'https://debuginfod.centos.org/ ', 'sven_root_path': '/home/gfekete/sven', '_ModuleTable047_': 'LApbInB5LWV0LXhtbGZpbGUiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1ldC14bWxmaWxlLzEuMC4xLmx1YSIsCmZ1bGxOYW1lID0gInB5LWV0LXhtbGZpbGUvMS4wLjEiLApsb2FkT3JkZXIgPSA2NSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1ldC14bWxmaWxlLzEuMC4xIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMC4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsi', 'restart_interval_enkfgfs': 3, 'SENDDBN_NTC': False, 'FIXaer': '/work2/noaa/global/mterry/global-workflow_forked/fix/aer', 'FHMAX_HF_GFS': 48, 'COM_ICE_LETKF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf', 'LMOD_FAMILY_MPI': 'intel-oneapi-mpi', 'netcdf_fortran_ver': '4.6.1', 'FIXmom': '/work2/noaa/global/mterry/global-workflow_forked/fix/mom6', 'cyc': 12, 'BACIO_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib/libbacio_4.a', 'FC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'FHOUT_WAV_GFS': 3, 'SDATE_GFS': datetime.datetime(2021, 3, 23, 12, 0), 'BACIO_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib/libbacio_8.a', 'sp_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr', 'OPS_RES': 'C768', 'NMEM_ENS_GFS': 30, 'which_declare': 'declare -f', '_ModuleTable056_': 'OS4wL3B5LXBhbmRhcy8xLjUuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1wYW5kYXMvMS41LjMiLApsb2FkT3JkZXIgPSA3MiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1wYW5kYXMvMS41LjMiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA1LjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKWyJweS1weXRob24tZGF0ZXV0aWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1weXRob24tZGF0ZXV0aWwvMi44', 'COM_WAVE_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded', 'LMOD_FAMILY_G2TMPLVIRT': 'g2tmpl', 'QUEUE_DTN': 'batch', 'DO_GENESIS_FSU': False, 'prod_util_ver': '2.1.1', 'MPI_F90': 'mpiifort', 'LMOD_FAMILY_COMPILER_VERSION': '2023.1.0', '_ModuleTable042_': 'MDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcGlnei8yLjcubHVhIiwKZnVsbE5hbWUgPSAicGlnei8yLjciLApsb2FkT3JkZXIgPSA4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gNCwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInBpZ3ovMi43IiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4qemZpbmFsIiwKfSwKcHJlcG9icyA9IHsKZm4gPSAiL3dvcmsyL25vYWEvZ2xvYmFsL3JvbGUtZ2xvYmFsL2dpdC9wcmVwb2JzL3YxLjEuMC9tb2R1bGVmaWxlcy9wcmVwb2JzLzEuMS4wLmx1YSIsCmZ1bGxOYW1lID0gInByZXBvYnMvMS4xLjAiLApsb2FkT3JkZXIgPSA4MSwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0', 'openjpeg_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu', 'USHfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/ush', 'YAML_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib', 'DO_GSISOILDA': False, 'nemsio_ver': '2.5.4', 'DO_NPOESS': False, 'SLURM_MPI_TYPE': 'pmi2', 'OMP_STACKSIZE': 2048000, 'py_et_xmlfile_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq', 'FORECAST_HOUR': 72, 'ATARDIR': '/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW', 'EXECgfs': '/work2/noaa/global/mterry/global-workflow_forked/exec', 'py_python_dateutil_ver': '2.8.2', 'assim_freq': 6, 'DOIAU': True, 'FHOUT_WAV': 1, 'py_pyyaml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2', 'met_ver': '9.1.3', 'gsi_ncdiag_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe', 'DO_FETCH_LOCAL': False, 'NLSPATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64/locale/%l_%t/%N:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin/locale/%l_%t/%N', 'netcdf_fortran_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe', 'SP_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_4.a', 'COM_OCEAN_BMATRIX_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean', 'SLURM_CONF': '/var/spool/slurmd/conf-cache/slurm.conf', 'PATH': '/apps/other/globus-cli-3.35.2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r/ush:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/bin:/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/bin/intel64:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/bin:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/bin:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin:/usr/sbin:/usr/bin:/apps/sbin:/apps/bin:/opt/slurm/bin:/home/gfekete/sven/bin', 'SP_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_8.a', 'netcdf_c_ver': '4.9.2', 'SLURM_JOB_NAME': 'C48_S2SW_gfs_ocean_prod_f072_12', 'COM_ICE_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history', 'NLN': '/bin/ln -sf', 'MODULEPATH': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles:/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0-a66eaip/g2tmpl/1.13.0/intel/2021.9.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/g2tmpl/1.13.0/intel/2021.9.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/intel-oneapi-mpi/2021.9.0-a66eaip/oneapi/2023.1.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core:/work2/noaa/global/mterry/global-workflow_forked/modulefiles:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core:/apps/other/modulefiles:/apps/containers/modulefiles:/apps/licensed/modulefiles:/apps/contrib/modulefiles', 'CC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icc', 'grib_util_ver': '1.3.0', 'py_f90nml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z', 'DO_MERGENSST': False, 'SLURM_NTASKS_PER_NODE': 1, '_LMFILES_': '/apps/other/modulefiles/contrib/0.1:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core/intel-oneapi-compilers/2023.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core/stack-intel/2021.9.0.lua:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0/intel-oneapi-mpi/2021.9.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/stack-intel-oneapi-mpi/2021.9.0.lua:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core/intel-oneapi-mkl/2023.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/zlib/1.2.13.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/pigz/2.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/zstd/1.5.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/tar/1.34.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gettext/0.21.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libxcrypt/4.4.35.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/sqlite/3.43.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/util-linux-uuid/2.38.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/python/3.11.6.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libjpeg/2.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/jasper/2.0.32.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libpng/1.6.37.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/openjpeg/2.3.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/eccodes/2.32.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/fftw/3.3.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/nghttp2/1.57.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/curl/8.4.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/proj/9.2.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/udunits/2.2.28.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/cdo/2.2.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/hdf5/1.14.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/snappy/1.1.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/c-blosc/1.21.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-c/4.9.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-fortran/4.6.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/parallel-netcdf/1.12.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/parallelio/2.5.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/esmf/8.6.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/antlr/2.7.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gsl/2.7.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/nco/5.0.6.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/bacio/2.4.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/w3emc/2.10.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/prod_util/2.1.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2/3.4.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/sp/2.5.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/ip/4.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/grib-util/1.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2tmpl/1.13.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/gsi-ncdiag/1.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/crtm-fix/2.4.0.1_emc.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/git-lfs/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/crtm/2.4.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/openblas/0.3.24.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-setuptools/63.4.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-numpy/1.23.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/bufr/11.7.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gmake/4.2.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/wgrib2/2.0.8.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-f90nml/1.4.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-cftime/1.0.3.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/py-netcdf4/1.5.8.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libyaml/0.2.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pyyaml/6.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-markupsafe/2.1.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-jinja2/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-bottleneck/1.3.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-numexpr/2.8.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-et-xmlfile/1.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-openpyxl/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pytz/2023.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pyxlsb/1.0.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlrd/2.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlsxwriter/3.1.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlwt/1.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pandas/1.5.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-six/1.16.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-python-dateutil/2.8.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2c/1.8.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-cxx4/4.3.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/met/9.1.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/metplus/3.1.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-packaging/23.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xarray/2023.7.0.lua:/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles/prepobs/1.1.0.lua:/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles/fit2obs/1.1.7.1.lua:/apps/other/modulefiles/globus-cli/3.35.2.lua:/work2/noaa/global/mterry/global-workflow_forked/modulefiles/module_base.hercules.lua', 'globus_cli_ver': 3.27, '_ModuleTable028_': 'L2xpYnBuZy8xLjYuMzcubHVhIiwKZnVsbE5hbWUgPSAibGlicG5nLzEuNi4zNyIsCmxvYWRPcmRlciA9IDE4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImxpYnBuZy8xLjYuMzciLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA2LjAwMDAwMDAzNy4qemZpbmFsIiwKfSwKbGlieGNyeXB0ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2xpYnhjcnlwdC80LjQuMzUubHVhIiwKZnVsbE5hbWUgPSAibGlieGNyeXB0', 'py_numpy_ver': '1.23.4', '__LMOD_REF_COUNT_HDF5_PLUGIN_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/plugins:1', 'bufr_ver': '11.7.0', 'COMINnam': '/work2/noaa/global/role-global/data/external_gempak/nam', 'DBUS_SESSION_BUS_ADDRESS': 'unix:path=/run/user/9583/bus', 'DO_AERO_FCST': False, 'py_pandas_ver': '1.5.3', '_ModuleTable039_': 'dGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9vcGVuanBlZy8yLjMuMS5sdWEiLApmdWxsTmFtZSA9ICJvcGVuanBlZy8yLjMuMSIsCmxvYWRPcmRlciA9IDE5LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm9wZW5qcGVnLzIuMy4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsicGFyYWxsZWwtbmV0Y2RmIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMv', 'COM_ICE_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice', 'UUID_ORION_DTN': '8a10dd4f-24ee-4794-a39d-9c313ab6a34b', 'MPI_F77': 'mpiifort', 'LMOD_CMD': '/apps/other/lmod/lmod/libexec/lmod', '_ModuleTable045_': 'aW5hbCIsCn0sClsicHktYm90dGxlbmVjayJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LWJvdHRsZW5lY2svMS4zLjcubHVhIiwKZnVsbE5hbWUgPSAicHktYm90dGxlbmVjay8xLjMuNyIsCmxvYWRPcmRlciA9IDYzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWJvdHRsZW5lY2svMS4zLjciLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAzLjAwMDAwMDAwNy4qemZpbmFs', 'LMOD_MPI_VERSION': '2021.9.0-a66eaip', 'LMOD_SYSTEM_NAME': 'hercules', 'RUN_ENVIR': 'emc', 'w3emc_ver': '2.10.0', '__LMOD_REF_COUNT_LD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/host/linux64/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/lib:1;/usr/lib64:1;/usr/lib:1;/opt/slurm/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:2', 'DOHYBVAR_OCN': False, 'NUM_SND_COLLECTIVES': 9, 'walltime': '00:15:00', 'netcdf_cxx4_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu', 'ens_tracker_ver': 'v1.2.0', 'LMOD_SYSTEM_DEFAULT_MODULES': 'contrib', 'OMPI_MCA_plm_slurm_args': '--external-launcher', '_ModuleTable078_': 'cyIKLCAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLWE2NmVhaXAvZzJ0bXBsLzEuMTMuMC9pbnRlbC8yMDIxLjkuMCIKLCAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9nMnRtcGwvMS4xMy4wL2ludGVsLzIwMjEuOS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVz', 'FIXcpl': '/work2/noaa/global/mterry/global-workflow_forked/fix/cpl', 'G2C_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64/libg2c.so', '_ModuleTable019_': 'cmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9ncmliLXV0aWwvMS4zLjAubHVhIiwKZnVsbE5hbWUgPSAiZ3JpYi11dGlsLzEuMy4wIiwKbG9hZE9yZGVyID0gNDQsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ3JpYi11dGlsLzEuMy4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKWyJnc2ktbmNkaWFnIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMv', 'SLURM_JOB_GID': 17000, 'SLURM_GET_USER_ENV': 1, 'IAUFHRS_ENKF': [3, 6, 9], 'CPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/include:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/include:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/include', 'G2TMPL_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include', 'DO_JEDISNOWDA': False, 'SP_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_d.a', 'pigz_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq', 'COM_ATMOS_BUFR_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr', 'COM_OCEAN_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2', 'CNVGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/cnvgrib', 'COM_ATMOS_TRACK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks', 'DO_VERFRAD': True, 'pgmerr': 'errfile', 'MPI_BUFS_PER_HOST': 2048, 'DO_PREP_OBS_AERO': False, 'LMOD_FAMILY_G2TMPLVIRT_VERSION': '1.13.0', 'PNG_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu', 'COM_ATMOS_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart', 'DEGRIB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/degrib2', 'DO_ANLSTAT': False, 'LMOD_MPI_NAME': 'intel-oneapi-mpi', 'py_f90nml_ver': '1.4.3', 'NMEM_ENS_GFS_OFFSET': 20, 'SMOOTH_ENKF': False, 'FHMAX_GOES': 120, 'OLDPWD': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', 'QUEUE': 'batch', 'COM_OBS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/obs', 'DO_VMINMON': True, 'metplus_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r', 'VERBOSE': True, 'SLURM_JOB_NODELIST': 'hercules-02-53', 'CLUSTER': 'hercules', 'DBNROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn', 'FHOUT_HF_WAV': 1, 'PARMgfs': '/work2/noaa/global/mterry/global-workflow_forked/parm', 'I_MPI_HYDRA_BOOTSTRAP': 'slurm', 'APRUN_OCNICEPOST': 'srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1', 'BASH_FUNC_ml%%': '() { eval "$($LMOD_DIR/ml_cmd "$@")"\n}', 'BASH_FUNC_which%%': '() { ( alias;\n eval ${which_declare} ) | /usr/bin/which --tty-only --read-alias --read-functions --show-tilde --show-dot $@\n}', 'BASH_FUNC_module%%': '() { if [ -z "${LMOD_SH_DBG_ON+x}" ]; then\n case "$-" in \n *v*x*)\n __lmod_sh_dbg=\'vx\'\n ;;\n *v*)\n __lmod_sh_dbg=\'v\'\n ;;\n *x*)\n __lmod_sh_dbg=\'x\'\n ;;\n esac;\n fi;\n if [ -n "${__lmod_sh_dbg:-}" ]; then\n set +$__lmod_sh_dbg;\n echo "Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for Lmod\'s output" 1>&2;\n fi;\n eval "$($LMOD_CMD $LMOD_SHELL_PRGM "$@")" && eval "$(${LMOD_SETTARG_CMD:-:} -s sh)";\n __lmod_my_status=$?;\n if [ -n "${__lmod_sh_dbg:-}" ]; then\n echo "Shell debugging restarted" 1>&2;\n set -$__lmod_sh_dbg;\n fi;\n unset __lmod_sh_dbg;\n return $__lmod_my_status\n}', 'BASH_FUNC_declare_from_tmpl%%': '() { if [[ ${DEBUG_WORKFLOW:-"NO"} == "NO" ]]; then\n set +x;\n fi;\n local opts="-g";\n local OPTIND=1;\n while getopts "rx" option; do\n opts="${opts}${option}";\n done;\n shift $((OPTIND-1));\n for input in "$@";\n do\n IFS=\':\' read -ra args <<< "${input}";\n local com_var="${args[0]}";\n local template;\n local value;\n if (( ${#args[@]} > 1 )); then\n template="${args[1]}";\n else\n template="${com_var}_TMPL";\n fi;\n if [[ ! -v "${template}" ]]; then\n echo "FATAL ERROR in declare_from_tmpl: Requested template ${template} not defined!";\n exit 2;\n fi;\n value=$(echo "${!template}" | envsubst);\n declare ${opts} "${com_var}"="${value}";\n echo "declare_from_tmpl :: ${com_var}=${value}";\n done;\n set_trace\n}', 'BASH_FUNC_err_exit%%': ['() { set +eux;\n msg1=${*:-Job ${jobid} failed};\n if [[ -n "${pgm}" ]]; then\n msg1+="', 'ERROR IN ${pgm}";\n fi;\n if [[ -n "${err}" ]]; then\n msg1+=" RETURN CODE ${err}";\n fi;\n msg2="\n -------------------------------------------------------------\n -- FATAL ERROR: ${msg1}\n -- ABNORMAL EXIT at $(date) on ${HOSTNAME}\n -------------------------------------------------------------\n ";\n echo "${msg2}" 1>&2;\n module list;\n echo "" 1>&2;\n echo "${msg1}" 1>&2;\n if [[ -n "${DATA}" ]]; then\n echo "${DATA}" 1>&2;\n ls -ltr "${DATA}" 1>&2;\n else\n echo "WARNING: DATA variable not defined" 1>&2;\n fi;\n if [[ -n "${pgmout}" ]]; then\n if [[ -s errfile ]]; then\n echo "----- contents of errfile -----" >> "${pgmout}";\n cat errfile >> "${pgmout}";\n fi;\n cat "${pgmout}" 1>&2;\n else\n if [[ -s errfile ]]; then\n cat errfile 1>&2;\n fi;\n fi;\n if [[ "${SENDECF}" == "YES" ]]; then\n timeout 30 ecflow_client --msg "${ECF_NAME}: ${msg1}";\n timeout 30 ssh "${ECF_HOST}" "echo \\"${msg}2\\" >> ${ECF_JOBOUT:?}";\n fi;\n if [[ "${SENDECF}" == "YES" ]]; then\n ecflow_client --kill="${ECF_NAME:?}";\n fi;\n if [[ -n "${PBS_JOBID}" ]]; then\n qdel "${PBS_JOBID}";\n else\n if [[ -n "${SLURM_JOB_ID}" ]]; then\n scancel "${SLURM_JOB_ID}";\n fi;\n fi\n}'], 'BASH_FUNC_wait_for_file%%': '() { set +x;\n local file_name=${1:?"wait_for_file() requires a file name"};\n local sleep_interval=${2:-60};\n local max_tries=${3:-100};\n for ((iter=0; iter' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.oceanice_products -+++ config.oceanice_products[5]echo 'BEGIN: config.oceanice_products' -BEGIN: config.oceanice_products -+++ config.oceanice_products[8]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources oceanice_products -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=oceanice_products -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[978]walltime=00:15:00 -++++ config.resources[979]ntasks=1 -++++ config.resources[980]tasks_per_node=1 -++++ config.resources[981]threads_per_task=1 -++++ config.resources[982]memory=96GB -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n '' ]] -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export memory -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.oceanice_products[11]export MAX_TASKS=25 -+++ config.oceanice_products[11]MAX_TASKS=25 -+++ config.oceanice_products[13]export OCEANICEPRODUCTS_CONFIG=/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml -+++ config.oceanice_products[13]OCEANICEPRODUCTS_CONFIG=/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml -+++ config.oceanice_products[16]export NFHRS_PER_GROUP=3 -+++ config.oceanice_products[16]NFHRS_PER_GROUP=3 -+++ config.oceanice_products[18]echo 'END: config.oceanice_products' -END: config.oceanice_products -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env oceanice_products -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=oceanice_products -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 100663296 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 1 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 1 ]] -+++ HERCULES.env[34]max_threads_per_task=80 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 80 ]] -+++ HERCULES.env[40][[ 1 -gt 80 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 1' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[286]export NTHREADS_OCNICEPOST=1 -+++ HERCULES.env[286]NTHREADS_OCNICEPOST=1 -+++ HERCULES.env[287]export 'APRUN_OCNICEPOST=srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1' -+++ HERCULES.env[287]APRUN_OCNICEPOST='srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1' -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_OCEANICE_PRODUCTS[11]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[11]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[11]declare_from_tmpl -rx COMIN_OCEAN_HISTORY:COM_OCEAN_HISTORY_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_OCEAN_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/ocean/history -+ JGLOBAL_OCEANICE_PRODUCTS[12]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[12]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[12]declare_from_tmpl -rx COMOUT_OCEAN_GRIB:COM_OCEAN_GRIB_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_OCEAN_GRIB=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ocean/grib2 -+ JGLOBAL_OCEANICE_PRODUCTS[13]YMD=20210323 -+ JGLOBAL_OCEANICE_PRODUCTS[13]HH=12 -+ JGLOBAL_OCEANICE_PRODUCTS[13]declare_from_tmpl -rx COMOUT_OCEAN_NETCDF:COM_OCEAN_NETCDF_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_OCEAN_NETCDF=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ocean/netcdf -+ JGLOBAL_OCEANICE_PRODUCTS[17]/work2/noaa/global/mterry/global-workflow_forked/scripts/exglobal_oceanice_products.py -2025-07-28 21:34:36,923 - INFO - root : BEGIN: __main__.main -2025-07-28 21:34:36,924 - DEBUG - root : ( ) -2025-07-28 21:34:36,932 - INFO - oceanice_products: BEGIN: OceanIceProducts.__init__ -2025-07-28 21:34:36,932 - DEBUG - oceanice_products: ( , {'SHELL': '/bin/bash', 'PDY': datetime.datetime(2021, 3, 23, 0, 0), 'DO_JEDIATMENS': False, 'sqlite_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4', 'COPYGB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/copygb', 'WORK_hercules': '/work/hercules', 'TOCGRIB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib2', 'G2_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/include_d', 'PIO_TYPENAME_VALID_VALUES': ['netcdf', 'netcdf4p', 'netcdf4c', 'pnetcdf'], 'WGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/wgrib', 'SERIAL_CXX': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icpc', 'FHMAX_WAV': 9, 'FIXlut': '/work2/noaa/global/mterry/global-workflow_forked/fix/lut', 'fftw_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj', 'snappy_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx', '_ModuleTable016_': 'bGxOYW1lID0gImdldHRleHQvMC4yMS4xIiwKbG9hZE9yZGVyID0gMTEsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ2V0dGV4dC8wLjIxLjEiLAp3ViA9ICIwMDAwMDAwMDAuMDAwMDAwMDIxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJnaXQtbGZzIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvZ2l0LWxmcy8zLjEuMi5sdWEiLApmdWxsTmFtZSA9ICJnaXQtbGZzLzMuMS4yIiwK', 'MAKE_NSSTBUFR': False, 'cdo_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e', 'curl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops', 'PIO_VERSION_MAJOR': 2, 'threads_per_task': 1, 'SLURM_JOB_USER': 'mterry', '_ModuleTable067_': 'NC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3NuYXBweS8xLjEuMTAubHVhIiwKZnVsbE5hbWUgPSAic25hcHB5LzEuMS4xMCIsCmxvYWRPcmRlciA9IDI4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInNuYXBweS8xLjEuMTAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAxLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKc3AgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45', 'COM_ICE_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}', 'COM_OCEAN_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}', '_ModuleTable022_': 'ZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9oZGY1LzEuMTQuMC5sdWEiLApmdWxsTmFtZSA9ICJoZGY1LzEuMTQuMCIsCmxvYWRPcmRlciA9IDI3LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImhkZjUvMS4xNC4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAxNC4qemZpbmFsIiwKfSwKWyJpbnRlbC1vbmVhcGktY29tcGlsZXJzIl0gPSB7CmZuID0gIi9hcHBzL3NwYWNrLW1hbmFnZWQvbW9kdWxlZmlsZXMvbGludXgtcm9ja3k5LXg4Nl82NC9Db3JlL2ludGVsLW9uZWFwaS1jb21waWxlcnMvMjAyMy4xLjAubHVhIiwKZnVsbE5hbWUgPSAiaW50ZWwtb25lYXBpLWNvbXBpbGVy', 'SLURM_TASKS_PER_NODE': 1, 'COM_ATMOS_OZNMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon', 'COM_ATMOS_GEMPAK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}', 'SLURM_JOB_UID': 9583, 'HISTCONTROL': 'ignoredups', 'SLURM_EXPORT_ENV': 'NONE', 'g2c_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp', 'COM_OCEAN_NETCDF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf', 'GRB2INDEX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grb2index', 'l4densvar': True, 'FHOUT_HF_GFS': 1, 'I_MPI_EXTRA_FILESYSTEM': 1, 'SLURM_TASK_PID': 1366570, 'EXECprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/exec', 'COMOUT_OCEAN_NETCDF': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ocean/netcdf', 'g2tmpl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel', 'MPI_BUFS_PER_PROC': 2048, 'IAU_DELTHRS_ENKF': 6, 'COM_WAVE_WMO_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo', '_ModuleTable027_': 'LTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9saWJqcGVnLzIuMS4wLmx1YSIsCmZ1bGxOYW1lID0gImxpYmpwZWcvMi4xLjAiLApsb2FkT3JkZXIgPSAxNiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJsaWJqcGVnLzIuMS4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKbGlicG5nID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4w', 'NMEM_ENS': 0, '__LMOD_REF_COUNT_CLASSPATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/mpi.jar:1', '_ModuleTable050_': 'ZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbWFya3Vwc2FmZS8yLjEuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1tYXJrdXBzYWZlLzIuMS4zIiwKbG9hZE9yZGVyID0gNjEsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktbWFya3Vwc2FmZS8yLjEuMyIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDEuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LW5ldGNkZjQiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFj', 'err': 0, '_ModuleTable032_': 'CnVzZXJOYW1lID0gIm1ldHBsdXMvMy4xLjEiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJtb2R1bGVfYmFzZS5oZXJjdWxlcyJdID0gewpmbiA9ICIvd29yazIvbm9hYS9nbG9iYWwvbXRlcnJ5L2dsb2JhbC13b3JrZmxvd19mb3JrZWQvbW9kdWxlZmlsZXMvbW9kdWxlX2Jhc2UuaGVyY3VsZXMubHVhIiwKZnVsbE5hbWUgPSAibW9kdWxlX2Jhc2UuaGVyY3VsZXMiLApsb2FkT3JkZXIgPSA4NCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDAsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJtb2R1bGVfYmFzZS5oZXJjdWxlcyIsCndWID0gIk0uKnpmaW5hbCIsCn0sCm5jbyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGlj', 'CHGRP_CMD': 'chgrp rstprod', 'PKG_CONFIG_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib/pkgconfig:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib/pkgconfig:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/pkgconfig:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib/pkgconfig:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/lib/pkgconfig', 'SLURM_LOCALID': 0, 'MAX_TASKS': 25, '_ModuleTable040_': 'aW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9wYXJhbGxlbC1uZXRjZGYvMS4xMi4yLmx1YSIsCmZ1bGxOYW1lID0gInBhcmFsbGVsLW5ldGNkZi8xLjEyLjIiLApsb2FkT3JkZXIgPSAzMiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwYXJhbGxlbC1uZXRjZGYvMS4xMi4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAxMi4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCnBhcmFsbGVsaW8gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0', 'SLURM_SUBMIT_DIR': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', '_ModuleTable077_': 'LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvenN0ZC8xLjUuMi5sdWEiLApmdWxsTmFtZSA9ICJ6c3RkLzEuNS4yIiwKbG9hZE9yZGVyID0gOSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMywKc3RhY2tEZXB0aCA9IDQsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJ6c3RkLzEuNS4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwNS4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCn0sCm1wYXRoQSA9IHsKIi93b3JrMi9ub2FhL2dsb2JhbC9yb2xlLWdsb2JhbC9naXQvRml0Mk9icy92MS4xLjcuMS9tb2R1bGVmaWxlcyIKLCAiL3dvcmsyL25vYWEvZ2xvYmFsL3JvbGUtZ2xvYmFsL2dpdC9wcmVwb2JzL3YxLjEuMC9tb2R1bGVmaWxl', 'HOSTNAME': 'hercules-01-10', 'HISTSIZE': 1000, 'wgrib2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln', 'FRAC_GRID': True, 'G2_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/include_4', 'intel_oneapi_mpi_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse', 'SLURMD_NODENAME': 'hercules-01-10', '_ModuleTable080_': 'bGVzL2xpbnV4LXJvY2t5OS14ODZfNjQvb25lYXBpLzIwMjMuMS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL0NvcmUiCiwgIi93b3JrMi9ub2FhL2dsb2JhbC9tdGVycnkvZ2xvYmFsLXdvcmtmbG93X2ZvcmtlZC9tb2R1bGVmaWxlcyIKLCAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L0NvcmUiCiwgIi9hcHBzL290aGVyL21vZHVsZWZpbGVzIiwgIi9hcHBzL2NvbnRhaW5lcnMvbW9kdWxlZmlsZXMiCiwgIi9hcHBzL2xpY2Vuc2VkL21vZHVsZWZpbGVzIiwg', '__LMOD_REF_COUNT_OCL_ICD_FILENAMES': 'libintelocl_emu.so:1;libalteracl.so:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64/libintelocl.so:1', 'c_blosc_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn', 'FHOUT_OCN': 3, 'COM_OBS_JEDI': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi', 'eccodes_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj', '_ModuleTable037_': 'ay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9uZ2h0dHAyLzEuNTcuMC5sdWEiLApmdWxsTmFtZSA9ICJuZ2h0dHAyLzEuNTcuMCIsCmxvYWRPcmRlciA9IDIyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gNCwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm5naHR0cDIvMS41Ny4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDA1Ny4qemZpbmFsIiwKfSwKb3BlbmJsYXMgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2kt', 'netcdf_c_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy', '_ModuleTable070_': 'c3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzdGFjay1pbnRlbC8yMDIxLjkuMCIsCndWID0gIjAwMDAwMjAyMS4wMDAwMDAwMDkuKnpmaW5hbCIsCn0sClsic3RhY2staW50ZWwtb25lYXBpLW1waSJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3N0YWNrLWludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAubHVhIiwKZnVsbE5hbWUgPSAic3RhY2staW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMCIsCmxvYWRPcmRlciA9IDUsCnByb3BUID0g', 'SLURM_JOB_START_TIME': 1753756458, '_ModuleTable011_': 'b3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZXNtZi84LjYuMCIsCndWID0gIjAwMDAwMDAwOC4wMDAwMDAwMDYuKnpmaW5hbCIsCn0sCmZmdHcgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9mZnR3LzMuMy4xMC5sdWEiLApmdWxsTmFtZSA9ICJmZnR3LzMuMy4xMCIsCmxvYWRPcmRlciA9IDIxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwK', 'DO_NEST': False, 'STMP': '/work2/noaa/stmp/mterry/HERCULES', 'HYDRA_LAUNCHER_EXTRA_ARGS': '--external-launcher', 'ACL_BOARD_VENDOR_PATH': '/opt/Intel/OpenCLFPGA/oneAPI/Boards', '_ModuleTable060_': 'ZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1weXlhbWwvNi4wIiwKd1YgPSAiMDAwMDAwMDA2Lip6ZmluYWwiLAp9LApbInB5LXNldHVwdG9vbHMiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1zZXR1cHRvb2xzLzYzLjQuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1zZXR1cHRvb2xzLzYzLjQuMyIsCmxvYWRPcmRlciA9IDUxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIs', '_ModuleTable007_': 'ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9jcnRtLzIuNC4wLjEubHVhIiwKZnVsbE5hbWUgPSAiY3J0bS8yLjQuMC4xIiwKbG9hZE9yZGVyID0gNDksCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY3J0bS8yLjQuMC4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNC4wMDAwMDAwMDAuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbImNydG0tZml4Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45', 'G2TMPL_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib/libg2tmpl.a', 'DO_WAVE': True, 'SCRgfs': '/work2/noaa/global/mterry/global-workflow_forked/scripts', 'SP_INC8': 'include_8', '_ModuleTable002_': 'b2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9hbnRsci8yLjcuNy5sdWEiLApmdWxsTmFtZSA9ICJhbnRsci8yLjcuNyIsCmxvYWRPcmRlciA9IDM1LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImFudGxyLzIuNy43IiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4wMDAwMDAwMDcuKnpmaW5hbCIsCn0sCmJhY2lvID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2JhY2lvLzIuNC4xLmx1', 'MPICC': 'mpiicc', 'HDF5_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh', '_ModuleTable055_': 'L21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXBhY2thZ2luZy8yMy4xLmx1YSIsCmZ1bGxOYW1lID0gInB5LXBhY2thZ2luZy8yMy4xIiwKbG9hZE9yZGVyID0gNzksCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcGFja2FnaW5nLzIzLjEiLAp3ViA9ICIwMDAwMDAwMjMuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbInB5LXBhbmRhcyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEu', 'SP_INC4': 'include_4', 'FPGA_VARS_DIR': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga', 'FHOUT': 3, '__LMOD_REF_COUNT_MODULEPATH': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles:1;/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0-a66eaip/g2tmpl/1.13.0/intel/2021.9.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/g2tmpl/1.13.0/intel/2021.9.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/intel-oneapi-mpi/2021.9.0-a66eaip/oneapi/2023.1.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core:1;/work2/noaa/global/mterry/global-workflow_forked/modulefiles:1;/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core:1;/apps/other/modulefiles:1;/apps/containers/modulefiles:1;/apps/licensed/modulefiles:1;/apps/contrib/modulefiles:1', 'cycle': 't12z', '_ModuleTable043_': 'aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcmVwb2JzLzEuMS4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKcHJvZF91dGlsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3Byb2RfdXRpbC8yLjEuMS5sdWEiLApmdWxsTmFtZSA9ICJwcm9kX3V0aWwvMi4xLjEiLApsb2FkT3JkZXIgPSA0MCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcm9kX3V0aWwvMi4x', 'MPI_CC': 'mpiicc', 'PARTITION_DTN': '', 'sp_ver': '2.5.0', 'SHELLOPTS': 'braceexpand:errexit:hashall:interactive-comments:nounset:xtrace', 'DO_AWIPS': False, 'EDATE': datetime.datetime(2021, 3, 23, 12, 0), '__LMOD_REF_COUNT___INTEL_POST_CFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1'], 'I_MPI_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0', 'CRTM_FIX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2/fix', '_ModuleTable062_': 'MTYuMCIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMTYuKnpmaW5hbCIsCn0sClsicHkteGFycmF5Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHkteGFycmF5LzIwMjMuNy4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXhhcnJheS8yMDIzLjcuMCIsCmxvYWRPcmRlciA9IDgwLApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXhhcnJheS8yMDIzLjcuMCIsCndWID0gIjAwMDAwMjAyMy4wMDAwMDAwMDcuKnpm', 'proj_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez', 'FSYNC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file', '_ModuleTable005_': 'cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJjLWJsb3NjLzEuMjEuNSIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMjEuMDAwMDAwMDA1Lip6ZmluYWwiLAp9LApjZG8gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9jZG8vMi4yLjAubHVhIiwKZnVsbE5hbWUgPSAiY2RvLzIuMi4wIiwKbG9hZE9yZGVyID0gMjYsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY2RvLzIu', 'g2tmpl_ver': '1.13.0', 'FHOUT_GOES': 3, '_ModuleTable021_': 'MC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nc2wvMi43LjEubHVhIiwKZnVsbE5hbWUgPSAiZ3NsLzIuNy4xIiwKbG9hZE9yZGVyID0gMzYsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ3NsLzIuNy4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sCmhkZjUgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxl', 'EUPD_CYC': 'gdas', 'fms_ver': '2023.02.01', 'util_linux_uuid_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj', '_ModuleTable052_': 'Y2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbnVtZXhwci8yLjguNC5sdWEiLApmdWxsTmFtZSA9ICJweS1udW1leHByLzIuOC40IiwKbG9hZE9yZGVyID0gNjQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktbnVtZXhwci8yLjguNCIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDguMDAwMDAwMDA0Lip6ZmluYWwiLAp9LApbInB5LW51bXB5Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2kt', '_ModuleTable030_': 'cFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJsaWJ5YW1sLzAuMi41IiwKd1YgPSAiMDAwMDAwMDAwLjAwMDAwMDAwMi4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCm1ldCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL21ldC85LjEuMy5sdWEiLApmdWxsTmFtZSA9ICJtZXQvOS4xLjMiLApsb2FkT3JkZXIgPSA3NywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0', 'COM_ATMOS_IMAGERY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery', 'I_MPI_EXTRA_FILESYSTEM_LIST': 'lustre', 'COM_WAVE_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history', 'SLURM_CLUSTER_NAME': 'hercules', '__LMOD_REF_COUNT_ACLOCAL_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/aclocal:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/aclocal:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/aclocal:2', 'SERIAL_CC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icc', 'COM_OBSPROC_TMPL': '${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos', 'SLURM_JOB_END_TIME': 1753757358, '_ModuleTable075_': 'ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3dncmliMi8yLjAuOC5sdWEiLApmdWxsTmFtZSA9ICJ3Z3JpYjIvMi4wLjgiLApsb2FkT3JkZXIgPSA1NSwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJ3Z3JpYjIvMi4wLjgiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKemxpYiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJj', 'FHMAX_GFS': 120, 'SENDAWIP': False, 'G2_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64/libg2_d.a', 'spack_mod_path': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core', 'OCNRES': 500, 'COM_MED_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart', 'SLURM_CPUS_ON_NODE': 1, 'LMOD_FAMILY_METAMPI': 'stack-intel-oneapi-mpi', 'LEVS': 128, 'FIXgsi': '/work2/noaa/global/mterry/global-workflow_forked/fix/gsi', 'NFHRS_PER_GROUP': 3, 'COM_WAVE_PREP_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep', 'FI_PROVIDER_PATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib/prov:/usr/lib64/libfabric', 'python_ver': '3.11.6', 'IAU_OFFSET': 0, 'py_packaging_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6', 'DIAGUTIL_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/sys_check/sys_check.sh', 'python_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2', 'SCRIPTScfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/scripts', 'py_markupsafe_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7', 'G2C_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include', 'py_xlrd_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f', 'LMOD_G2TMPLVIRT_VERSION': '1.13.0', 'pid': 1367274, 'OUTPUT_GRID': 'gaussian_grid', 'jobid': 'oceanice_products_ocean_f102.1366596', 'SLURM_JOB_CPUS_PER_NODE': 1, 'spack_env': 'gsi-addon-dev-fms-2024.01', 'INTEL_ONEAPI_MPI_ROOT': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse', 'LMOD_FAMILY_METACOMPILER': 'stack-intel', 'SP_INCd': 'include_d', 'py_six_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o', 'COM_ATMOS_GOES_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim', 'MPICXX': 'mpiicpc', 'restart_interval_gdas': 3, 'fit2obs_ver': '1.1.7.1', 'LMOD_DIR': '/apps/other/lmod/lmod/libexec', '_ModuleTable068_': 'LjAvc3AvMi41LjAubHVhIiwKZnVsbE5hbWUgPSAic3AvMi41LjAiLApsb2FkT3JkZXIgPSA0MiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMiwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzcC8yLjUuMCIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCnNxbGl0ZSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9zcWxpdGUvMy40My4yLmx1YSIsCmZ1bGxOYW1lID0gInNxbGl0ZS8zLjQzLjIiLApsb2FkT3Jk', 'REPLAY_ICS': False, 'COM_ICE_NETCDF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf', 'G2_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64/libg2_4.a', '_ModuleTable058_': 'InB5LXB5dHovMjAyMy4zIiwKbG9hZE9yZGVyID0gNjcsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcHl0ei8yMDIzLjMiLAp3ViA9ICIwMDAwMDIwMjMuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LXB5eGxzYiJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXB5eGxzYi8xLjAuMTAubHVhIiwKZnVsbE5hbWUgPSAicHktcHl4bHNiLzEuMC4xMCIsCmxvYWRPcmRlciA9', 'py_setuptools_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr', 'FHMIN': 0, 'COM_ICE_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input', '_ModuleTable048_': 'cHktZjkwbm1sIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktZjkwbm1sLzEuNC4zLmx1YSIsCmZ1bGxOYW1lID0gInB5LWY5MG5tbC8xLjQuMyIsCmxvYWRPcmRlciA9IDU2LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWY5MG5tbC8xLjQuMyIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMDQuMDAwMDAwMDAzLip6ZmluYWwiLAp9LApbInB5LWppbmphMiJdID0gewpmbiA9ICIvd29yay9ub2FhL2Vw', 'intel_mkl_ver': '2023.1.0', 'PRTE_MCA_plm_slurm_args': '--external-launcher', 'PWD': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ocean_f102.1366596', 'SLURM_GTIDS': 0, 'LOGNAME': 'mterry', 'MAKE_ACFTBUFR': False, 'W3EMC_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_8', 'HOMEDIR': '/work2/noaa/global/mterry', 'IAU_DELTHRS': 6, 'W3EMC_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_4', 'XDG_SESSION_TYPE': 'unspecified', 'ESMFMKFILE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib/esmf.mk', 'FHMAX': 9, 'WRITE_NSFLIP': True, 'SLURM_JOB_PARTITION': 'hercules', 'YAML_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x', 'PARTITION_SERVICE': 'service', 'MODULESHOME': '/apps/other/lmod/lmod', '__LMOD_REF_COUNT_DYLD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:1', 'COM_WAVE_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart', 'BUFR_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_4.so', 'BUFR_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_8.so', 'HOMEcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1', 'COM_OCEAN_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon', 'DEBUG_POSTSCRIPT': False, 'MANPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/share/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/man:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/share/man:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/man:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/documentation/en/man/common:/apps/other/lmod/lmod/share/man:/usr/share/man:/apps/share/man:/apps/man:/opt/slurm/share/man::', 'libyaml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x', 'jasper_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt', 'py_xlsxwriter_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok', 'SLURM_TRES_PER_TASK': 'cpu=1', 'nco_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq', 'stack_intel_ver': '2021.9.0', 'SLURM_OOM_KILL_STEP': 0, 'BUFR_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m', 'bufr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m', 'CLUSTERS_DTN': '', 'grib_util_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35', 'nghttp2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky', 'DO_TEST_MODE': True, 'PACKAGEROOT': '/work2/noaa/global/role-global/nwpara', 'lobsdiag_forenkf': True, 'SLURM_JOB_NUM_NODES': 1, 'py_netcdf4_ver': '1.5.8', 'ICERES': 500, 'CXX': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icpc', 'GRIB2GRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grib2grib2', 'DMPDIR': '/work/noaa/rstprod/dump', 'LSOIL_INCR': 2, 'metplus_ver': '3.1.1', '_ModuleTable008_': 'LjAvY3J0bS1maXgvMi40LjAuMV9lbWMubHVhIiwKZnVsbE5hbWUgPSAiY3J0bS1maXgvMi40LjAuMV9lbWMiLApsb2FkT3JkZXIgPSA0NywKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJjcnRtLWZpeC8yLjQuMC4xX2VtYyIsCndWID0gIjAwMDAwMDAwMi4wMDAwMDAwMDQuMDAwMDAwMDAwLjAwMDAwMDAwMS4qXy4qZW1jLip6ZmluYWwiLAp9LApjdXJsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIw', 'COM_OCEAN_LETKF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf', 'zlib_ver': '1.2.13', 'BUFR_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/libbufr_d.so', 'ENKF_SPREAD': True, 'zlib_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j', '__LMOD_REF_COUNT_CMAKE_PREFIX_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r:2;/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/IntelDPCPP:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72:1', '_ModuleTable035_': 'c3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9uZXRjZGYtY3h4NC80LjMuMS5sdWEiLApmdWxsTmFtZSA9ICJuZXRjZGYtY3h4NC80LjMuMSIsCmxvYWRPcmRlciA9IDc2LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm5ldGNkZi1jeHg0LzQuMy4xIiwKd1YgPSAiMDAwMDAwMDA0LjAwMDAwMDAwMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsibmV0Y2RmLWZvcnRyYW4iXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFj', 'FIXgfs': '/work2/noaa/global/mterry/global-workflow_forked/fix', 'DO_COUPLED': True, 'SLURM_JOBID': 5951722, 'SERIAL_FC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'COM_ATMOS_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input', 'W3EMC_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/include_d', '_ModuleTable013_': 'IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nMi8zLjQuNS5sdWEiLApmdWxsTmFtZSA9ICJnMi8zLjQuNSIsCmxvYWRPcmRlciA9IDQxLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImcyLzMuNC41IiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAwNC4wMDAwMDAwMDUuKnpmaW5hbCIsCn0sCmcyYyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJj', 'COM_ATMOS_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history', 'FIXam': '/work2/noaa/global/mterry/global-workflow_forked/fix/am', 'hdf5_ver': '1.14.0', 'restart_interval_enkfgdas': 3, 'COM_ATMOS_RADMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon', 'DO_CALC_INCREMENT': False, 'DO_STARTMEM_FROM_JEDIICE': False, 'FIXprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/fix', 'SLURM_JOB_QOS': 'batch', 'I_MPI_HYDRA_BOOTSTRAP_EXEC_EXTRA_ARGS': '--external-launcher', 'obsproc_run_ver': '1.2.0', 'COM_ATMOS_MINMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon', 'crtm_ver': '2.4.0.1', 'FIXorog': '/work2/noaa/global/mterry/global-workflow_forked/fix/orog', 'COM_OCEAN_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history', '__LMOD_REF_COUNT_PATH': '/apps/other/globus-cli-3.35.2/bin:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r/ush:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/bin:2;/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/bin:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/bin:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/bin/intel64:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/bin:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/bin:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/bin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/bin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin:1;/usr/sbin:1;/usr/bin:1;/apps/sbin:1;/apps/bin:1;/opt/slurm/bin:1;/home/gfekete/sven/bin:1', '_ModuleTable018_': 'KnpmaW5hbCIsCn0sCmdtYWtlID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2dtYWtlLzQuMi4xLmx1YSIsCmZ1bGxOYW1lID0gImdtYWtlLzQuMi4xIiwKbG9hZE9yZGVyID0gNTQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ21ha2UvNC4yLjEiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDAyLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJncmliLXV0aWwiXSA9IHsKZm4gPSAiL3dv', 'COMROOT': '/work2/noaa/global/mterry/RUNTESTS/COMROOT', 'HOME': '/home/mterry', 'FI_PROVIDER': 'mlx', 'met_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b', 'COM_RTOFS_TMPL': '${DMPDIR}', '_ModuleTable015_': 'bnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2cydG1wbC8xLjEzLjAubHVhIiwKZnVsbE5hbWUgPSAiZzJ0bXBsLzEuMTMuMCIsCmxvYWRPcmRlciA9IDQ1LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImcydG1wbC8xLjEzLjAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDEzLip6ZmluYWwiLAp9LApnZXR0ZXh0ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2dldHRleHQvMC4yMS4xLmx1YSIsCmZ1', '_ModuleTable_Sz_': 81, 'LANG': 'C.UTF-8', '__LMOD_REF_COUNT_LIBRARY_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:1', 'DO_BUFRSND': False, 'memory': '96GB', '_ModuleTable064_': 'PSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHkteGxzeHdyaXRlci8zLjEuNy5sdWEiLApmdWxsTmFtZSA9ICJweS14bHN4d3JpdGVyLzMuMS43IiwKbG9hZE9yZGVyID0gNzAsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHkteGxzeHdyaXRlci8zLjEuNyIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDA3Lip6ZmluYWwiLAp9LApbInB5LXhsd3QiXSA9IHsKZm4gPSAi', 'DYLD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib', 'PARTITION_BATCH': 'hercules', '_ModuleTable023_': 'cy8yMDIzLjEuMCIsCmxvYWRPcmRlciA9IDIsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiaW50ZWwtb25lYXBpLWNvbXBpbGVycy8yMDIzLjEuMCIsCndWID0gIjAwMDAwMjAyMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsiaW50ZWwtb25lYXBpLW1rbCJdID0gewpmbiA9ICIvYXBwcy9zcGFjay1tYW5hZ2VkL21vZHVsZWZpbGVzL2xpbnV4LXJvY2t5OS14ODZfNjQvQ29yZS9pbnRlbC1vbmVhcGktbWtsLzIwMjMuMS4wLmx1YSIsCmZ1bGxOYW1lID0gImludGVsLW9uZWFwaS1ta2wvMjAyMy4xLjAiLApsb2FkT3JkZXIgPSA2LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVz', '_ModuleTable081_': 'Ii9hcHBzL2NvbnRyaWIvbW9kdWxlZmlsZXMiLAp9LApzeXN0ZW1CYXNlTVBBVEggPSAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L0NvcmU6L2FwcHMvb3RoZXIvbW9kdWxlZmlsZXM6L2FwcHMvY29udGFpbmVycy9tb2R1bGVmaWxlczovYXBwcy9saWNlbnNlZC9tb2R1bGVmaWxlcyIsCn0K', 'REDERR': '2>', 'PSLOT': 'C48_S2SW', '__LMOD_REF_COUNT_PKG_CONFIG_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib/pkgconfig:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib/pkgconfig:2;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/pkgconfig:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib/pkgconfig:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/pkgconfig:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib/pkgconfig:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/lib/pkgconfig:1', '_ModuleTable036_': 'ay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL25ldGNkZi1mb3J0cmFuLzQuNi4xLmx1YSIsCmZ1bGxOYW1lID0gIm5ldGNkZi1mb3J0cmFuLzQuNi4xIiwKbG9hZE9yZGVyID0gMzEsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAibmV0Y2RmLWZvcnRyYW4vNC42LjEiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDA2LjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKbmdodHRwMiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFj', 'COM_ATMOS_GRIB_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}', 'USHcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/ush', 'DO_GEMPAK': False, 'SLURM_PROCID': 0, 'USHgfs': '/work2/noaa/global/mterry/global-workflow_forked/ush', 'DOIBP_WAV': False, 'prepobs_run_ver': '1.1.0', 'EXPDIR': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', 'job': 'oceanice_products', 'imp_physics': 8, 'LMOD_SHELL_PRGM': 'bash', 'IAU_FHROT': 0, 'MPIF90': 'mpiifort', 'IP_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_4.a', 'waveGRD': 'uglo_100km', 'COM_CHEM_BMAT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix', 'jasper_ver': '2.0.32', 'SCRATCH': '/scratch/hercules', 'IP_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_8.a', 'CLIENT_GLOBUS_UUID': '869912fe-f6de-46c0-af10-b22efd84a022', 'ACCOUNT': 'fv3-cpu', 'HOMEgfs': '/work2/noaa/global/mterry/global-workflow_forked', 'NHOUR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/nhour', 'binary_diag': False, 'COM_WAVE_GRID_RES_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}', 'SCRIPTSprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/scripts', 'MP_STDOUTMODE': 'ORDERED', 'sigio_ver': '2.3.2', '__LMOD_REF_COUNT_CPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/include:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/include:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/include:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/include:1', 'LMOD_SETTARG_FULL_SUPPORT': False, 'COMINsyn': '/work2/noaa/global/role-global/com/gfs/prod/syndat', 'OFFSET_START_HOUR': 0, 'REDOUT': '1>', 'PTMP': '/work2/noaa/stmp/mterry/HERCULES', '_ModuleTable076_': 'dWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC96bGliLzEuMi4xMy5sdWEiLApmdWxsTmFtZSA9ICJ6bGliLzEuMi4xMyIsCmxvYWRPcmRlciA9IDcsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEwLApzdGFja0RlcHRoID0gNSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInpsaWIvMS4yLjEzIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMi4wMDAwMDAwMTMuKnpmaW5hbCIsCn0sCnpzdGQgPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2', 'g2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato', 'MODE': 'forecast-only', 'DATAROOT': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312', 'bacio_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj', 'APP': 'S2SW', '_ModuleTable029_': 'LzQuNC4zNSIsCmxvYWRPcmRlciA9IDEyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImxpYnhjcnlwdC80LjQuMzUiLAp3ViA9ICIwMDAwMDAwMDQuMDAwMDAwMDA0LjAwMDAwMDAzNS4qemZpbmFsIiwKfSwKbGlieWFtbCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9saWJ5YW1sLzAuMi41Lmx1YSIsCmZ1bGxOYW1lID0gImxpYnlhbWwvMC4yLjUiLApsb2FkT3JkZXIgPSA1OSwKcHJv', 'openblas_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f', 'stack_impi_ver': '2021.9.0', 'COM_CHEM_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem', 'TMPDIR': '/local/scratch/mterry/5951722', 'HDF5_PLUGIN_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/plugins', 'DO_TRACKER': True, 'CMAKE_PREFIX_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r:/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/IntelDPCPP:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72', 'crtm_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r', '_ModuleTable038_': 'YWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvb3BlbmJsYXMvMC4zLjI0Lmx1YSIsCmZ1bGxOYW1lID0gIm9wZW5ibGFzLzAuMy4yNCIsCmxvYWRPcmRlciA9IDUwLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm9wZW5ibGFzLzAuMy4yNCIsCndWID0gIjAwMDAwMDAwMC4wMDAwMDAwMDMuMDAwMDAwMDI0Lip6ZmluYWwiLAp9LApvcGVuanBlZyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5z', 'nco_ver': '5.0.6', '_ModuleTable046_': 'IiwKfSwKWyJweS1jZnRpbWUiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1jZnRpbWUvMS4wLjMuNC5sdWEiLApmdWxsTmFtZSA9ICJweS1jZnRpbWUvMS4wLjMuNCIsCmxvYWRPcmRlciA9IDU3LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWNmdGltZS8xLjAuMy40IiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMC4wMDAwMDAwMDMuMDAwMDAwMDA0Lip6ZmluYWwiLAp9', 'GRBINDEX': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/grbindex', 'libjpeg_turbo_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy', 'MPIF77': 'mpiifort', '_ModuleTable073_': 'LjAwMDAwMDAwMi4wMDAwMDAwMjguKnpmaW5hbCIsCn0sClsidXRpbC1saW51eC11dWlkIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvdXRpbC1saW51eC11dWlkLzIuMzguMS5sdWEiLApmdWxsTmFtZSA9ICJ1dGlsLWxpbnV4LXV1aWQvMi4zOC4xIiwKbG9hZE9yZGVyID0gMTQsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAidXRpbC1saW51eC11dWlkLzIuMzguMSIsCndWID0gIjAw', '_ModuleTable010_': 'Mi4wIiwKbG9hZE9yZGVyID0gMjAsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZWNjb2Rlcy8yLjMyLjAiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDMyLip6ZmluYWwiLAp9LAplc21mID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAvZXNtZi84LjYuMC5sdWEiLApmdWxsTmFtZSA9ICJlc21mLzguNi4wIiwKbG9hZE9yZGVyID0gMzQsCnBy', 'NTHSTACK': 1024000000, 'SLURM_CPUS_PER_TASK': 1, 'FIXcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/fix', 'SLURM_NTASKS': 1, 'DO_ICE': True, 'NET': 'gfs', 'COM_ATMOS_GENESIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital', 'HOMEpost': '/work2/noaa/global/mterry/global-workflow_forked', '_ModuleTable061_': 'CnVzZXJOYW1lID0gInB5LXNldHVwdG9vbHMvNjMuNC4zIiwKd1YgPSAiMDAwMDAwMDYzLjAwMDAwMDAwNC4wMDAwMDAwMDMuKnpmaW5hbCIsCn0sClsicHktc2l4Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktc2l4LzEuMTYuMC5sdWEiLApmdWxsTmFtZSA9ICJweS1zaXgvMS4xNi4wIiwKbG9hZE9yZGVyID0gNzMsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktc2l4LzEu', 'bacio_ver': '2.4.1', 'cmake_ver': '3.23.1', 'UTILROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq', '_ModuleTable024_': 'ZXJOYW1lID0gImludGVsLW9uZWFwaS1ta2wvMjAyMy4xLjAiLAp3ViA9ICIwMDAwMDIwMjMuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApbImludGVsLW9uZWFwaS1tcGkiXSA9IHsKZm4gPSAiL2FwcHMvc3BhY2stbWFuYWdlZC9tb2R1bGVmaWxlcy9saW51eC1yb2NreTkteDg2XzY0L29uZWFwaS8yMDIzLjEuMC9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLmx1YSIsCmZ1bGxOYW1lID0gImludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAiLApsb2FkT3JkZXIgPSA0LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAiLAp3ViA9ICIwMDAwMDIwMjEuMDAwMDAwMDA5Lip6ZmluYWwi', 'tar_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths', 'ntasks': 1, 'max_tasks_per_node': 80, 'NCDUMP': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump', '_ModuleTable057_': 'LjIubHVhIiwKZnVsbE5hbWUgPSAicHktcHl0aG9uLWRhdGV1dGlsLzIuOC4yIiwKbG9hZE9yZGVyID0gNzQsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAicHktcHl0aG9uLWRhdGV1dGlsLzIuOC4yIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwOC4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sClsicHktcHl0eiJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXB5dHovMjAyMy4zLmx1YSIsCmZ1bGxOYW1lID0g', 'py_numpy_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne', 'QUILTING': True, 'IP_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64/libip_d.a', 'SLURM_TOPOLOGY_ADDR': 'hercules-01-10', 'libxcrypt_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr', 'CRTM_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include', 'COMIN_OCEAN_HISTORY': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/ocean/history', 'LMOD_VERSION': '8.7.14', 'spack_stack_ver': '1.6.0', '_ModuleTable041_': 'LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9wYXJhbGxlbGlvLzIuNS4xMC5sdWEiLApmdWxsTmFtZSA9ICJwYXJhbGxlbGlvLzIuNS4xMCIsCmxvYWRPcmRlciA9IDMzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInBhcmFsbGVsaW8vMi41LjEwIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNS4wMDAwMDAwMTAuKnpmaW5hbCIsCn0sCnBpZ3ogPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0y', 'HOMEprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0', 'COMPONENT': 'ocean', 'BASE_GIT': '/work2/noaa/global/role-global/git', 'FCST_SEGMENTS': [0, 120], 'libpng_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu', '_ModuleTable044_': 'LjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAxLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKcHJvaiA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9wcm9qLzkuMi4xLmx1YSIsCmZ1bGxOYW1lID0gInByb2ovOS4yLjEiLApsb2FkT3JkZXIgPSAyNCwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJwcm9qLzkuMi4xIiwKd1YgPSAiMDAwMDAwMDA5LjAwMDAwMDAwMi4wMDAwMDAwMDEuKnpm', 'DO_GENESIS': True, 'BUFR_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_4', '__LMOD_REF_COUNT_DIAGUTIL_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/sys_check/sys_check.sh:1', 'COM_WAVE_GEMPAK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak', 'COM_ATMOS_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2', 'RUN': 'gfs', 'BUFR_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_8', 'COM_ICE_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon', 'py_pandas_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw', '__INTEL_POST_CFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64'], 'ARCDIR': '/work2/noaa/global/mterry/archive/C48_S2SW', '__LMOD_REF_COUNT_NLSPATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64/locale/%l_%t/%N:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin/locale/%l_%t/%N:1', '_ModuleTable069_': 'ZXIgPSAxMywKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMiwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJzcWxpdGUvMy40My4yIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDA0My4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sClsic3RhY2staW50ZWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9Db3JlL3N0YWNrLWludGVsLzIwMjEuOS4wLmx1YSIsCmZ1bGxOYW1lID0gInN0YWNrLWludGVsLzIwMjEuOS4wIiwKbG9hZE9yZGVyID0gMywKcHJvcFQgPSB7fSwK', 'SENDECF': False, 'parallelio_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed', 'py_pytz_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s', 'ILPOST': 3, 'WGRIB2_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib/libwgrib2.a', 'parallel_netcdf_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3', 'CMPLR_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0', 'prod_util_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq', 'COMINukmet': '/work2/noaa/global/role-global/data/external_gempak/ukmet', 'libpng_ver': '1.6.37', 'COPYGB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/copygb2', 'NMV': '/bin/mv', 'NOSCRUB': '/work2/noaa/global/mterry', 'SCRATCH_hercules': '/scratch/hercules', 'W3EMC_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_d.a', 'SENDSDM': False, 'envir': 'prod', 'DO_METP': False, 'CRTM_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib/libcrtm.a', 'LOGSCRIPT': '', 'INTEL_ONEAPI_COMPILERS_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72', '_ModuleTable074_': 'MDAwMDAwMi4wMDAwMDAwMzguMDAwMDAwMDAxLip6ZmluYWwiLAp9LAp3M2VtYyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC93M2VtYy8yLjEwLjAubHVhIiwKZnVsbE5hbWUgPSAidzNlbWMvMi4xMC4wIiwKbG9hZE9yZGVyID0gMzksCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAidzNlbWMvMi4xMC4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKd2dyaWIy', 'py_jinja2_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj', 'FPGA_VARS_ARGS': '', 'INCVARS_ZERO_STRAT': ["'sphum_inc'", "'liq_wat_inc'", "'icmr_inc'", "'rwmr_inc'", "'snmr_inc'", "'grle_inc'"], 'COM_CONF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf', '_ModuleTable003_': 'YSIsCmZ1bGxOYW1lID0gImJhY2lvLzIuNC4xIiwKbG9hZE9yZGVyID0gMzgsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDIsCnN0YWNrRGVwdGggPSAzLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiYmFjaW8vMi40LjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDA0LjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKYnVmciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9idWZyLzExLjcuMC5sdWEiLApmdWxsTmFtZSA9ICJidWZyLzExLjcuMCIsCmxvYWRPcmRlciA9', 'HYDRA_BOOTSTRAP': 'slurm', '__LMOD_REF_COUNT___INTEL_POST_FFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1'], 'COM_CHEM_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon', 'MDATE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/mdate', 'WGRIB2': 'wgrib2', '_ModuleTable054_': 'LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktb3BlbnB5eGwvMy4xLjIubHVhIiwKZnVsbE5hbWUgPSAicHktb3BlbnB5eGwvMy4xLjIiLApsb2FkT3JkZXIgPSA2NiwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1vcGVucHl4bC8zLjEuMiIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbInB5LXBhY2thZ2luZyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxs', 'EXP_WARM_START': False, 'MODULEPATH_ROOT': '/apps/other/modulefiles', 'COM_SNOW_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon', 'CHGRP_RSTPROD': True, 'py_pyxlsb_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge', 'BACIO_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/include_8', 'restart_interval_gfs': 12, 'DO_GOES': False, 'pgm': '', 'SLURM_TOPOLOGY_ADDR_PATTERN': 'node', 'DO_LAND_IAU': False, 'ntiles': 6, 'FHCYC': 24, 'ROTDIR': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW', 'FHOUT_AERO': 3, 'py_jinja2_ver': '3.1.2', 'BACIO_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/include_4', 'INCREMENTS_TO_ZERO': ["'liq_wat_inc'", "'icmr_inc'", "'rwmr_inc'", "'snmr_inc'", "'grle_inc'"], 'FHR_LIST': 102, 'USHprepobs': '/work/noaa/global/glopara/git_rocky9/prepobs/v1.1.0/ush', 'BUFR_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include/bufr_d', 'hdf5_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh', 'INTEL_ONEAPI_MKL_ROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci', 'SCRIPTSfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/scripts', '__LMOD_REF_COUNT_FI_PROVIDER_PATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib/prov:1;/usr/lib64/libfabric:1', 'FIXcice': '/work2/noaa/global/mterry/global-workflow_forked/fix/cice', 'IAUFHRS': [6], 'XDG_SESSION_CLASS': 'background', 'INTERVAL_GFS': 6, 'py_openpyxl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe', 'FHMAX_FITS': 120, 'COM_ATMOS_MASTER_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master', 'LMOD_PKG': '/apps/other/lmod/lmod', 'MPI_MEMMAP_OFF': 1, 'UUID_HERCULES_DTN': '869912fe-f6de-46c0-af10-b22efd84a022', '_ModuleTable079_': 'L3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAiCiwgIi9hcHBzL3NwYWNrLW1hbmFnZWQvbW9kdWxlZmlsZXMvbGludXgtcm9ja3k5LXg4Nl82NC9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLWE2NmVhaXAvb25lYXBpLzIwMjMuMS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wIgosICIvYXBwcy9zcGFjay1tYW5hZ2VkL21vZHVsZWZp', 'SDATE': datetime.datetime(2021, 3, 23, 12, 0), 'SLURM_SCRIPT_CONTEXT': 'prolog_task', 'MPI_CXX': 'mpiicpc', 'lwrite4danl': True, 'SLURM_MEM_PER_NODE': 98304, 'CASE_ENS': '{{ CASE_ENS }}', '_ModuleTable049_': 'aWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LWppbmphMi8zLjEuMi5sdWEiLApmdWxsTmFtZSA9ICJweS1qaW5qYTIvMy4xLjIiLApsb2FkT3JkZXIgPSA2MiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1qaW5qYTIvMy4xLjIiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDAxLjAwMDAwMDAwMi4qemZpbmFsIiwKfSwKWyJweS1tYXJrdXBzYWZlIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3Vs', 'YAML_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include', 'PYTHONPATH': '/apps/other/globus-cli-3.35.2:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:/work2/noaa/global/mterry/global-workflow_forked/sorc/wxflow/src:/work2/noaa/global/mterry/global-workflow_forked/ush/python', '__LMOD_REF_COUNT_PYTHONPATH': '/apps/other/globus-cli-3.35.2:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64/python3.11/site-packages:3;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib/python3.11/site-packages:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib/python3.11/site-packages:1', 'py_xarray_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4', 'nst_anl': True, 'FHOUT_GFS': 3, 'WORK': '/work/hercules', 'W3EMC_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_4.a', 'F77': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'BASE_DATA': '/work2/noaa/global/role-global/data', 'py_xlwt_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56', 'ACLOCAL_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/aclocal:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/aclocal:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/aclocal', 'QUEUE_SERVICE': 'batch', 'W3EMC_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64/libw3emc_8.a', 'DATA': '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/oceanice_products_ocean_f102.1366596', 'LESSOPEN': '||/usr/bin/lesspipe.sh %s', 'DO_VERFOZN': True, 'antlr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2', 'crtm_fix_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-fix-2.4.0.1_emc-2os2hw2', 'TOCGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib', 'USER': 'mterry', 'HPSS_PROJECT': 'emc-global', 'FHMAX_WAV_GFS': 120, 'FIXgdas': '/work2/noaa/global/mterry/global-workflow_forked/fix/gdas', 'launcher': 'srun -l --export=ALL --hint=nomultithread', 'NDATE': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/ndate', 'LIBRARY_PATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib', 'gsl_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z', 'SLURM_NODELIST': 'hercules-01-10', 'DO_FIT2OBS': True, 'IP_INC4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_4', 'DOIAU_ENKF': True, 'COM_ATMOS_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos', 'py_xarray_ver': '2023.7.0', '__INTEL_POST_FFLAGS': ['-Wl', '-rpath', '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64'], 'netcdf_diag': True, 'DOBNDPNT_WAVE': True, 'FETCHDIR': '/NCEPDEV/emc-global/1year/David.Grumm/test_data', 'ENVIRONMENT': 'BATCH', 'gsi_ncdiag_ver': '1.1.2', 'IP_INC8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_8', 'py_bottleneck_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a', 'w3emc_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo', 'py_netcdf4_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i', 'wgrib2_ver': '2.0.8', 'TOCGRIB2SUPER': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/tocgrib2super', 'COM_ICE_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2', 'KEEPDATA': False, 'udunits_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo', 'cdo_ver': '2.2.0', 'COM_OCEAN_INPUT_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input', 'FNTSFA': ' ', 'LOADEDMODULES': 'contrib/0.1:intel-oneapi-compilers/2023.1.0:stack-intel/2021.9.0:intel-oneapi-mpi/2021.9.0:stack-intel-oneapi-mpi/2021.9.0:intel-oneapi-mkl/2023.1.0:zlib/1.2.13:pigz/2.7:zstd/1.5.2:tar/1.34:gettext/0.21.1:libxcrypt/4.4.35:sqlite/3.43.2:util-linux-uuid/2.38.1:python/3.11.6:libjpeg/2.1.0:jasper/2.0.32:libpng/1.6.37:openjpeg/2.3.1:eccodes/2.32.0:fftw/3.3.10:nghttp2/1.57.0:curl/8.4.0:proj/9.2.1:udunits/2.2.28:cdo/2.2.0:hdf5/1.14.0:snappy/1.1.10:c-blosc/1.21.5:netcdf-c/4.9.2:netcdf-fortran/4.6.1:parallel-netcdf/1.12.2:parallelio/2.5.10:esmf/8.6.0:antlr/2.7.7:gsl/2.7.1:nco/5.0.6:bacio/2.4.1:w3emc/2.10.0:prod_util/2.1.1:g2/3.4.5:sp/2.5.0:ip/4.3.0:grib-util/1.3.0:g2tmpl/1.13.0:gsi-ncdiag/1.1.2:crtm-fix/2.4.0.1_emc:git-lfs/3.1.2:crtm/2.4.0.1:openblas/0.3.24:py-setuptools/63.4.3:py-numpy/1.23.4:bufr/11.7.0:gmake/4.2.1:wgrib2/2.0.8:py-f90nml/1.4.3:py-cftime/1.0.3.4:py-netcdf4/1.5.8:libyaml/0.2.5:py-pyyaml/6.0:py-markupsafe/2.1.3:py-jinja2/3.1.2:py-bottleneck/1.3.7:py-numexpr/2.8.4:py-et-xmlfile/1.0.1:py-openpyxl/3.1.2:py-pytz/2023.3:py-pyxlsb/1.0.10:py-xlrd/2.0.1:py-xlsxwriter/3.1.7:py-xlwt/1.3.0:py-pandas/1.5.3:py-six/1.16.0:py-python-dateutil/2.8.2:g2c/1.8.0:netcdf-cxx4/4.3.1:met/9.1.3:metplus/3.1.1:py-packaging/23.1:py-xarray/2023.7.0:prepobs/1.1.0:fit2obs/1.1.7.1:globus-cli/3.35.2:module_base.hercules', 'SLURM_JOB_ACCOUNT': 'fv3-cpu', 'DO_OCN': True, 'SLURM_PRIO_PROCESS': 0, 'HOMEfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1', 'gmake_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq', 'FIXfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/fix', 'py_python_dateutil_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy', 'tasks_per_node': 1, 'SLURM_NPROCS': 1, 'COM_CHEM_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history', 'LMOD_ROOT': '/apps/other/lmod', 'DOHYBVAR': '{{ DOHYBVAR }}', 'GSL_ROOT_DIR': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z', 'SERIAL_F77': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'IP_INCd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/include_d', 'SHLVL': 4, '_ModuleTable071_': 'e30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAic3RhY2staW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMCIsCndWID0gIjAwMDAwMjAyMS4wMDAwMDAwMDkuKnpmaW5hbCIsCn0sCnRhciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC90YXIvMS4zNC5sdWEiLApmdWxsTmFtZSA9ICJ0YXIvMS4zNCIsCmxvYWRPcmRlciA9IDEwLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIs', '_ModuleTable012_': 'c3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImZmdHcvMy4zLjEwIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAwMy4wMDAwMDAwMTAuKnpmaW5hbCIsCn0sCmZpdDJvYnMgPSB7CmZuID0gIi93b3JrMi9ub2FhL2dsb2JhbC9yb2xlLWdsb2JhbC9naXQvRml0Mk9icy92MS4xLjcuMS9tb2R1bGVmaWxlcy9maXQyb2JzLzEuMS43LjEubHVhIiwKZnVsbE5hbWUgPSAiZml0Mm9icy8xLjEuNy4xIiwKbG9hZE9yZGVyID0gODIsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZml0Mm9icy8xLjEuNy4xIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4wMDAwMDAwMDcuMDAwMDAwMDAxLip6ZmluYWwiLAp9LApnMiA9', 'INCVARS_EFOLD': 5, 'SLURM_NNODES': 1, 'RESERVATION': '', 'FHMIN_GFS': 0, 'FHMAX_HF_WAV': 48, 'FIXugwd': '/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd', 'BASH_ENV': '/apps/other/lmod/lmod/init/bash', 'CDATE': datetime.datetime(2021, 3, 23, 12, 0), '_ModuleTable063_': 'aW5hbCIsCn0sClsicHkteGxyZCJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LXhscmQvMi4wLjEubHVhIiwKZnVsbE5hbWUgPSAicHkteGxyZC8yLjAuMSIsCmxvYWRPcmRlciA9IDY5LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXhscmQvMi4wLjEiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAwMS4qemZpbmFsIiwKfSwKWyJweS14bHN4d3JpdGVyIl0g', '_ModuleTable006_': 'Mi4wIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMi4qemZpbmFsIiwKfSwKY29udHJpYiA9IHsKZm4gPSAiL2FwcHMvb3RoZXIvbW9kdWxlZmlsZXMvY29udHJpYi8wLjEiLApmdWxsTmFtZSA9ICJjb250cmliLzAuMSIsCmxvYWRPcmRlciA9IDEsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAwLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiY29udHJpYiIsCndWID0gIjAwMDAwMDAwMC4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sCmNydG0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50', 'FHOUT_OCN_GFS': 6, '_ModuleTable026_': 'cmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2phc3Blci8yLjAuMzIubHVhIiwKZnVsbE5hbWUgPSAiamFzcGVyLzIuMC4zMiIsCmxvYWRPcmRlciA9IDE3LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImphc3Blci8yLjAuMzIiLAp3ViA9ICIwMDAwMDAwMDIuMDAwMDAwMDAwLjAwMDAwMDAzMi4qemZpbmFsIiwKfSwKbGlianBlZyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1z', 'DONST': True, '_ModuleTable051_': 'ay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL3B5LW5ldGNkZjQvMS41LjgubHVhIiwKZnVsbE5hbWUgPSAicHktbmV0Y2RmNC8xLjUuOCIsCmxvYWRPcmRlciA9IDU4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LW5ldGNkZjQvMS41LjgiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA1LjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKWyJweS1udW1leHByIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3Rh', 'LMOD_sys': 'Linux', '_ModuleTable033_': 'L3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL25jby81LjAuNi5sdWEiLApmdWxsTmFtZSA9ICJuY28vNS4wLjYiLApsb2FkT3JkZXIgPSAzNywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJuY28vNS4wLjYiLAp3ViA9ICIwMDAwMDAwMDUuMDAwMDAwMDAwLjAwMDAwMDAwNi4qemZpbmFsIiwKfSwKWyJuZXRjZGYtYyJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1', 'FHMIN_WAV': 0, 'LMOD_FAMILY_METACOMPILER_VERSION': '2021.9.0', 'I_MPI_PMI_LIBRARY': '/opt/slurm/lib/libpmi2.so', 'machine': 'HERCULES', 'py_pyyaml_ver': 6.0, 'OCL_ICD_FILENAMES': 'libintelocl_emu.so:libalteracl.so:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64/libintelocl.so', '_ModuleTable017_': 'bG9hZE9yZGVyID0gNDgsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ2l0LWxmcy8zLjEuMiIsCndWID0gIjAwMDAwMDAwMy4wMDAwMDAwMDEuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbImdsb2J1cy1jbGkiXSA9IHsKZm4gPSAiL2FwcHMvb3RoZXIvbW9kdWxlZmlsZXMvZ2xvYnVzLWNsaS8zLjM1LjIubHVhIiwKZnVsbE5hbWUgPSAiZ2xvYnVzLWNsaS8zLjM1LjIiLApsb2FkT3JkZXIgPSA4MywKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJnbG9idXMtY2xpIiwKd1YgPSAiMDAwMDAwMDAzLjAwMDAwMDAzNS4wMDAwMDAwMDIu', 'DO_FETCH_HPSS': False, 'PROJ_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/proj', 'FHOUT_ICE': 3, 'COM_WAVE_STATION_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station', 'COMINecmwf': '/work2/noaa/global/role-global/data/external_gempak/ecmwf', 'COM_TOP_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}', 'WGRIB2_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include', 'mpmd_opt': '--multi-prog --output=mpmd.%j.%t.out', 'DO_PREP_SFC': False, 'LMOD_G2TMPLVIRT_NAME': 'g2tmpl', '_ModuleTable066_': 'bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5dGhvbi8zLjExLjYubHVhIiwKZnVsbE5hbWUgPSAicHl0aG9uLzMuMTEuNiIsCmxvYWRPcmRlciA9IDE1LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5dGhvbi8zLjExLjYiLAp3ViA9ICIwMDAwMDAwMDMuMDAwMDAwMDExLjAwMDAwMDAwNi4qemZpbmFsIiwKfSwKc25hcHB5ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAy', '_ModuleTable009_': 'MjEuOS4wL2N1cmwvOC40LjAubHVhIiwKZnVsbE5hbWUgPSAiY3VybC84LjQuMCIsCmxvYWRPcmRlciA9IDIzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImN1cmwvOC40LjAiLAp3ViA9ICIwMDAwMDAwMDguMDAwMDAwMDA0Lip6ZmluYWwiLAp9LAplY2NvZGVzID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2VjY29kZXMvMi4zMi4wLmx1YSIsCmZ1bGxOYW1lID0gImVjY29kZXMvMi4z', 'py_numexpr_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq', '__LMOD_REF_COUNT_MANPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/share/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/man:2;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/share/man:2;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/man:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/documentation/en/man/common:1;/apps/other/lmod/lmod/share/man:1;/usr/share/man:1;/apps/share/man:1;/apps/man:1;/opt/slurm/share/man:1', 'DOLETKF_OCN': False, 'MPI_GROUP_MAX': 256, 'py_cftime_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i', 'XDG_SESSION_ID': 'c142', 'zstd_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4', 'CLUSTERS': '', '_ModuleTable034_': 'bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLW9uZWFwaS1tcGkvMjAyMS45LjAvaW50ZWwvMjAyMS45LjAvbmV0Y2RmLWMvNC45LjIubHVhIiwKZnVsbE5hbWUgPSAibmV0Y2RmLWMvNC45LjIiLApsb2FkT3JkZXIgPSAzMCwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJuZXRjZGYtYy80LjkuMiIsCndWID0gIjAwMDAwMDAwNC4wMDAwMDAwMDkuMDAwMDAwMDAyLip6ZmluYWwiLAp9LApbIm5ldGNkZi1jeHg0Il0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2st', 'COM_OCEAN_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean', 'CLUSTERS_SERVICE': '', 'pgmout': 'OUTPUT.1367274', 'EXECfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/exec', '_ModuleTable031_': 'aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJtZXQvOS4xLjMiLAp3ViA9ICIwMDAwMDAwMDkuMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKbWV0cGx1cyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wL2ludGVsLzIwMjEuOS4wL21ldHBsdXMvMy4xLjEubHVhIiwKZnVsbE5hbWUgPSAibWV0cGx1cy8zLjEuMSIsCmxvYWRPcmRlciA9IDc4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIs', 'COM_ATMOS_ANLMON_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon', 'HOMEobsproc': '/work2/noaa/global/role-global/git/obsproc/v1.2.0', 'esmf_ver': '8.6.0', 'DO_ARCHCOM': False, 'SLURM_SUBMIT_HOST': 'hercules-login-1.hpc.msstate.edu', 'WRITE_DOPOST': True, 'EXECcfs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/exec', 'DUMP_SUFFIX': '', '_ModuleTable072_': 'CnVzZXJOYW1lID0gInRhci8xLjM0IiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAzNC4qemZpbmFsIiwKfSwKdWR1bml0cyA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC91ZHVuaXRzLzIuMi4yOC5sdWEiLApmdWxsTmFtZSA9ICJ1ZHVuaXRzLzIuMi4yOCIsCmxvYWRPcmRlciA9IDI1LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAyLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInVkdW5pdHMvMi4yLjI4IiwKd1YgPSAiMDAwMDAwMDAy', 'COMOUT_OCEAN_GRIB': '/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/ocean/grib2', 'CASE': 'C48', 'SENDDBN': False, 'gettext_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43', 'COM_ICE_BMATRIX_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice', 'git_lfs_ROOT': '/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l', '_ModuleTable001_': 'X01vZHVsZVRhYmxlXyA9IHsKTVR2ZXJzaW9uID0gMywKY19yZWJ1aWxkVGltZSA9IGZhbHNlLApjX3Nob3J0VGltZSA9IGZhbHNlLApkZXB0aFQgPSB7fSwKZmFtaWx5ID0gewpNZXRhQ29tcGlsZXIgPSAic3RhY2staW50ZWwiLApNZXRhTVBJID0gInN0YWNrLWludGVsLW9uZWFwaS1tcGkiLApjb21waWxlciA9ICJpbnRlbC1vbmVhcGktY29tcGlsZXJzIiwKZzJ0bXBsdmlydCA9ICJnMnRtcGwiLAptcGkgPSAiaW50ZWwtb25lYXBpLW1waSIsCn0sCm1UID0gewphbnRsciA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9t', 'COM_ICE_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart', '_ModuleTable025_': 'LAp9LAppcCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9pcC80LjMuMC5sdWEiLApmdWxsTmFtZSA9ICJpcC80LjMuMCIsCmxvYWRPcmRlciA9IDQzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImlwLzQuMy4wIiwKd1YgPSAiMDAwMDAwMDA0LjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKamFzcGVyID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hl', 'COM_OCEAN_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart', 'ARCHCOM_TO': 'globus_hpss', 'COM_ATMOS_WMO_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo', 'ip_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf', 'CLASSPATH': '/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/mpi.jar', 'INTELFPGAOCLSDKROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga', 'LD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/host/linux64/lib:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/lib:/usr/lib64:/usr/lib:/opt/slurm/lib:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib', 'DO_CA': True, 'g2_ver': '3.4.5', 'LMOD_FAMILY_COMPILER': 'intel-oneapi-compilers', 'XDG_RUNTIME_DIR': '/run/user/9583', 'BASE_ENV': '/work2/noaa/global/mterry/global-workflow_forked/env', 'SLURM_JOB_ID': 5951722, '_ModuleTable014_': 'dWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9nMmMvMS44LjAubHVhIiwKZnVsbE5hbWUgPSAiZzJjLzEuOC4wIiwKbG9hZE9yZGVyID0gNzUsCnByb3BUID0ge30sCnJlZl9jb3VudCA9IDEsCnN0YWNrRGVwdGggPSAyLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZzJjLzEuOC4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwOC4qemZpbmFsIiwKfSwKZzJ0bXBsID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9p', 'OCEANICEPRODUCTS_CONFIG': '/work2/noaa/global/mterry/global-workflow_forked/parm/post/oceanice_products_gfs.yaml', 'NTHREADS_OCNICEPOST': 1, 'DO_JEDIOCNVAR': False, 'FIXreg2grb2': '/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2', 'SLURM_NODEID': 0, 'ip_ver': '4.3.0', 'KMP_AFFINITY': 'scatter', 'FHOUT_ICE_GFS': 6, '_ModuleTable065_': 'L3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS14bHd0LzEuMy4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXhsd3QvMS4zLjAiLApsb2FkT3JkZXIgPSA3MSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS14bHd0LzEuMy4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKcHl0aG9uID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1', 'DO_ATM': True, '_ModuleTable004_': 'IDUzLApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImJ1ZnIvMTEuNy4wIiwKd1YgPSAiMDAwMDAwMDExLjAwMDAwMDAwNy4qemZpbmFsIiwKfSwKWyJjLWJsb3NjIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvYy1ibG9zYy8xLjIxLjUubHVhIiwKZnVsbE5hbWUgPSAiYy1ibG9zYy8xLjIxLjUiLApsb2FkT3JkZXIgPSAyOSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDIsCnN0YXR1', 'esmf_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep', 'LMOD_FAMILY_MPI_VERSION': '2021.9.0', '_ModuleTable020_': 'c3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwtb25lYXBpLW1waS8yMDIxLjkuMC9pbnRlbC8yMDIxLjkuMC9nc2ktbmNkaWFnLzEuMS4yLmx1YSIsCmZ1bGxOYW1lID0gImdzaS1uY2RpYWcvMS4xLjIiLApsb2FkT3JkZXIgPSA0NiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJnc2ktbmNkaWFnLzEuMS4yIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMS4wMDAwMDAwMDIuKnpmaW5hbCIsCn0sCmdzbCA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYu', 'DO_JEDIATMVAR': False, 'DO_AERO_ANL': False, 'NCLEN': '/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen', 'PS4': '+ $(basename ${BASH_SOURCE[0]:-${FUNCNAME[0]:-"Unknown"}})[${LINENO}]', 'COM_SNOW_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow', '_ModuleTable053_': 'YWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcHktbnVtcHkvMS4yMy40Lmx1YSIsCmZ1bGxOYW1lID0gInB5LW51bXB5LzEuMjMuNCIsCmxvYWRPcmRlciA9IDUyLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSA4LApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LW51bXB5LzEuMjMuNCIsCndWID0gIjAwMDAwMDAwMS4wMDAwMDAwMjMuMDAwMDAwMDA0Lip6ZmluYWwiLAp9LApbInB5LW9wZW5weXhsIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0', '__LMOD_REF_COUNT_ACL_BOARD_VENDOR_PATH': '/opt/Intel/OpenCLFPGA/oneAPI/Boards:1', 'LMOD_FAMILY_METAMPI_VERSION': '2021.9.0', '_ModuleTable059_': 'IDY4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LXB5eGxzYi8xLjAuMTAiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAwLjAwMDAwMDAxMC4qemZpbmFsIiwKfSwKWyJweS1weXlhbWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1weXlhbWwvNi4wLmx1YSIsCmZ1bGxOYW1lID0gInB5LXB5eWFtbC82LjAiLApsb2FkT3JkZXIgPSA2MCwKcHJvcFQgPSB7fSwKc3RhY2tE', 'DO_CALC_INCREMENT_ENKF_GFS': False, 'MKLROOT': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0', 'DEBUGINFOD_URLS': 'https://debuginfod.centos.org/ ', 'sven_root_path': '/home/gfekete/sven', '_ModuleTable047_': 'LApbInB5LWV0LXhtbGZpbGUiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1ldC14bWxmaWxlLzEuMC4xLmx1YSIsCmZ1bGxOYW1lID0gInB5LWV0LXhtbGZpbGUvMS4wLjEiLApsb2FkT3JkZXIgPSA2NSwKcHJvcFQgPSB7fSwKcmVmX2NvdW50ID0gMSwKc3RhY2tEZXB0aCA9IDMsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1ldC14bWxmaWxlLzEuMC4xIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMC4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsi', 'restart_interval_enkfgfs': 3, 'SENDDBN_NTC': False, 'FIXaer': '/work2/noaa/global/mterry/global-workflow_forked/fix/aer', 'FHMAX_HF_GFS': 48, 'COM_ICE_LETKF_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf', 'LMOD_FAMILY_MPI': 'intel-oneapi-mpi', 'netcdf_fortran_ver': '4.6.1', 'FIXmom': '/work2/noaa/global/mterry/global-workflow_forked/fix/mom6', 'cyc': 12, 'BACIO_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib/libbacio_4.a', 'FC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/ifort', 'FHOUT_WAV_GFS': 3, 'SDATE_GFS': datetime.datetime(2021, 3, 23, 12, 0), 'BACIO_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib/libbacio_8.a', 'sp_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr', 'OPS_RES': 'C768', 'NMEM_ENS_GFS': 30, 'which_declare': 'declare -f', '_ModuleTable056_': 'OS4wL3B5LXBhbmRhcy8xLjUuMy5sdWEiLApmdWxsTmFtZSA9ICJweS1wYW5kYXMvMS41LjMiLApsb2FkT3JkZXIgPSA3MiwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0aCA9IDEsCnN0YXR1cyA9ICJhY3RpdmUiLAp1c2VyTmFtZSA9ICJweS1wYW5kYXMvMS41LjMiLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA1LjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKWyJweS1weXRob24tZGF0ZXV0aWwiXSA9IHsKZm4gPSAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9weS1weXRob24tZGF0ZXV0aWwvMi44', 'COM_WAVE_GRID_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded', 'LMOD_FAMILY_G2TMPLVIRT': 'g2tmpl', 'QUEUE_DTN': 'batch', 'DO_GENESIS_FSU': False, 'prod_util_ver': '2.1.1', 'MPI_F90': 'mpiifort', 'LMOD_FAMILY_COMPILER_VERSION': '2023.1.0', '_ModuleTable042_': 'MDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMvaW50ZWwvMjAyMS45LjAvcGlnei8yLjcubHVhIiwKZnVsbE5hbWUgPSAicGlnei8yLjciLApsb2FkT3JkZXIgPSA4LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gNCwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInBpZ3ovMi43IiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwNy4qemZpbmFsIiwKfSwKcHJlcG9icyA9IHsKZm4gPSAiL3dvcmsyL25vYWEvZ2xvYmFsL3JvbGUtZ2xvYmFsL2dpdC9wcmVwb2JzL3YxLjEuMC9tb2R1bGVmaWxlcy9wcmVwb2JzLzEuMS4wLmx1YSIsCmZ1bGxOYW1lID0gInByZXBvYnMvMS4xLjAiLApsb2FkT3JkZXIgPSA4MSwKcHJvcFQgPSB7fSwKc3RhY2tEZXB0', 'openjpeg_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu', 'USHfit2obs': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/ush', 'YAML_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib', 'DO_GSISOILDA': False, 'nemsio_ver': '2.5.4', 'DO_NPOESS': False, 'SLURM_MPI_TYPE': 'pmi2', 'OMP_STACKSIZE': 2048000, 'py_et_xmlfile_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq', 'FORECAST_HOUR': 102, 'ATARDIR': '/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW', 'EXECgfs': '/work2/noaa/global/mterry/global-workflow_forked/exec', 'py_python_dateutil_ver': '2.8.2', 'assim_freq': 6, 'DOIAU': True, 'FHOUT_WAV': 1, 'py_pyyaml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2', 'met_ver': '9.1.3', 'gsi_ncdiag_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe', 'DO_FETCH_LOCAL': False, 'NLSPATH': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64/locale/%l_%t/%N:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin/locale/%l_%t/%N', 'netcdf_fortran_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe', 'SP_LIB4': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_4.a', 'COM_OCEAN_BMATRIX_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean', 'SLURM_CONF': '/var/spool/slurmd/conf-cache/slurm.conf', 'PATH': '/apps/other/globus-cli-3.35.2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r/ush:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/met-9.1.3-tbjrp4b/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gmake-4.2.1-2w4aspq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/bin:/apps/spack-managed/gcc-11.3.1/git-lfs-3.1.2-sjfqfgha27na65g3lrcqamncnryjoa7l/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/tar-1.34-rlgkths/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/bin:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/bin/intel64:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/bin:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/bin:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/bin:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64:/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin:/usr/sbin:/usr/bin:/apps/sbin:/apps/bin:/opt/slurm/bin:/home/gfekete/sven/bin', 'SP_LIB8': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_8.a', 'netcdf_c_ver': '4.9.2', 'SLURM_JOB_NAME': 'C48_S2SW_gfs_ocean_prod_f102_12', 'COM_ICE_HISTORY_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history', 'NLN': '/bin/ln -sf', 'MODULEPATH': '/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles:/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0-a66eaip/g2tmpl/1.13.0/intel/2021.9.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/g2tmpl/1.13.0/intel/2021.9.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/intel-oneapi-mpi/2021.9.0-a66eaip/oneapi/2023.1.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core:/work2/noaa/global/mterry/global-workflow_forked/modulefiles:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core:/apps/other/modulefiles:/apps/containers/modulefiles:/apps/licensed/modulefiles:/apps/contrib/modulefiles', 'CC': '/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/bin/intel64/icc', 'grib_util_ver': '1.3.0', 'py_f90nml_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z', 'DO_MERGENSST': False, 'SLURM_NTASKS_PER_NODE': 1, '_LMFILES_': '/apps/other/modulefiles/contrib/0.1:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core/intel-oneapi-compilers/2023.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/Core/stack-intel/2021.9.0.lua:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/oneapi/2023.1.0/intel-oneapi-mpi/2021.9.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/stack-intel-oneapi-mpi/2021.9.0.lua:/apps/spack-managed/modulefiles/linux-rocky9-x86_64/Core/intel-oneapi-mkl/2023.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/zlib/1.2.13.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/pigz/2.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/zstd/1.5.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/tar/1.34.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gettext/0.21.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libxcrypt/4.4.35.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/sqlite/3.43.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/util-linux-uuid/2.38.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/python/3.11.6.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libjpeg/2.1.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/jasper/2.0.32.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libpng/1.6.37.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/openjpeg/2.3.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/eccodes/2.32.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/fftw/3.3.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/nghttp2/1.57.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/curl/8.4.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/proj/9.2.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/udunits/2.2.28.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/cdo/2.2.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/hdf5/1.14.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/snappy/1.1.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/c-blosc/1.21.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-c/4.9.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-fortran/4.6.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/parallel-netcdf/1.12.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/parallelio/2.5.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/esmf/8.6.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/antlr/2.7.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gsl/2.7.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/nco/5.0.6.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/bacio/2.4.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/w3emc/2.10.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/prod_util/2.1.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2/3.4.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/sp/2.5.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/ip/4.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/grib-util/1.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2tmpl/1.13.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/gsi-ncdiag/1.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/crtm-fix/2.4.0.1_emc.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/git-lfs/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/crtm/2.4.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/openblas/0.3.24.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-setuptools/63.4.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-numpy/1.23.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/bufr/11.7.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/gmake/4.2.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/wgrib2/2.0.8.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-f90nml/1.4.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-cftime/1.0.3.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/py-netcdf4/1.5.8.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/libyaml/0.2.5.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pyyaml/6.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-markupsafe/2.1.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-jinja2/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-bottleneck/1.3.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-numexpr/2.8.4.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-et-xmlfile/1.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-openpyxl/3.1.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pytz/2023.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pyxlsb/1.0.10.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlrd/2.0.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlsxwriter/3.1.7.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xlwt/1.3.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-pandas/1.5.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-six/1.16.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-python-dateutil/2.8.2.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/g2c/1.8.0.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/netcdf-cxx4/4.3.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/met/9.1.3.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel-oneapi-mpi/2021.9.0/intel/2021.9.0/metplus/3.1.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-packaging/23.1.lua:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/modulefiles/intel/2021.9.0/py-xarray/2023.7.0.lua:/work2/noaa/global/role-global/git/prepobs/v1.1.0/modulefiles/prepobs/1.1.0.lua:/work2/noaa/global/role-global/git/Fit2Obs/v1.1.7.1/modulefiles/fit2obs/1.1.7.1.lua:/apps/other/modulefiles/globus-cli/3.35.2.lua:/work2/noaa/global/mterry/global-workflow_forked/modulefiles/module_base.hercules.lua', 'globus_cli_ver': 3.27, '_ModuleTable028_': 'L2xpYnBuZy8xLjYuMzcubHVhIiwKZnVsbE5hbWUgPSAibGlicG5nLzEuNi4zNyIsCmxvYWRPcmRlciA9IDE4LApwcm9wVCA9IHt9LApzdGFja0RlcHRoID0gMSwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gImxpYnBuZy8xLjYuMzciLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDA2LjAwMDAwMDAzNy4qemZpbmFsIiwKfSwKbGlieGNyeXB0ID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL2xpYnhjcnlwdC80LjQuMzUubHVhIiwKZnVsbE5hbWUgPSAibGlieGNyeXB0', 'py_numpy_ver': '1.23.4', '__LMOD_REF_COUNT_HDF5_PLUGIN_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/plugins:1', 'bufr_ver': '11.7.0', 'COMINnam': '/work2/noaa/global/role-global/data/external_gempak/nam', 'DBUS_SESSION_BUS_ADDRESS': 'unix:path=/run/user/9583/bus', 'DO_AERO_FCST': False, 'py_pandas_ver': '1.5.3', '_ModuleTable039_': 'dGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9vcGVuanBlZy8yLjMuMS5sdWEiLApmdWxsTmFtZSA9ICJvcGVuanBlZy8yLjMuMSIsCmxvYWRPcmRlciA9IDE5LApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMywKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gIm9wZW5qcGVnLzIuMy4xIiwKd1YgPSAiMDAwMDAwMDAyLjAwMDAwMDAwMy4wMDAwMDAwMDEuKnpmaW5hbCIsCn0sClsicGFyYWxsZWwtbmV0Y2RmIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMvc3BhY2stc3RhY2stMS42LjAvZW52cy9nc2ktYWRkb24tZGV2LWZtcy0yMDI0LjAxL2luc3RhbGwvbW9kdWxlZmlsZXMv', 'COM_ICE_ANALYSIS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice', 'UUID_ORION_DTN': '8a10dd4f-24ee-4794-a39d-9c313ab6a34b', 'MPI_F77': 'mpiifort', 'LMOD_CMD': '/apps/other/lmod/lmod/libexec/lmod', '_ModuleTable045_': 'aW5hbCIsCn0sClsicHktYm90dGxlbmVjayJdID0gewpmbiA9ICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVzL3NwYWNrLXN0YWNrLTEuNi4wL2VudnMvZ3NpLWFkZG9uLWRldi1mbXMtMjAyNC4wMS9pbnN0YWxsL21vZHVsZWZpbGVzL2ludGVsLzIwMjEuOS4wL3B5LWJvdHRsZW5lY2svMS4zLjcubHVhIiwKZnVsbE5hbWUgPSAicHktYm90dGxlbmVjay8xLjMuNyIsCmxvYWRPcmRlciA9IDYzLApwcm9wVCA9IHt9LApyZWZfY291bnQgPSAxLApzdGFja0RlcHRoID0gMiwKc3RhdHVzID0gImFjdGl2ZSIsCnVzZXJOYW1lID0gInB5LWJvdHRsZW5lY2svMS4zLjciLAp3ViA9ICIwMDAwMDAwMDEuMDAwMDAwMDAzLjAwMDAwMDAwNy4qemZpbmFs', 'LMOD_MPI_VERSION': '2021.9.0-a66eaip', 'LMOD_SYSTEM_NAME': 'hercules', 'RUN_ENVIR': 'emc', 'w3emc_ver': '2.10.0', '__LMOD_REF_COUNT_LD_LIBRARY_PATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xarray-2023.7.0-x743hx4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-packaging-23.1-cvb54w6/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-python-dateutil-2.8.2-crskgyy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-six-1.16.0-dtjcn7o/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pandas-1.5.3-7tjsvkw/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlwt-1.3.0-ghvbw56/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlsxwriter-3.1.7-rqogaok/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-xlrd-2.0.1-g52jf7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyxlsb-1.0.10-ow7w6ge/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pytz-2023.3-kmu3i3s/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-openpyxl-3.1.2-vfourqe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-et-xmlfile-1.0.1-7wze7yq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numexpr-2.8.4-nobvkrq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-bottleneck-1.3.7-ftw2r3a/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-jinja2-3.1.2-eoja5pj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-markupsafe-2.1.3-v6xcnb7/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-pyyaml-6.0-t2st4x2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-netcdf4-1.5.8-bgfhv5i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-cftime-1.0.3.4-ienmq3i/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-f90nml-1.4.3-3ib7u6z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-numpy-1.23.4-ftta6ne/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/py-setuptools-63.4.3-5extjhr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/ip-4.3.0-5c2fcrf/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/g2-3.4.5-anw5ato/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/w3emc-2.10.0-zc33leo/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/bacio-2.4.1-kocc5wj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nco-5.0.6-r3dr4pq/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/lib64:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/lib/intel64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/libfabric/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib/release:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/lib:1;/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/x64:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/lib/oclfpga/host/linux64/lib:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/compiler/2023.1.0/linux/compiler/lib/intel64_lin:1;/apps/spack-managed/gcc-11.3.1/intel-oneapi-compilers-2023.1.0-sb753366rvywq75zeg4ml5k5c72xgj72/lib:1;/usr/lib64:1;/usr/lib:1;/opt/slurm/lib:1;/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/lib:2', 'DOHYBVAR_OCN': False, 'NUM_SND_COLLECTIVES': 9, 'walltime': '00:15:00', 'netcdf_cxx4_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu', 'ens_tracker_ver': 'v1.2.0', 'LMOD_SYSTEM_DEFAULT_MODULES': 'contrib', 'OMPI_MCA_plm_slurm_args': '--external-launcher', '_ModuleTable078_': 'cyIKLCAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC1vbmVhcGktbXBpLzIwMjEuOS4wLWE2NmVhaXAvZzJ0bXBsLzEuMTMuMC9pbnRlbC8yMDIxLjkuMCIKLCAiL3dvcmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9nMnRtcGwvMS4xMy4wL2ludGVsLzIwMjEuOS4wIgosICIvd29yay9ub2FhL2VwaWMvcm9sZS1lcGljL3NwYWNrLXN0YWNrL2hlcmN1bGVz', 'FIXcpl': '/work2/noaa/global/mterry/global-workflow_forked/fix/cpl', 'G2C_LIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/lib64/libg2c.so', '_ModuleTable019_': 'cmsvbm9hYS9lcGljL3JvbGUtZXBpYy9zcGFjay1zdGFjay9oZXJjdWxlcy9zcGFjay1zdGFjay0xLjYuMC9lbnZzL2dzaS1hZGRvbi1kZXYtZm1zLTIwMjQuMDEvaW5zdGFsbC9tb2R1bGVmaWxlcy9pbnRlbC8yMDIxLjkuMC9ncmliLXV0aWwvMS4zLjAubHVhIiwKZnVsbE5hbWUgPSAiZ3JpYi11dGlsLzEuMy4wIiwKbG9hZE9yZGVyID0gNDQsCnByb3BUID0ge30sCnN0YWNrRGVwdGggPSAxLApzdGF0dXMgPSAiYWN0aXZlIiwKdXNlck5hbWUgPSAiZ3JpYi11dGlsLzEuMy4wIiwKd1YgPSAiMDAwMDAwMDAxLjAwMDAwMDAwMy4qemZpbmFsIiwKfSwKWyJnc2ktbmNkaWFnIl0gPSB7CmZuID0gIi93b3JrL25vYWEvZXBpYy9yb2xlLWVwaWMvc3BhY2stc3RhY2svaGVyY3VsZXMv', 'SLURM_JOB_GID': 17000, 'SLURM_GET_USER_ENV': 1, 'IAUFHRS_ENKF': [3, 6, 9], 'CPATH': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-cxx4-4.3.1-ulwplbu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2c-1.8.0-hmyvqtp/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libyaml-0.2.5-o55kg5x/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/wgrib2-2.0.8-53fnkln/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/bufr-11.7.0-7qdgt6m/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openblas-0.3.24-zsule7f/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/crtm-2.4.0.1-wlgbu5r/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsi-ncdiag-1.1.2-pggoyfe/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gsl-2.7.1-uoz625z/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/antlr-2.7.7-jly57m2/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/esmf-8.6.0-rqrapep/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallelio-2.5.10-rdwrsed/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/parallel-netcdf-1.12.2-x3m7oy3/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-fortran-4.6.1-ndsugbe/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/c-blosc-1.21.5-2ziecvn/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/snappy-1.1.10-5lcp3lx/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/hdf5-1.14.0-htxkrrh/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/cdo-2.2.0-rqg7s5e/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/udunits-2.2.28-sprwyzo/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/proj-9.2.1-ka244ez/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/curl-8.4.0-g3ljops/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/nghttp2-1.57.0-64rceky/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/fftw-3.3.10-botgfbj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/eccodes-2.32.0-uxjtmuj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/openjpeg-2.3.1-ae7flgu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/jasper-2.0.32-jk3acwt/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libjpeg-turbo-2.1.0-6ziftfy/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/python-3.11.6-ajcuas2/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/util-linux-uuid-2.38.1-rdrx3jj/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sqlite-3.43.2-afpdum4/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libxcrypt-4.4.35-5pcpqqr/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/gettext-0.21.1-djc4g43/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zstd-1.5.2-pyuk7p4/include:/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/zlib-1.2.13-xsm2i7j/include:/apps/spack-managed/gcc-11.3.1/intel-oneapi-mkl-2023.1.0-4cujjco7etbwl34hwrtw3ree7dwhxnci/mkl/2023.1.0/include:/apps/spack-managed/oneapi-2023.1.0/intel-oneapi-mpi-2021.9.0-a66eaipzsnyrdgaqzxmqmqz64qzvhkse/mpi/2021.9.0/include', 'G2TMPL_INC': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/g2tmpl-1.13.0-ifdzkel/include', 'DO_JEDISNOWDA': False, 'SP_LIBd': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/sp-2.5.0-hidqzcr/lib64/libsp_d.a', 'pigz_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/pigz-2.7-cwgjiyq', 'COM_ATMOS_BUFR_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr', 'COM_OCEAN_GRIB_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2', 'CNVGRIB': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/cnvgrib', 'COM_ATMOS_TRACK_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks', 'DO_VERFRAD': True, 'pgmerr': 'errfile', 'MPI_BUFS_PER_HOST': 2048, 'DO_PREP_OBS_AERO': False, 'LMOD_FAMILY_G2TMPLVIRT_VERSION': '1.13.0', 'PNG_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/libpng-1.6.37-bnpuvlu', 'COM_ATMOS_RESTART_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart', 'DEGRIB2': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/grib-util-1.3.0-76tgs35/bin/degrib2', 'DO_ANLSTAT': False, 'LMOD_MPI_NAME': 'intel-oneapi-mpi', 'py_f90nml_ver': '1.4.3', 'NMEM_ENS_GFS_OFFSET': 20, 'SMOOTH_ENKF': False, 'FHMAX_GOES': 120, 'OLDPWD': '/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW', 'QUEUE': 'batch', 'COM_OBS_TMPL': '${ROTDIR}/${RUN}.${YMD}/${HH}/obs', 'DO_VMINMON': True, 'metplus_ROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/gsi-addon-dev-fms-2024.01/install/intel/2021.9.0/metplus-3.1.1-jdsse5r', 'VERBOSE': True, 'SLURM_JOB_NODELIST': 'hercules-01-10', 'CLUSTER': 'hercules', 'DBNROOT': '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn', 'FHOUT_HF_WAV': 1, 'PARMgfs': '/work2/noaa/global/mterry/global-workflow_forked/parm', 'I_MPI_HYDRA_BOOTSTRAP': 'slurm', 'APRUN_OCNICEPOST': 'srun -l --export=ALL --hint=nomultithread -n 1 --cpus-per-task=1', 'BASH_FUNC_ml%%': '() { eval "$($LMOD_DIR/ml_cmd "$@")"\n}', 'BASH_FUNC_which%%': '() { ( alias;\n eval ${which_declare} ) | /usr/bin/which --tty-only --read-alias --read-functions --show-tilde --show-dot $@\n}', 'BASH_FUNC_module%%': '() { if [ -z "${LMOD_SH_DBG_ON+x}" ]; then\n case "$-" in \n *v*x*)\n __lmod_sh_dbg=\'vx\'\n ;;\n *v*)\n __lmod_sh_dbg=\'v\'\n ;;\n *x*)\n __lmod_sh_dbg=\'x\'\n ;;\n esac;\n fi;\n if [ -n "${__lmod_sh_dbg:-}" ]; then\n set +$__lmod_sh_dbg;\n echo "Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for Lmod\'s output" 1>&2;\n fi;\n eval "$($LMOD_CMD $LMOD_SHELL_PRGM "$@")" && eval "$(${LMOD_SETTARG_CMD:-:} -s sh)";\n __lmod_my_status=$?;\n if [ -n "${__lmod_sh_dbg:-}" ]; then\n echo "Shell debugging restarted" 1>&2;\n set -$__lmod_sh_dbg;\n fi;\n unset __lmod_sh_dbg;\n return $__lmod_my_status\n}', 'BASH_FUNC_declare_from_tmpl%%': '() { if [[ ${DEBUG_WORKFLOW:-"NO"} == "NO" ]]; then\n set +x;\n fi;\n local opts="-g";\n local OPTIND=1;\n while getopts "rx" option; do\n opts="${opts}${option}";\n done;\n shift $((OPTIND-1));\n for input in "$@";\n do\n IFS=\':\' read -ra args <<< "${input}";\n local com_var="${args[0]}";\n local template;\n local value;\n if (( ${#args[@]} > 1 )); then\n template="${args[1]}";\n else\n template="${com_var}_TMPL";\n fi;\n if [[ ! -v "${template}" ]]; then\n echo "FATAL ERROR in declare_from_tmpl: Requested template ${template} not defined!";\n exit 2;\n fi;\n value=$(echo "${!template}" | envsubst);\n declare ${opts} "${com_var}"="${value}";\n echo "declare_from_tmpl :: ${com_var}=${value}";\n done;\n set_trace\n}', 'BASH_FUNC_err_exit%%': ['() { set +eux;\n msg1=${*:-Job ${jobid} failed};\n if [[ -n "${pgm}" ]]; then\n msg1+="', 'ERROR IN ${pgm}";\n fi;\n if [[ -n "${err}" ]]; then\n msg1+=" RETURN CODE ${err}";\n fi;\n msg2="\n -------------------------------------------------------------\n -- FATAL ERROR: ${msg1}\n -- ABNORMAL EXIT at $(date) on ${HOSTNAME}\n -------------------------------------------------------------\n ";\n echo "${msg2}" 1>&2;\n module list;\n echo "" 1>&2;\n echo "${msg1}" 1>&2;\n if [[ -n "${DATA}" ]]; then\n echo "${DATA}" 1>&2;\n ls -ltr "${DATA}" 1>&2;\n else\n echo "WARNING: DATA variable not defined" 1>&2;\n fi;\n if [[ -n "${pgmout}" ]]; then\n if [[ -s errfile ]]; then\n echo "----- contents of errfile -----" >> "${pgmout}";\n cat errfile >> "${pgmout}";\n fi;\n cat "${pgmout}" 1>&2;\n else\n if [[ -s errfile ]]; then\n cat errfile 1>&2;\n fi;\n fi;\n if [[ "${SENDECF}" == "YES" ]]; then\n timeout 30 ecflow_client --msg "${ECF_NAME}: ${msg1}";\n timeout 30 ssh "${ECF_HOST}" "echo \\"${msg}2\\" >> ${ECF_JOBOUT:?}";\n fi;\n if [[ "${SENDECF}" == "YES" ]]; then\n ecflow_client --kill="${ECF_NAME:?}";\n fi;\n if [[ -n "${PBS_JOBID}" ]]; then\n qdel "${PBS_JOBID}";\n else\n if [[ -n "${SLURM_JOB_ID}" ]]; then\n scancel "${SLURM_JOB_ID}";\n fi;\n fi\n}'], 'BASH_FUNC_wait_for_file%%': '() { set +x;\n local file_name=${1:?"wait_for_file() requires a file name"};\n local sleep_interval=${2:-60};\n local max_tries=${3:-100};\n for ((iter=0; iter' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wave -+++ config.wave[6]echo 'BEGIN: config.wave' -BEGIN: config.wave -+++ config.wave[13]export RUNRSTwave=gdas -+++ config.wave[13]RUNRSTwave=gdas -+++ config.wave[16]export MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[16]MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[19]case "${waveGRD}" in -+++ config.wave[64]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[64]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[65]export wavepostGRD= -+++ config.wave[65]wavepostGRD= -+++ config.wave[66]export waveuoutpGRD=uglo_100km -+++ config.wave[66]waveuoutpGRD=uglo_100km -+++ config.wave[75]export WAVEWND_DID= -+++ config.wave[75]WAVEWND_DID= -+++ config.wave[76]export WAVEWND_FID= -+++ config.wave[76]WAVEWND_FID= -+++ config.wave[79][[ gfs == \g\f\s ]] -+++ config.wave[80]export FHMAX_WAV=120 -+++ config.wave[80]FHMAX_WAV=120 -+++ config.wave[82]export WAVHINDH=0 -+++ config.wave[82]WAVHINDH=0 -+++ config.wave[83]export FHMAX_WAV_IBP=180 -+++ config.wave[83]FHMAX_WAV_IBP=180 -+++ config.wave[84](( FHMAX_WAV < FHMAX_WAV_IBP )) -+++ config.wave[84]export FHMAX_WAV_IBP=120 -+++ config.wave[84]FHMAX_WAV_IBP=120 -+++ config.wave[87]export DTFLD_WAV=3600 -+++ config.wave[87]DTFLD_WAV=3600 -+++ config.wave[88]export DTPNT_WAV=3600 -+++ config.wave[88]DTPNT_WAV=3600 -+++ config.wave[89]export FHINCP_WAV=1 -+++ config.wave[89]FHINCP_WAV=1 -+++ config.wave[92]export 'OUTPARS_WAV=WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[92]OUTPARS_WAV='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[95][[ gfs == \g\d\a\s ]] -+++ config.wave[99](( INTERVAL_GFS > 0 )) -+++ config.wave[100]export WAVHCYC=6 -+++ config.wave[100]WAVHCYC=6 -+++ config.wave[101]export FHMAX_WAV_CUR=192 -+++ config.wave[101]FHMAX_WAV_CUR=192 -+++ config.wave[109]export RSTTYPE_WAV=T -+++ config.wave[109]RSTTYPE_WAV=T -+++ config.wave[110][[ gfs != gfs ]] -+++ config.wave[115]rst_dt_gfs=43200 -+++ config.wave[116][[ 43200 -gt 0 ]] -+++ config.wave[117]export DT_1_RST_WAV=0 -+++ config.wave[117]DT_1_RST_WAV=0 -+++ config.wave[120]export DT_2_RST_WAV=43200 -+++ config.wave[120]DT_2_RST_WAV=43200 -+++ config.wave[126]export RSTIOFF_WAV=0 -+++ config.wave[126]RSTIOFF_WAV=0 -+++ config.wave[131]export RUNMEM=-1 -+++ config.wave[131]RUNMEM=-1 -+++ config.wave[134](( RUNMEM == -1 )) -+++ config.wave[136]export waveMEMB= -+++ config.wave[136]waveMEMB= -+++ config.wave[143]export WW3ATMINP=CPL -+++ config.wave[143]WW3ATMINP=CPL -+++ config.wave[144][[ YES == \Y\E\S ]] -+++ config.wave[145]export WW3ICEINP=CPL -+++ config.wave[145]WW3ICEINP=CPL -+++ config.wave[146]export WAVEICE_FID= -+++ config.wave[146]WAVEICE_FID= -+++ config.wave[152][[ YES == \Y\E\S ]] -+++ config.wave[153]export WW3CURINP=CPL -+++ config.wave[153]WW3CURINP=CPL -+++ config.wave[154]export WAVECUR_FID= -+++ config.wave[154]WAVECUR_FID= -+++ config.wave[161]export WW3ATMIENS=F -+++ config.wave[161]WW3ATMIENS=F -+++ config.wave[162]export WW3ICEIENS=F -+++ config.wave[162]WW3ICEIENS=F -+++ config.wave[163]export WW3CURIENS=F -+++ config.wave[163]WW3CURIENS=F -+++ config.wave[165]export GOFILETYPE=1 -+++ config.wave[165]GOFILETYPE=1 -+++ config.wave[166]export POFILETYPE=1 -+++ config.wave[166]POFILETYPE=1 -+++ config.wave[170]export FUNIPNT=T -+++ config.wave[170]FUNIPNT=T -+++ config.wave[172]export IOSRV=1 -+++ config.wave[172]IOSRV=1 -+++ config.wave[174]export FPNTPROC=T -+++ config.wave[174]FPNTPROC=T -+++ config.wave[176]export FGRDPROC=F -+++ config.wave[176]FGRDPROC=F -+++ config.wave[178]export FLAGMASKCOMP=F -+++ config.wave[178]FLAGMASKCOMP=F -+++ config.wave[180]export FLAGMASKOUT=F -+++ config.wave[180]FLAGMASKOUT=F -+++ config.wave[182]echo 'END: config.wave' -END: config.wave -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wavepostsbs -+++ config.wavepostsbs[6]echo 'BEGIN: config.wavepostsbs' -BEGIN: config.wavepostsbs -+++ config.wavepostsbs[9]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources wavepostsbs -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=wavepostsbs -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[156]ntasks=8 -++++ config.resources[157]threads_per_task=1 -++++ config.resources[158]tasks_per_node=80 -++++ config.resources[159]NTASKS=8 -++++ config.resources[160]memory=20GB -++++ config.resources[161]walltime=00:15:00 -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export NTASKS -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export memory -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.wavepostsbs[12]export MAX_TASKS=25 -+++ config.wavepostsbs[12]MAX_TASKS=25 -+++ config.wavepostsbs[15]export WAV_SUBGRBSRC= -+++ config.wavepostsbs[15]WAV_SUBGRBSRC= -+++ config.wavepostsbs[16]export WAV_SUBGRB= -+++ config.wavepostsbs[16]WAV_SUBGRB= -+++ config.wavepostsbs[19]export DOFLD_WAV=YES -+++ config.wavepostsbs[19]DOFLD_WAV=YES -+++ config.wavepostsbs[20]export DOPNT_WAV=YES -+++ config.wavepostsbs[20]DOPNT_WAV=YES -+++ config.wavepostsbs[21]export DOGRB_WAV=YES -+++ config.wavepostsbs[21]DOGRB_WAV=YES -+++ config.wavepostsbs[22][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+++ config.wavepostsbs[23]export DOGRI_WAV=YES -+++ config.wavepostsbs[23]DOGRI_WAV=YES -+++ config.wavepostsbs[27]export DOSPC_WAV=YES -+++ config.wavepostsbs[27]DOSPC_WAV=YES -+++ config.wavepostsbs[28]export DOBLL_WAV=YES -+++ config.wavepostsbs[28]DOBLL_WAV=YES -+++ config.wavepostsbs[30]echo 'END: config.wavepostsbs' -END: config.wavepostsbs -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env wavepostsbs -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=wavepostsbs -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 20971520 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 8 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[34]max_threads_per_task=1 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 1 ]] -+++ HERCULES.env[40][[ 1 -gt 1 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 8' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[63]export USE_CFP=YES -+++ HERCULES.env[63]USE_CFP=YES -+++ HERCULES.env[64][[ wavepostsbs == \w\a\v\e\p\r\e\p ]] -+++ HERCULES.env[65]export 'wavempexec=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[65]wavempexec='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[66]export 'wave_mpmd=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[66]wave_mpmd='--multi-prog --output=mpmd.%j.%t.out' -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[4]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ JGLOBAL_WAVE_POST_SBS[7]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[7]HH=12 -+ JGLOBAL_WAVE_POST_SBS[7]declare_from_tmpl -rx COMIN_WAVE_PREP:COM_WAVE_PREP_TMPL COMIN_WAVE_HISTORY:COM_WAVE_HISTORY_TMPL COMOUT_WAVE_GRID:COM_WAVE_GRID_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_WAVE_PREP=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep -declare_from_tmpl :: COMIN_WAVE_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history -declare_from_tmpl :: COMOUT_WAVE_GRID=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded -+ JGLOBAL_WAVE_POST_SBS[12][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n '' ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_atlocn_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_atlocn_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_atlocn_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_epacif_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_epacif_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_epacif_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_wcoast_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_wcoast_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_wcoast_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p50 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p50:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p50=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p50 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=arctic.9km -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_arctic_9km:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_arctic_9km=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_arctic_9km -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_gsouth_0p25:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_gsouth_0p25=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 ]] -+ JGLOBAL_WAVE_POST_SBS[28]/work2/noaa/global/mterry/global-workflow_forked/scripts/exgfs_wave_post_gridded_sbs.sh -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ exgfs_wave_post_gridded_sbs.sh[24]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ exgfs_wave_post_gridded_sbs.sh[26]DOGRI_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[27]DOGRB_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[29]export waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[29]waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[30]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[30]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[31]export wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[31]wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[33]cat - INFO: Grid information: - INFO: Native wave grids: uglo_100km - INFO: Interpolated grids: at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m - INFO: Post-process grids: -++ exgfs_wave_post_gridded_sbs.sh[40]printf %03i 36 -+ exgfs_wave_post_gridded_sbs.sh[40]fhr3=036 -++ exgfs_wave_post_gridded_sbs.sh[41]date -u -d '20210323 12 + 36 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[41]valid_time=2021032500 -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[49]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f036.bin ./out_grd.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f036.bin ./out_grd.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[52][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[62]cat - INFO: Summary: - INFO: Grid interp: DOGRI_WAV="YES" - INFO: Grib files: DOGRB_WAV="YES" - INFO: Fields to be included in grib files: - INFO: OUTPARS_WAV="WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA" -+ exgfs_wave_post_gridded_sbs.sh[70][[ YES == \N\O ]] -+ exgfs_wave_post_gridded_sbs.sh[76]rm -f 'cmdfile.*' cmdfile -+ exgfs_wave_post_gridded_sbs.sh[77]count=0 -+ exgfs_wave_post_gridded_sbs.sh[80][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[81]dt_int=3600. -+ exgfs_wave_post_gridded_sbs.sh[82]n_int=9999 -++ exgfs_wave_post_gridded_sbs.sh[83]date -u -d '20210325 00 - 0 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[83]ymdh_int=2021032500 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=1 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032500 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032500 36 atlocn 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib2_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib2_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=2 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032500 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032500 36 epacif 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib2_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib2_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=3 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032500 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032500 36 wcoast 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib2_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib2_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=4 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032500 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032500 36 global 0p50 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib2_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib2_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=5 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032500 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032500 36 arctic 9km '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib2_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib2_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=6 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032500 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032500 36 global 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib2_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib2_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=7 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032500 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032500 36 gsouth 0p25 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib2_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib2_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[100][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[113][[ 8 -lt 7 ]] -+ exgfs_wave_post_gridded_sbs.sh[121]echo 'INFO: Running MPMD job with 7 commands' -INFO: Running MPMD job with 7 commands -+ exgfs_wave_post_gridded_sbs.sh[122]/work2/noaa/global/mterry/global-workflow_forked/ush/run_mpmd.sh /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/cmdfile -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ run_mpmd.sh[31]source /work2/noaa/global/mterry/global-workflow_forked/ush/preamble.sh -++ preamble.sh[20]set +x -Begin run_mpmd.sh at Tue Jul 29 03:04:38 UTC 2025 -+ run_mpmd.sh[33]cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/cmdfile -+ run_mpmd.sh[36][[ YES != \Y\E\S ]] -+ run_mpmd.sh[46]export OMP_NUM_THREADS=1 -+ run_mpmd.sh[46]OMP_NUM_THREADS=1 -++ run_mpmd.sh[49]wc -l -+ run_mpmd.sh[49]nprocs=7 -+ run_mpmd.sh[52]mpmd_cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mpmd_cmdfile -+ run_mpmd.sh[53][[ -s /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mpmd_cmdfile ]] -+ run_mpmd.sh[55]cat - INFO: Executing MPMD job, STDOUT redirected for each process separately - INFO: On failure, logs for each job will be available in /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mpmd.proc_num.out - INFO: The proc_num corresponds to the line in '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mpmd_cmdfile' -+ run_mpmd.sh[61][[ srun -l --export=ALL --hint=nomultithread =~ ^srun.* ]] -+ run_mpmd.sh[65]nm=0 -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '0 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/cmdfile.1' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '1 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/cmdfile.2' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '2 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/cmdfile.3' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '3 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/cmdfile.4' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '4 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/cmdfile.5' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '5 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/cmdfile.6' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '6 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/cmdfile.7' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[72]set +e -+ run_mpmd.sh[74]srun -l --export=ALL --hint=nomultithread --multi-prog --output=mpmd.%j.%t.out -n 7 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mpmd_cmdfile -+ run_mpmd.sh[75]err=0 -+ run_mpmd.sh[76]set_strict -+ preamble.sh[35][[ YES == \Y\E\S ]] -+ preamble.sh[37]set -eu -+ run_mpmd.sh[103][[ 0 -eq 0 ]] -+ run_mpmd.sh[104]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mpmd_cmdfile -++ run_mpmd.sh[105]find . -name 'mpmd.*.out' -+ run_mpmd.sh[105]out_files='./mpmd.5951789.1.out -./mpmd.5951789.0.out -./mpmd.5951789.3.out -./mpmd.5951789.2.out -./mpmd.5951789.4.out -./mpmd.5951789.6.out -./mpmd.5951789.5.out' -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951789.1.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951789.1.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951789.0.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951789.0.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951789.3.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951789.3.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951789.2.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951789.2.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951789.4.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951789.4.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951789.6.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951789.6.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951789.5.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951789.5.out -+ run_mpmd.sh[110]cat mpmd.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + cmdfile.2[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032500 3600. 9999 -1: + cmdfile.2[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grid_interp_sbs.sh[25]grdID=ep_10m -1: + wave_grid_interp_sbs.sh[26]valid_time=2021032500 -1: + wave_grid_interp_sbs.sh[27]dt=3600. -1: + wave_grid_interp_sbs.sh[28]nst=9999 -1: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551 -1: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.uglo_100km ./out_grd.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mod_def.uglo_100km ./mod_def.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mod_def.ep_10m ./mod_def.ep_10m -1: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ]] -1: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m'\''' -1: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m' -1: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[51]weights_found=1 -1: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 000000' -1: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 000000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/ep_10m/g ww3_gint.inp.tmpl -1: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -1: $ Input file for interpolation of uglo_100km to ep_10m -1: $------------------------------------------------ -1: $ Start Time 3600. NSteps -1: 20210325 000000 3600. 9999 -1: $ Total number of grids -1: 2 -1: $ Grid extensions -1: 'uglo_100km' -1: 'ep_10m' -1: $ -1: 0 -1: $ -1: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[70]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.232246 ']' -1: ++ prep_step[4]echo gfs_ww3_gint.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -1: INFO: Executing 'gfs_ww3_gint.x' -1: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[73]cat grid_interp.ep_10m.out -1: -1: *** WAVEWATCH III Grid interpolation *** -1: =============================================== -1: -1: Comment character is '$' -1: -1: Time Information : -1: --------------------------------------------- -1: Starting Time : 2021/03/25 00:00:00 UTC -1: Interval (in sec) : 3600.00 -1: Number of requests : 9999 -1: --------------------------------------------- -1: Number of grids (including output grid) = 2 -1: -1: -1: Extension for grid 1 is --> uglo_100km -1: -1: Grid Particulars are : -1: Dimensions = 45166 1 -1: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Extension for grid 2 is --> ep_10m -1: -1: Grid Particulars are : -1: Dimensions = 511 301 -1: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -1: -1: -1: Preparing interpolation weights for output grid -1: Total number of wet points for interpolation 7439 -1: -1: -1: Variable: Grid Interpolation Map Units: 0.100E+01 -1: -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 301 | | -1: | | -1: | 0 0 | -1: 262 | 0 0 0 0 | -1: | 0 0 0 0 | -1: | 0 0 0 0 | -1: 223 | 0 0 0 | -1: | 0 | -1: | 0 | -1: 184 | | -1: | 0 0 0 | -1: | 0 0 0 0 | -1: 145 | 0 | -1: | | -1: | | -1: 106 | | -1: | | -1: | | -1: 67 | | -1: | 0 | -1: | | -1: 28 | 0 | -1: | | -1: | | -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: -1: -1: Interpolating fields .... -1: -1: Output group 1 -1: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -1: Output group 2 -1: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -1: Output group 3 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 4 -1: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -1: Output group 5 -1: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -1: Output group 6 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 7 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 8 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 9 -1: Output variables skipped -1: Output group 10 -1: Output variables skipped -1: ------------------------------------------------ -1: 1Current vel. -1: 1Wind speed -1: 1Ice concentration -1: 2Wave height -1: 2Mean wave period(+2) -1: 2Mean wave period(+1) -1: 2Peak frequency -1: 2Mean wave dir. a1b1 -1: 2Peak direction -1: 4Part. wave height -1: 4Part. peak period -1: 4Part. mean direction -1: 5Charnock parameter -1: ------------------------------------------------ -1: OUTPUT TIME : 2021/03/25 00:00:00 UTC -1: -1: End of file reached -1: -1: -1: *** End of Grid interpolation Routine *** -1: =============================================== -1: -1: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -1: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -1: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_ep_10m/out_grd.ep_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.ep_10m -1: + cmdfile.2[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032500 36 epacif 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + cmdfile.2[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib2_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grib2_sbs.sh[30]grdID=ep_10m -1: + wave_grib2_sbs.sh[31]GRIDNR=255 -1: + wave_grib2_sbs.sh[32]MODNR=11 -1: + wave_grib2_sbs.sh[33]valid_time=2021032500 -1: + wave_grib2_sbs.sh[34]fhr=36 -1: + wave_grib2_sbs.sh[35]grid_region=epacif -1: + wave_grib2_sbs.sh[36]grid_res=0p16 -1: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551 -1: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_ep_10m -1: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_ep_10m -1: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_ep_10m -1: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_ep_10m -1: ++ wave_grib2_sbs.sh[47]printf %03i 36 -1: + wave_grib2_sbs.sh[47]FH3=036 -1: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_epacif_0p16 -1: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.epacif.0p16.f036.grib2 -1: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f036.grib2 ]] -1: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mod_def.ep_10m ./mod_def.ww3 -1: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.ep_10m ./out_grd.ww3 -1: + wave_grib2_sbs.sh[73]ngrib=1 -1: + wave_grib2_sbs.sh[74]dtgrib=3600 -1: + wave_grib2_sbs.sh[75]tstart='20210325 000000' -1: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 000000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.ep_10m.inp.tmpl -1: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -1: $ WAVEWATCH-III gridded output input file -1: $ ---------------------------------------- -1: 20210325 000000 3600 1 -1: N -1: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -1: $ -1: 20210325 000000 7 11 255 0 0 -1: $ -1: $ end of input file -1: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[88]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.232246 ']' -1: ++ prep_step[4]echo gfs_ww3_grib.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[90]export err=0 -1: + wave_grib2_sbs.sh[90]err=0 -1: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[95]cat grib2_epacif_036.out -1: -1: *** WAVEWATCH III GRIB output postp. *** -1: ============================================== -1: -1: Comment character is '$' -1: -1: Grid name : East Pacific 10 min wave grid -1: -1: LINEIN: -1: 20210325 000000 3600 1 -1: -1: 20210325000000 3600 1 -1: GEN_PRO -99999 -1: -1: Output time data : -1: ----------------------------------------------------- -1: First time : 2021/03/25 00:00:00 UTC -1: Interval : 01:00:00 -1: Number of requests : 1 -1: Fields : Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: Requested output fields not yet available: -1: ----------------------------------------------------- -1: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -1: -1: Successfully requested output fields : -1: ----------------------------------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: -1: Additional GRIB parameters : -1: ----------------------------------------------------- -1: Run time : 2021/03/25 00:00:00 UTC -1: GRIB center ID : 7 -1: GRIB gen. proc. ID : 11 -1: GRIB grid ID : 255 -1: GRIB GDS parameter : 0 -1: Fields in file : -1: -------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: CHOSEN GRID TYPE: : LLRECTILINEAR -1: -1: -1: -1: Generating file -1: ----------------------------------------------------- -1: Data for 2021/03/25 00:00:00 UTC 0H forecast. -1: -1: End of program -1: ========================================= -1: WAVEWATCH III GRIB output -1: -1: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -1: + wave_grib2_sbs.sh[102][[ 36 -gt 0 ]] -1: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '36 hour fcst' -grib gfs.wave.t12z.epacif.0p16.f036.grib2 -1: 1:0:d=2021032312:SPC:surface:36 hour fcst: -1: 2:21875:d=2021032312:DIRC:surface:36 hour fcst: -1: 3:49766:d=2021032312:UOGRD:surface:36 hour fcst: -1: 4:71631:d=2021032312:VOGRD:surface:36 hour fcst: -1: 5:93446:d=2021032312:WIND:surface:36 hour fcst: -1: 6:117909:d=2021032312:WDIR:surface:36 hour fcst: -1: 7:145197:d=2021032312:UGRD:surface:36 hour fcst: -1: 8:169859:d=2021032312:VGRD:surface:36 hour fcst: -1: 9:194355:d=2021032312:ICEC:surface:36 hour fcst: -1: 10:213763:d=2021032312:HTSGW:surface:36 hour fcst: -1: 11:236714:d=2021032312:IMWF:surface:36 hour fcst: -1: 12:259861:d=2021032312:MWSPER:surface:36 hour fcst: -1: 13:283108:d=2021032312:PERPW:surface:36 hour fcst: -1: 14:306680:d=2021032312:WWSDIR:surface:36 hour fcst: -1: 15:334246:d=2021032312:DIRPW:surface:36 hour fcst: -1: 16:362431:d=2021032312:WVHGT:surface:36 hour fcst: -1: 17:385366:d=2021032312:SWELL:1 in sequence:36 hour fcst: -1: 18:406751:d=2021032312:SWELL:2 in sequence:36 hour fcst: -1: 19:427228:d=2021032312:SWELL:3 in sequence:36 hour fcst: -1: 20:446886:d=2021032312:WVPER:surface:36 hour fcst: -1: 21:470512:d=2021032312:SWPER:1 in sequence:36 hour fcst: -1: 22:492337:d=2021032312:SWPER:2 in sequence:36 hour fcst: -1: 23:513155:d=2021032312:SWPER:3 in sequence:36 hour fcst: -1: 24:532856:d=2021032312:WVDIR:surface:36 hour fcst: -1: 25:559153:d=2021032312:SWDIR:1 in sequence:36 hour fcst: -1: 26:583714:d=2021032312:SWDIR:2 in sequence:36 hour fcst: -1: 27:606329:d=2021032312:SWDIR:3 in sequence:36 hour fcst: -1: + wave_grib2_sbs.sh[104]err=0 -1: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.epacif.0p16.f036.grib2 -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f036.grib2 ]] -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f036.grib2.idx ]] -1: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.epacif.0p16.f036.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f036.grib2 -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f036.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f036.grib2 = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f036.grib2 ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f036.grib2 -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f036.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f036.grib2.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f036.grib2.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f036.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f036.grib2 -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.epacif.0p16.f036.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f036.grib2.idx -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f036.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f036.grib2.idx = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f036.grib2.idx ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f036.grib2.idx -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f036.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f036.grib2.idx.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f036.grib2.idx.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f036.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f036.grib2.idx -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.epacif.0p16.f036.grib2 and gfs.wave.t12z.epacif.0p16.f036.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_ep_10m to COM' -1: INFO: Copied gfs.wave.t12z.epacif.0p16.f036.grib2 and gfs.wave.t12z.epacif.0p16.f036.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_ep_10m to COM -1: + wave_grib2_sbs.sh[130][[ ep_10m == '' ]] -1: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -1: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.epacif.0p16.f036.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -1: INFO: gfs.wave.t12z.epacif.0p16.f036.grib2 is global.0p50 or SENDDBN is NO, no alert sent -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + cmdfile.1[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032500 3600. 9999 -0: + cmdfile.1[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grid_interp_sbs.sh[25]grdID=at_10m -0: + wave_grid_interp_sbs.sh[26]valid_time=2021032500 -0: + wave_grid_interp_sbs.sh[27]dt=3600. -0: + wave_grid_interp_sbs.sh[28]nst=9999 -0: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551 -0: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.uglo_100km ./out_grd.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mod_def.uglo_100km ./mod_def.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mod_def.at_10m ./mod_def.at_10m -0: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ]] -0: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m'\''' -0: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m' -0: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[51]weights_found=1 -0: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 000000' -0: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 000000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/at_10m/g ww3_gint.inp.tmpl -0: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -0: $ Input file for interpolation of uglo_100km to at_10m -0: $------------------------------------------------ -0: $ Start Time 3600. NSteps -0: 20210325 000000 3600. 9999 -0: $ Total number of grids -0: 2 -0: $ Grid extensions -0: 'uglo_100km' -0: 'at_10m' -0: $ -0: 0 -0: $ -0: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[70]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.232246 ']' -0: ++ prep_step[4]echo gfs_ww3_gint.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -0: INFO: Executing 'gfs_ww3_gint.x' -0: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[73]cat grid_interp.at_10m.out -0: -0: *** WAVEWATCH III Grid interpolation *** -0: =============================================== -0: -0: Comment character is '$' -0: -0: Time Information : -0: --------------------------------------------- -0: Starting Time : 2021/03/25 00:00:00 UTC -0: Interval (in sec) : 3600.00 -0: Number of requests : 9999 -0: --------------------------------------------- -0: Number of grids (including output grid) = 2 -0: -0: -0: Extension for grid 1 is --> uglo_100km -0: -0: Grid Particulars are : -0: Dimensions = 45166 1 -0: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Extension for grid 2 is --> at_10m -0: -0: Grid Particulars are : -0: Dimensions = 301 331 -0: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -0: -0: -0: Preparing interpolation weights for output grid -0: Total number of wet points for interpolation 29591 -0: -0: -0: Variable: Grid Interpolation Map Units: 0.100E+01 -0: -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 331 | | -0: | | -0: | | -0: 289 | | -0: | | -0: | 0 0 0 | -0: 247 | 0 0 0 0 | -0: | 0 0 0 0 0 | -0: | 0 0 0 0 | -0: 205 | 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 163 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 121 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 79 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 | -0: | 0 0 | -0: 37 | 0 | -0: | | -0: | | -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: -0: -0: Interpolating fields .... -0: -0: Output group 1 -0: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -0: Output group 2 -0: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -0: Output group 3 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 4 -0: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -0: Output group 5 -0: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -0: Output group 6 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 7 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 8 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 9 -0: Output variables skipped -0: Output group 10 -0: Output variables skipped -0: ------------------------------------------------ -0: 1Current vel. -0: 1Wind speed -0: 1Ice concentration -0: 2Wave height -0: 2Mean wave period(+2) -0: 2Mean wave period(+1) -0: 2Peak frequency -0: 2Mean wave dir. a1b1 -0: 2Peak direction -0: 4Part. wave height -0: 4Part. peak period -0: 4Part. mean direction -0: 5Charnock parameter -0: ------------------------------------------------ -0: OUTPUT TIME : 2021/03/25 00:00:00 UTC -0: -0: End of file reached -0: -0: -0: *** End of Grid interpolation Routine *** -0: =============================================== -0: -0: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -0: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -0: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_at_10m/out_grd.at_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.at_10m -0: + cmdfile.1[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032500 36 atlocn 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + cmdfile.1[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib2_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grib2_sbs.sh[30]grdID=at_10m -0: + wave_grib2_sbs.sh[31]GRIDNR=255 -0: + wave_grib2_sbs.sh[32]MODNR=11 -0: + wave_grib2_sbs.sh[33]valid_time=2021032500 -0: + wave_grib2_sbs.sh[34]fhr=36 -0: + wave_grib2_sbs.sh[35]grid_region=atlocn -0: + wave_grib2_sbs.sh[36]grid_res=0p16 -0: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551 -0: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_at_10m -0: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_at_10m -0: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_at_10m -0: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_at_10m -0: ++ wave_grib2_sbs.sh[47]printf %03i 36 -0: + wave_grib2_sbs.sh[47]FH3=036 -0: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_atlocn_0p16 -0: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.atlocn.0p16.f036.grib2 -0: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f036.grib2 ]] -0: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mod_def.at_10m ./mod_def.ww3 -0: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.at_10m ./out_grd.ww3 -0: + wave_grib2_sbs.sh[73]ngrib=1 -0: + wave_grib2_sbs.sh[74]dtgrib=3600 -0: + wave_grib2_sbs.sh[75]tstart='20210325 000000' -0: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 000000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.at_10m.inp.tmpl -0: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -0: $ WAVEWATCH-III gridded output input file -0: $ ---------------------------------------- -0: 20210325 000000 3600 1 -0: N -0: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -0: $ -0: 20210325 000000 7 11 255 0 0 -0: $ -0: $ end of input file -0: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[88]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.232246 ']' -0: ++ prep_step[4]echo gfs_ww3_grib.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[90]export err=0 -0: + wave_grib2_sbs.sh[90]err=0 -0: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[95]cat grib2_atlocn_036.out -0: -0: *** WAVEWATCH III GRIB output postp. *** -0: ============================================== -0: -0: Comment character is '$' -0: -0: Grid name : NW Atlantic 10 min wave grid -0: -0: LINEIN: -0: 20210325 000000 3600 1 -0: -0: 20210325000000 3600 1 -0: GEN_PRO -99999 -0: -0: Output time data : -0: ----------------------------------------------------- -0: First time : 2021/03/25 00:00:00 UTC -0: Interval : 01:00:00 -0: Number of requests : 1 -0: Fields : Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: Requested output fields not yet available: -0: ----------------------------------------------------- -0: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -0: -0: Successfully requested output fields : -0: ----------------------------------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: -0: Additional GRIB parameters : -0: ----------------------------------------------------- -0: Run time : 2021/03/25 00:00:00 UTC -0: GRIB center ID : 7 -0: GRIB gen. proc. ID : 11 -0: GRIB grid ID : 255 -0: GRIB GDS parameter : 0 -0: Fields in file : -0: -------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: CHOSEN GRID TYPE: : LLRECTILINEAR -0: -0: -0: -0: Generating file -0: ----------------------------------------------------- -0: Data for 2021/03/25 00:00:00 UTC 0H forecast. -0: -0: End of program -0: ========================================= -0: WAVEWATCH III GRIB output -0: -0: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -0: + wave_grib2_sbs.sh[102][[ 36 -gt 0 ]] -0: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '36 hour fcst' -grib gfs.wave.t12z.atlocn.0p16.f036.grib2 -0: 1:0:d=2021032312:SPC:surface:36 hour fcst: -0: 2:19990:d=2021032312:DIRC:surface:36 hour fcst: -0: 3:59229:d=2021032312:UOGRD:surface:36 hour fcst: -0: 4:79165:d=2021032312:VOGRD:surface:36 hour fcst: -0: 5:99132:d=2021032312:WIND:surface:36 hour fcst: -0: 6:129727:d=2021032312:WDIR:surface:36 hour fcst: -0: 7:172499:d=2021032312:UGRD:surface:36 hour fcst: -0: 8:202800:d=2021032312:VGRD:surface:36 hour fcst: -0: 9:233456:d=2021032312:ICEC:surface:36 hour fcst: -0: 10:246091:d=2021032312:HTSGW:surface:36 hour fcst: -0: 11:270352:d=2021032312:IMWF:surface:36 hour fcst: -0: 12:296421:d=2021032312:MWSPER:surface:36 hour fcst: -0: 13:322641:d=2021032312:PERPW:surface:36 hour fcst: -0: 14:350418:d=2021032312:WWSDIR:surface:36 hour fcst: -0: 15:391974:d=2021032312:DIRPW:surface:36 hour fcst: -0: 16:434425:d=2021032312:WVHGT:surface:36 hour fcst: -0: 17:458308:d=2021032312:SWELL:1 in sequence:36 hour fcst: -0: 18:476891:d=2021032312:SWELL:2 in sequence:36 hour fcst: -0: 19:491867:d=2021032312:SWELL:3 in sequence:36 hour fcst: -0: 20:505303:d=2021032312:WVPER:surface:36 hour fcst: -0: 21:532082:d=2021032312:SWPER:1 in sequence:36 hour fcst: -0: 22:552718:d=2021032312:SWPER:2 in sequence:36 hour fcst: -0: 23:569549:d=2021032312:SWPER:3 in sequence:36 hour fcst: -0: 24:583880:d=2021032312:WVDIR:surface:36 hour fcst: -0: 25:622387:d=2021032312:SWDIR:1 in sequence:36 hour fcst: -0: 26:651152:d=2021032312:SWDIR:2 in sequence:36 hour fcst: -0: 27:672820:d=2021032312:SWDIR:3 in sequence:36 hour fcst: -0: + wave_grib2_sbs.sh[104]err=0 -0: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.atlocn.0p16.f036.grib2 -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f036.grib2 ]] -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f036.grib2.idx ]] -0: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.atlocn.0p16.f036.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f036.grib2 -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f036.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f036.grib2 = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f036.grib2 ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f036.grib2 -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f036.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f036.grib2.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f036.grib2.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f036.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f036.grib2 -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.atlocn.0p16.f036.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f036.grib2.idx -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f036.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f036.grib2.idx = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f036.grib2.idx ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f036.grib2.idx -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f036.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f036.grib2.idx.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f036.grib2.idx.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f036.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f036.grib2.idx -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.atlocn.0p16.f036.grib2 and gfs.wave.t12z.atlocn.0p16.f036.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_at_10m to COM' -0: INFO: Copied gfs.wave.t12z.atlocn.0p16.f036.grib2 and gfs.wave.t12z.atlocn.0p16.f036.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_at_10m to COM -0: + wave_grib2_sbs.sh[130][[ at_10m == '' ]] -0: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -0: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.atlocn.0p16.f036.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -0: INFO: gfs.wave.t12z.atlocn.0p16.f036.grib2 is global.0p50 or SENDDBN is NO, no alert sent -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + cmdfile.4[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032500 3600. 9999 -3: + cmdfile.4[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grid_interp_sbs.sh[25]grdID=glo_30m -3: + wave_grid_interp_sbs.sh[26]valid_time=2021032500 -3: + wave_grid_interp_sbs.sh[27]dt=3600. -3: + wave_grid_interp_sbs.sh[28]nst=9999 -3: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551 -3: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.uglo_100km ./out_grd.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mod_def.uglo_100km ./mod_def.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mod_def.glo_30m ./mod_def.glo_30m -3: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ]] -3: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m'\''' -3: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m' -3: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[51]weights_found=1 -3: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 000000' -3: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 000000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/glo_30m/g ww3_gint.inp.tmpl -3: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -3: $ Input file for interpolation of uglo_100km to glo_30m -3: $------------------------------------------------ -3: $ Start Time 3600. NSteps -3: 20210325 000000 3600. 9999 -3: $ Total number of grids -3: 2 -3: $ Grid extensions -3: 'uglo_100km' -3: 'glo_30m' -3: $ -3: 0 -3: $ -3: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[70]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.232246 ']' -3: ++ prep_step[4]echo gfs_ww3_gint.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -3: INFO: Executing 'gfs_ww3_gint.x' -3: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[73]cat grid_interp.glo_30m.out -3: -3: *** WAVEWATCH III Grid interpolation *** -3: =============================================== -3: -3: Comment character is '$' -3: -3: Time Information : -3: --------------------------------------------- -3: Starting Time : 2021/03/25 00:00:00 UTC -3: Interval (in sec) : 3600.00 -3: Number of requests : 9999 -3: --------------------------------------------- -3: Number of grids (including output grid) = 2 -3: -3: -3: Extension for grid 1 is --> uglo_100km -3: -3: Grid Particulars are : -3: Dimensions = 45166 1 -3: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Extension for grid 2 is --> glo_30m -3: -3: Grid Particulars are : -3: Dimensions = 720 336 -3: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -3: -3: -3: Preparing interpolation weights for output grid -3: Total number of wet points for interpolation 167619 -3: -3: -3: Variable: Grid Interpolation Map Units: 0.100E+01 -3: -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 336 | | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 291 | 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 | -3: 246 | 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 201 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 156 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 111 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 66 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 21 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 | -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: -3: -3: Interpolating fields .... -3: -3: Output group 1 -3: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -3: Output group 2 -3: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -3: Output group 3 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 4 -3: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -3: Output group 5 -3: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -3: Output group 6 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 7 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 8 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 9 -3: Output variables skipped -3: Output group 10 -3: Output variables skipped -3: ------------------------------------------------ -3: 1Current vel. -3: 1Wind speed -3: 1Ice concentration -3: 2Wave height -3: 2Mean wave period(+2) -3: 2Mean wave period(+1) -3: 2Peak frequency -3: 2Mean wave dir. a1b1 -3: 2Peak direction -3: 4Part. wave height -3: 4Part. peak period -3: 4Part. mean direction -3: 5Charnock parameter -3: ------------------------------------------------ -3: OUTPUT TIME : 2021/03/25 00:00:00 UTC -3: -3: End of file reached -3: -3: -3: *** End of Grid interpolation Routine *** -3: =============================================== -3: -3: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -3: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -3: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_glo_30m/out_grd.glo_30m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.glo_30m -3: + cmdfile.4[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032500 36 global 0p50 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + cmdfile.4[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib2_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grib2_sbs.sh[30]grdID=glo_30m -3: + wave_grib2_sbs.sh[31]GRIDNR=255 -3: + wave_grib2_sbs.sh[32]MODNR=11 -3: + wave_grib2_sbs.sh[33]valid_time=2021032500 -3: + wave_grib2_sbs.sh[34]fhr=36 -3: + wave_grib2_sbs.sh[35]grid_region=global -3: + wave_grib2_sbs.sh[36]grid_res=0p50 -3: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551 -3: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_glo_30m -3: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_glo_30m -3: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_glo_30m -3: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_glo_30m -3: ++ wave_grib2_sbs.sh[47]printf %03i 36 -3: + wave_grib2_sbs.sh[47]FH3=036 -3: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p50 -3: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p50.f036.grib2 -3: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f036.grib2 ]] -3: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mod_def.glo_30m ./mod_def.ww3 -3: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.glo_30m ./out_grd.ww3 -3: + wave_grib2_sbs.sh[73]ngrib=1 -3: + wave_grib2_sbs.sh[74]dtgrib=3600 -3: + wave_grib2_sbs.sh[75]tstart='20210325 000000' -3: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 000000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.glo_30m.inp.tmpl -3: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -3: $ WAVEWATCH-III gridded output input file -3: $ ---------------------------------------- -3: 20210325 000000 3600 1 -3: N -3: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -3: $ -3: 20210325 000000 7 11 255 0 0 -3: $ -3: $ end of input file -3: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[88]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.232246 ']' -3: ++ prep_step[4]echo gfs_ww3_grib.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[90]export err=0 -3: + wave_grib2_sbs.sh[90]err=0 -3: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[95]cat grib2_global_036.out -3: -3: *** WAVEWATCH III GRIB output postp. *** -3: ============================================== -3: -3: Comment character is '$' -3: -3: Grid name : Global 30 min wave grid -3: -3: LINEIN: -3: 20210325 000000 3600 1 -3: -3: 20210325000000 3600 1 -3: GEN_PRO -99999 -3: -3: Output time data : -3: ----------------------------------------------------- -3: First time : 2021/03/25 00:00:00 UTC -3: Interval : 01:00:00 -3: Number of requests : 1 -3: Fields : Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: Requested output fields not yet available: -3: ----------------------------------------------------- -3: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -3: -3: Successfully requested output fields : -3: ----------------------------------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: -3: Additional GRIB parameters : -3: ----------------------------------------------------- -3: Run time : 2021/03/25 00:00:00 UTC -3: GRIB center ID : 7 -3: GRIB gen. proc. ID : 11 -3: GRIB grid ID : 255 -3: GRIB GDS parameter : 0 -3: Fields in file : -3: -------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: CHOSEN GRID TYPE: : LLRECTILINEAR -3: -3: -3: -3: Generating file -3: ----------------------------------------------------- -3: Data for 2021/03/25 00:00:00 UTC 0H forecast. -3: -3: End of program -3: ========================================= -3: WAVEWATCH III GRIB output -3: -3: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -3: + wave_grib2_sbs.sh[102][[ 36 -gt 0 ]] -3: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '36 hour fcst' -grib gfs.wave.t12z.global.0p50.f036.grib2 -3: 1:0:d=2021032312:SPC:surface:36 hour fcst: -3: 2:77844:d=2021032312:DIRC:surface:36 hour fcst: -3: 3:320085:d=2021032312:UOGRD:surface:36 hour fcst: -3: 4:395300:d=2021032312:VOGRD:surface:36 hour fcst: -3: 5:475056:d=2021032312:WIND:surface:36 hour fcst: -3: 6:632316:d=2021032312:WDIR:surface:36 hour fcst: -3: 7:869108:d=2021032312:UGRD:surface:36 hour fcst: -3: 8:1023440:d=2021032312:VGRD:surface:36 hour fcst: -3: 9:1180682:d=2021032312:ICEC:surface:36 hour fcst: -3: 10:1221080:d=2021032312:HTSGW:surface:36 hour fcst: -3: 11:1319784:d=2021032312:IMWF:surface:36 hour fcst: -3: 12:1428203:d=2021032312:MWSPER:surface:36 hour fcst: -3: 13:1538002:d=2021032312:PERPW:surface:36 hour fcst: -3: 14:1659701:d=2021032312:WWSDIR:surface:36 hour fcst: -3: 15:1868778:d=2021032312:DIRPW:surface:36 hour fcst: -3: 16:2089992:d=2021032312:WVHGT:surface:36 hour fcst: -3: 17:2191216:d=2021032312:SWELL:1 in sequence:36 hour fcst: -3: 18:2272122:d=2021032312:SWELL:2 in sequence:36 hour fcst: -3: 19:2321041:d=2021032312:SWELL:3 in sequence:36 hour fcst: -3: 20:2354290:d=2021032312:WVPER:surface:36 hour fcst: -3: 21:2471991:d=2021032312:SWPER:1 in sequence:36 hour fcst: -3: 22:2564577:d=2021032312:SWPER:2 in sequence:36 hour fcst: -3: 23:2621640:d=2021032312:SWPER:3 in sequence:36 hour fcst: -3: 24:2656938:d=2021032312:WVDIR:surface:36 hour fcst: -3: 25:2845902:d=2021032312:SWDIR:1 in sequence:36 hour fcst: -3: 26:2999683:d=2021032312:SWDIR:2 in sequence:36 hour fcst: -3: 27:3082983:d=2021032312:SWDIR:3 in sequence:36 hour fcst: -3: + wave_grib2_sbs.sh[104]err=0 -3: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p50.f036.grib2 -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f036.grib2 ]] -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f036.grib2.idx ]] -3: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p50.f036.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f036.grib2 -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f036.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f036.grib2 = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f036.grib2 ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f036.grib2 -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f036.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f036.grib2.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f036.grib2.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f036.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f036.grib2 -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p50.f036.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f036.grib2.idx -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f036.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f036.grib2.idx = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f036.grib2.idx ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f036.grib2.idx -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f036.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f036.grib2.idx.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f036.grib2.idx.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f036.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f036.grib2.idx -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p50.f036.grib2 and gfs.wave.t12z.global.0p50.f036.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_glo_30m to COM' -3: INFO: Copied gfs.wave.t12z.global.0p50.f036.grib2 and gfs.wave.t12z.global.0p50.f036.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_glo_30m to COM -3: + wave_grib2_sbs.sh[130][[ glo_30m == '' ]] -3: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -3: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p50.f036.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -3: INFO: gfs.wave.t12z.global.0p50.f036.grib2 is global.0p50 or SENDDBN is NO, no alert sent -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + cmdfile.3[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032500 3600. 9999 -2: + cmdfile.3[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grid_interp_sbs.sh[25]grdID=wc_10m -2: + wave_grid_interp_sbs.sh[26]valid_time=2021032500 -2: + wave_grid_interp_sbs.sh[27]dt=3600. -2: + wave_grid_interp_sbs.sh[28]nst=9999 -2: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551 -2: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.uglo_100km ./out_grd.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mod_def.uglo_100km ./mod_def.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mod_def.wc_10m ./mod_def.wc_10m -2: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ]] -2: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m'\''' -2: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m' -2: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[51]weights_found=1 -2: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 000000' -2: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 000000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/wc_10m/g ww3_gint.inp.tmpl -2: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -2: $ Input file for interpolation of uglo_100km to wc_10m -2: $------------------------------------------------ -2: $ Start Time 3600. NSteps -2: 20210325 000000 3600. 9999 -2: $ Total number of grids -2: 2 -2: $ Grid extensions -2: 'uglo_100km' -2: 'wc_10m' -2: $ -2: 0 -2: $ -2: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[70]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.232246 ']' -2: ++ prep_step[4]echo gfs_ww3_gint.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -2: INFO: Executing 'gfs_ww3_gint.x' -2: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[73]cat grid_interp.wc_10m.out -2: -2: *** WAVEWATCH III Grid interpolation *** -2: =============================================== -2: -2: Comment character is '$' -2: -2: Time Information : -2: --------------------------------------------- -2: Starting Time : 2021/03/25 00:00:00 UTC -2: Interval (in sec) : 3600.00 -2: Number of requests : 9999 -2: --------------------------------------------- -2: Number of grids (including output grid) = 2 -2: -2: -2: Extension for grid 1 is --> uglo_100km -2: -2: Grid Particulars are : -2: Dimensions = 45166 1 -2: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Extension for grid 2 is --> wc_10m -2: -2: Grid Particulars are : -2: Dimensions = 241 151 -2: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -2: -2: -2: Preparing interpolation weights for output grid -2: Total number of wet points for interpolation 11044 -2: -2: -2: Variable: Grid Interpolation Map Units: 0.100E+01 -2: -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: +---------------------------------------------------------------------------------------------------------------+ -2: 151 | 0 0 0 0 | -2: | 0 0 0 0 0 | -2: | 0 0 0 0 0 | -2: 130 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 109 | 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 88 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 | -2: 67 | 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 | -2: 46 | 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: 25 | 0 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 0 | -2: | | -2: 4 | | -2: +---------------------------------------------------------------------------------------------------------------+ -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: -2: -2: Interpolating fields .... -2: -2: Output group 1 -2: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -2: Output group 2 -2: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -2: Output group 3 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 4 -2: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -2: Output group 5 -2: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -2: Output group 6 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 7 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 8 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 9 -2: Output variables skipped -2: Output group 10 -2: Output variables skipped -2: ------------------------------------------------ -2: 1Current vel. -2: 1Wind speed -2: 1Ice concentration -2: 2Wave height -2: 2Mean wave period(+2) -2: 2Mean wave period(+1) -2: 2Peak frequency -2: 2Mean wave dir. a1b1 -2: 2Peak direction -2: 4Part. wave height -2: 4Part. peak period -2: 4Part. mean direction -2: 5Charnock parameter -2: ------------------------------------------------ -2: OUTPUT TIME : 2021/03/25 00:00:00 UTC -2: -2: End of file reached -2: -2: -2: *** End of Grid interpolation Routine *** -2: =============================================== -2: -2: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -2: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -2: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_wc_10m/out_grd.wc_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.wc_10m -2: + cmdfile.3[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032500 36 wcoast 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + cmdfile.3[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib2_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grib2_sbs.sh[30]grdID=wc_10m -2: + wave_grib2_sbs.sh[31]GRIDNR=255 -2: + wave_grib2_sbs.sh[32]MODNR=11 -2: + wave_grib2_sbs.sh[33]valid_time=2021032500 -2: + wave_grib2_sbs.sh[34]fhr=36 -2: + wave_grib2_sbs.sh[35]grid_region=wcoast -2: + wave_grib2_sbs.sh[36]grid_res=0p16 -2: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551 -2: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_wc_10m -2: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_wc_10m -2: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_wc_10m -2: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_wc_10m -2: ++ wave_grib2_sbs.sh[47]printf %03i 36 -2: + wave_grib2_sbs.sh[47]FH3=036 -2: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_wcoast_0p16 -2: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.wcoast.0p16.f036.grib2 -2: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f036.grib2 ]] -2: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mod_def.wc_10m ./mod_def.ww3 -2: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.wc_10m ./out_grd.ww3 -2: + wave_grib2_sbs.sh[73]ngrib=1 -2: + wave_grib2_sbs.sh[74]dtgrib=3600 -2: + wave_grib2_sbs.sh[75]tstart='20210325 000000' -2: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 000000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.wc_10m.inp.tmpl -2: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -2: $ WAVEWATCH-III gridded output input file -2: $ ---------------------------------------- -2: 20210325 000000 3600 1 -2: N -2: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -2: $ -2: 20210325 000000 7 11 255 0 0 -2: $ -2: $ end of input file -2: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[88]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.232246 ']' -2: ++ prep_step[4]echo gfs_ww3_grib.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[90]export err=0 -2: + wave_grib2_sbs.sh[90]err=0 -2: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[95]cat grib2_wcoast_036.out -2: -2: *** WAVEWATCH III GRIB output postp. *** -2: ============================================== -2: -2: Comment character is '$' -2: -2: Grid name : West Coast 10 min wave grid -2: -2: LINEIN: -2: 20210325 000000 3600 1 -2: -2: 20210325000000 3600 1 -2: GEN_PRO -99999 -2: -2: Output time data : -2: ----------------------------------------------------- -2: First time : 2021/03/25 00:00:00 UTC -2: Interval : 01:00:00 -2: Number of requests : 1 -2: Fields : Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: Requested output fields not yet available: -2: ----------------------------------------------------- -2: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -2: -2: Successfully requested output fields : -2: ----------------------------------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: -2: Additional GRIB parameters : -2: ----------------------------------------------------- -2: Run time : 2021/03/25 00:00:00 UTC -2: GRIB center ID : 7 -2: GRIB gen. proc. ID : 11 -2: GRIB grid ID : 255 -2: GRIB GDS parameter : 0 -2: Fields in file : -2: -------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: CHOSEN GRID TYPE: : LLRECTILINEAR -2: -2: -2: -2: Generating file -2: ----------------------------------------------------- -2: Data for 2021/03/25 00:00:00 UTC 0H forecast. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: -2: End of program -2: ========================================= -2: WAVEWATCH III GRIB output -2: -2: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -2: + wave_grib2_sbs.sh[102][[ 36 -gt 0 ]] -2: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '36 hour fcst' -grib gfs.wave.t12z.wcoast.0p16.f036.grib2 -2: 1:0:d=2021032312:SPC:surface:36 hour fcst: -2: 2:6967:d=2021032312:DIRC:surface:36 hour fcst: -2: 3:23628:d=2021032312:UOGRD:surface:36 hour fcst: -2: 4:30689:d=2021032312:VOGRD:surface:36 hour fcst: -2: 5:38038:d=2021032312:WIND:surface:36 hour fcst: -2: 6:49416:d=2021032312:WDIR:surface:36 hour fcst: -2: 7:65204:d=2021032312:UGRD:surface:36 hour fcst: -2: 8:76635:d=2021032312:VGRD:surface:36 hour fcst: -2: 9:88083:d=2021032312:ICEC:surface:36 hour fcst: -2: 10:92813:d=2021032312:HTSGW:surface:36 hour fcst: -2: 11:102207:d=2021032312:IMWF:surface:36 hour fcst: -2: 12:111759:d=2021032312:MWSPER:surface:36 hour fcst: -2: 13:121362:d=2021032312:PERPW:surface:36 hour fcst: -2: 14:131284:d=2021032312:WWSDIR:surface:36 hour fcst: -2: 15:146553:d=2021032312:DIRPW:surface:36 hour fcst: -2: 16:162261:d=2021032312:WVHGT:surface:36 hour fcst: -2: 17:171478:d=2021032312:SWELL:1 in sequence:36 hour fcst: -2: 18:177466:d=2021032312:SWELL:2 in sequence:36 hour fcst: -2: 19:182196:d=2021032312:SWELL:3 in sequence:36 hour fcst: -2: 20:186926:d=2021032312:WVPER:surface:36 hour fcst: -2: 21:196558:d=2021032312:SWPER:1 in sequence:36 hour fcst: -2: 22:202701:d=2021032312:SWPER:2 in sequence:36 hour fcst: -2: 23:207431:d=2021032312:SWPER:3 in sequence:36 hour fcst: -2: 24:212161:d=2021032312:WVDIR:surface:36 hour fcst: -2: 25:227034:d=2021032312:SWDIR:1 in sequence:36 hour fcst: -2: 26:234547:d=2021032312:SWDIR:2 in sequence:36 hour fcst: -2: 27:239277:d=2021032312:SWDIR:3 in sequence:36 hour fcst: -2: + wave_grib2_sbs.sh[104]err=0 -2: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.wcoast.0p16.f036.grib2 -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f036.grib2 ]] -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f036.grib2.idx ]] -2: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.wcoast.0p16.f036.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f036.grib2 -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f036.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f036.grib2 = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f036.grib2 ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f036.grib2 -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f036.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f036.grib2.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f036.grib2.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f036.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f036.grib2 -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.wcoast.0p16.f036.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f036.grib2.idx -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f036.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f036.grib2.idx = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f036.grib2.idx ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f036.grib2.idx -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f036.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f036.grib2.idx.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f036.grib2.idx.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f036.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f036.grib2.idx -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.wcoast.0p16.f036.grib2 and gfs.wave.t12z.wcoast.0p16.f036.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_wc_10m to COM' -2: INFO: Copied gfs.wave.t12z.wcoast.0p16.f036.grib2 and gfs.wave.t12z.wcoast.0p16.f036.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_wc_10m to COM -2: + wave_grib2_sbs.sh[130][[ wc_10m == '' ]] -2: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -2: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.wcoast.0p16.f036.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -2: INFO: gfs.wave.t12z.wcoast.0p16.f036.grib2 is global.0p50 or SENDDBN is NO, no alert sent -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + cmdfile.5[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032500 3600. 9999 -4: + cmdfile.5[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grid_interp_sbs.sh[25]grdID=aoc_9km -4: + wave_grid_interp_sbs.sh[26]valid_time=2021032500 -4: + wave_grid_interp_sbs.sh[27]dt=3600. -4: + wave_grid_interp_sbs.sh[28]nst=9999 -4: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551 -4: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.uglo_100km ./out_grd.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mod_def.uglo_100km ./mod_def.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mod_def.aoc_9km ./mod_def.aoc_9km -4: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km'\''' -4: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km' -4: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[51]weights_found=1 -4: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 000000' -4: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 000000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/aoc_9km/g ww3_gint.inp.tmpl -4: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -4: $ Input file for interpolation of uglo_100km to aoc_9km -4: $------------------------------------------------ -4: $ Start Time 3600. NSteps -4: 20210325 000000 3600. 9999 -4: $ Total number of grids -4: 2 -4: $ Grid extensions -4: 'uglo_100km' -4: 'aoc_9km' -4: $ -4: 0 -4: $ -4: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[70]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.232246 ']' -4: ++ prep_step[4]echo gfs_ww3_gint.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -4: INFO: Executing 'gfs_ww3_gint.x' -4: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[73]cat grid_interp.aoc_9km.out -4: -4: *** WAVEWATCH III Grid interpolation *** -4: =============================================== -4: -4: Comment character is '$' -4: -4: Time Information : -4: --------------------------------------------- -4: Starting Time : 2021/03/25 00:00:00 UTC -4: Interval (in sec) : 3600.00 -4: Number of requests : 9999 -4: --------------------------------------------- -4: Number of grids (including output grid) = 2 -4: -4: -4: Extension for grid 1 is --> uglo_100km -4: -4: Grid Particulars are : -4: Dimensions = 45166 1 -4: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Extension for grid 2 is --> aoc_9km -4: -4: Grid Particulars are : -4: Dimensions = 1006 1006 -4: Grid Type = 2 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -4: -4: -4: Preparing interpolation weights for output grid -4: Total number of wet points for interpolation 360052 -4: -4: -4: Variable: Grid Interpolation Map Units: 0.100E+01 -4: -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: *** | | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: 880 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 | -4: 754 | 0 0 0 | -4: | 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 628 | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 | -4: 502 | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 376 | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 250 | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 124 | 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 | -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: -4: -4: Interpolating fields .... -4: -4: Output group 1 -4: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -4: Output group 2 -4: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -4: Output group 3 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 4 -4: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -4: Output group 5 -4: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -4: Output group 6 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 7 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 8 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 9 -4: Output variables skipped -4: Output group 10 -4: Output variables skipped -4: ------------------------------------------------ -4: 1Current vel. -4: 1Wind speed -4: 1Ice concentration -4: 2Wave height -4: 2Mean wave period(+2) -4: 2Mean wave period(+1) -4: 2Peak frequency -4: 2Mean wave dir. a1b1 -4: 2Peak direction -4: 4Part. wave height -4: 4Part. peak period -4: 4Part. mean direction -4: 5Charnock parameter -4: ------------------------------------------------ -4: OUTPUT TIME : 2021/03/25 00:00:00 UTC -4: -4: End of file reached -4: -4: -4: *** End of Grid interpolation Routine *** -4: =============================================== -4: -4: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -4: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -4: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_aoc_9km/out_grd.aoc_9km /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.aoc_9km -4: + cmdfile.5[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032500 36 arctic 9km 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + cmdfile.5[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib2_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grib2_sbs.sh[30]grdID=aoc_9km -4: + wave_grib2_sbs.sh[31]GRIDNR=255 -4: + wave_grib2_sbs.sh[32]MODNR=11 -4: + wave_grib2_sbs.sh[33]valid_time=2021032500 -4: + wave_grib2_sbs.sh[34]fhr=36 -4: + wave_grib2_sbs.sh[35]grid_region=arctic -4: + wave_grib2_sbs.sh[36]grid_res=9km -4: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551 -4: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_aoc_9km -4: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_aoc_9km -4: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_aoc_9km -4: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_aoc_9km -4: ++ wave_grib2_sbs.sh[47]printf %03i 36 -4: + wave_grib2_sbs.sh[47]FH3=036 -4: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_arctic_9km -4: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.arctic.9km.f036.grib2 -4: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f036.grib2 ]] -4: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mod_def.aoc_9km ./mod_def.ww3 -4: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.aoc_9km ./out_grd.ww3 -4: + wave_grib2_sbs.sh[73]ngrib=1 -4: + wave_grib2_sbs.sh[74]dtgrib=3600 -4: + wave_grib2_sbs.sh[75]tstart='20210325 000000' -4: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 000000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.aoc_9km.inp.tmpl -4: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -4: $ WAVEWATCH-III gridded output input file -4: $ ---------------------------------------- -4: 20210325 000000 3600 1 -4: N -4: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -4: $ -4: 20210325 000000 7 11 255 0 20 -4: $ -4: 70 0 9.0 9.0 64 -4: $ 60 0 8.64919046313 8.64919046313 64 -4: $ end of input file -4: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[88]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.232246 ']' -4: ++ prep_step[4]echo gfs_ww3_grib.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[90]export err=0 -4: + wave_grib2_sbs.sh[90]err=0 -4: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[95]cat grib2_arctic_036.out -4: -4: *** WAVEWATCH III GRIB output postp. *** -4: ============================================== -4: -4: Comment character is '$' -4: -4: Grid name : Arctic Ocean PolarStereo 9km -4: -4: LINEIN: -4: 20210325 000000 3600 1 -4: -4: 20210325000000 3600 1 -4: GEN_PRO -99999 -4: -4: Output time data : -4: ----------------------------------------------------- -4: First time : 2021/03/25 00:00:00 UTC -4: Interval : 01:00:00 -4: Number of requests : 1 -4: Fields : Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: Requested output fields not yet available: -4: ----------------------------------------------------- -4: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -4: -4: Successfully requested output fields : -4: ----------------------------------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: -4: Additional GRIB parameters : -4: ----------------------------------------------------- -4: Run time : 2021/03/25 00:00:00 UTC -4: GRIB center ID : 7 -4: GRIB gen. proc. ID : 11 -4: GRIB grid ID : 255 -4: GRIB GDS parameter : 0 -4: Fields in file : -4: -------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: CHOSEN GRID TYPE: : POLARSTEREO -4: -4: -4: -4: Generating file -4: ----------------------------------------------------- -4: Data for 2021/03/25 00:00:00 UTC 0H forecast. -4: -4: End of program -4: ========================================= -4: WAVEWATCH III GRIB output -4: -4: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -4: + wave_grib2_sbs.sh[102][[ 36 -gt 0 ]] -4: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '36 hour fcst' -grib gfs.wave.t12z.arctic.9km.f036.grib2 -4: 1:0:d=2021032312:SPC:surface:36 hour fcst: -4: 2:191696:d=2021032312:DIRC:surface:36 hour fcst: -4: 3:662732:d=2021032312:UOGRD:surface:36 hour fcst: -4: 4:850949:d=2021032312:VOGRD:surface:36 hour fcst: -4: 5:1045142:d=2021032312:WIND:surface:36 hour fcst: -4: 6:1344351:d=2021032312:WDIR:surface:36 hour fcst: -4: 7:1767970:d=2021032312:UGRD:surface:36 hour fcst: -4: 8:2064125:d=2021032312:VGRD:surface:36 hour fcst: -4: 9:2361290:d=2021032312:ICEC:surface:36 hour fcst: -4: 10:2532857:d=2021032312:HTSGW:surface:36 hour fcst: -4: 11:2742936:d=2021032312:IMWF:surface:36 hour fcst: -4: 12:2956334:d=2021032312:MWSPER:surface:36 hour fcst: -4: 13:3171508:d=2021032312:PERPW:surface:36 hour fcst: -4: 14:3396330:d=2021032312:WWSDIR:surface:36 hour fcst: -4: 15:3694494:d=2021032312:DIRPW:surface:36 hour fcst: -4: 16:4004383:d=2021032312:WVHGT:surface:36 hour fcst: -4: 17:4211121:d=2021032312:SWELL:1 in sequence:36 hour fcst: -4: 18:4400365:d=2021032312:SWELL:2 in sequence:36 hour fcst: -4: 19:4546838:d=2021032312:SWELL:3 in sequence:36 hour fcst: -4: 20:4676896:d=2021032312:WVPER:surface:36 hour fcst: -4: 21:4894885:d=2021032312:SWPER:1 in sequence:36 hour fcst: -4: 22:5092425:d=2021032312:SWPER:2 in sequence:36 hour fcst: -4: 23:5245482:d=2021032312:SWPER:3 in sequence:36 hour fcst: -4: 24:5377736:d=2021032312:WVDIR:surface:36 hour fcst: -4: 25:5660104:d=2021032312:SWDIR:1 in sequence:36 hour fcst: -4: 26:5923827:d=2021032312:SWDIR:2 in sequence:36 hour fcst: -4: 27:6100889:d=2021032312:SWDIR:3 in sequence:36 hour fcst: -4: + wave_grib2_sbs.sh[104]err=0 -4: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.arctic.9km.f036.grib2 -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f036.grib2 ]] -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f036.grib2.idx ]] -4: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.arctic.9km.f036.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f036.grib2 -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f036.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f036.grib2 = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f036.grib2 ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f036.grib2 -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f036.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f036.grib2.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f036.grib2.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f036.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f036.grib2 -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.arctic.9km.f036.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f036.grib2.idx -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f036.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f036.grib2.idx = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f036.grib2.idx ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f036.grib2.idx -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f036.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f036.grib2.idx.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f036.grib2.idx.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f036.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f036.grib2.idx -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.arctic.9km.f036.grib2 and gfs.wave.t12z.arctic.9km.f036.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_aoc_9km to COM' -4: INFO: Copied gfs.wave.t12z.arctic.9km.f036.grib2 and gfs.wave.t12z.arctic.9km.f036.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_aoc_9km to COM -4: + wave_grib2_sbs.sh[130][[ aoc_9km == '' ]] -4: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -4: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.arctic.9km.f036.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -4: INFO: gfs.wave.t12z.arctic.9km.f036.grib2 is global.0p50 or SENDDBN is NO, no alert sent -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + cmdfile.7[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032500 3600. 9999 -6: + cmdfile.7[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grid_interp_sbs.sh[25]grdID=gsh_15m -6: + wave_grid_interp_sbs.sh[26]valid_time=2021032500 -6: + wave_grid_interp_sbs.sh[27]dt=3600. -6: + wave_grid_interp_sbs.sh[28]nst=9999 -6: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551 -6: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.uglo_100km ./out_grd.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mod_def.uglo_100km ./mod_def.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mod_def.gsh_15m ./mod_def.gsh_15m -6: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m'\''' -6: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m' -6: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[51]weights_found=1 -6: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 000000' -6: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 000000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gsh_15m/g ww3_gint.inp.tmpl -6: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -6: $ Input file for interpolation of uglo_100km to gsh_15m -6: $------------------------------------------------ -6: $ Start Time 3600. NSteps -6: 20210325 000000 3600. 9999 -6: $ Total number of grids -6: 2 -6: $ Grid extensions -6: 'uglo_100km' -6: 'gsh_15m' -6: $ -6: 0 -6: $ -6: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[70]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.232246 ']' -6: ++ prep_step[4]echo gfs_ww3_gint.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -6: INFO: Executing 'gfs_ww3_gint.x' -6: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[73]cat grid_interp.gsh_15m.out -6: -6: *** WAVEWATCH III Grid interpolation *** -6: =============================================== -6: -6: Comment character is '$' -6: -6: Time Information : -6: --------------------------------------------- -6: Starting Time : 2021/03/25 00:00:00 UTC -6: Interval (in sec) : 3600.00 -6: Number of requests : 9999 -6: --------------------------------------------- -6: Number of grids (including output grid) = 2 -6: -6: -6: Extension for grid 1 is --> uglo_100km -6: -6: Grid Particulars are : -6: Dimensions = 45166 1 -6: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Extension for grid 2 is --> gsh_15m -6: -6: Grid Particulars are : -6: Dimensions = 1440 277 -6: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -6: -6: -6: Preparing interpolation weights for output grid -6: Total number of wet points for interpolation 317192 -6: -6: -6: Variable: Grid Interpolation Map Units: 0.100E+01 -6: -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 277 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 241 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 205 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 169 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 133 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 97 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 61 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 | -6: 25 | 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 | -6: | | -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: -6: -6: Interpolating fields .... -6: -6: Output group 1 -6: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -6: Output group 2 -6: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -6: Output group 3 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 4 -6: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -6: Output group 5 -6: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -6: Output group 6 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 7 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 8 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 9 -6: Output variables skipped -6: Output group 10 -6: Output variables skipped -6: ------------------------------------------------ -6: 1Current vel. -6: 1Wind speed -6: 1Ice concentration -6: 2Wave height -6: 2Mean wave period(+2) -6: 2Mean wave period(+1) -6: 2Peak frequency -6: 2Mean wave dir. a1b1 -6: 2Peak direction -6: 4Part. wave height -6: 4Part. peak period -6: 4Part. mean direction -6: 5Charnock parameter -6: ------------------------------------------------ -6: OUTPUT TIME : 2021/03/25 00:00:00 UTC -6: -6: End of file reached -6: -6: -6: *** End of Grid interpolation Routine *** -6: =============================================== -6: -6: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -6: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -6: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_gsh_15m/out_grd.gsh_15m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.gsh_15m -6: + cmdfile.7[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032500 36 gsouth 0p25 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + cmdfile.7[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib2_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grib2_sbs.sh[30]grdID=gsh_15m -6: + wave_grib2_sbs.sh[31]GRIDNR=255 -6: + wave_grib2_sbs.sh[32]MODNR=11 -6: + wave_grib2_sbs.sh[33]valid_time=2021032500 -6: + wave_grib2_sbs.sh[34]fhr=36 -6: + wave_grib2_sbs.sh[35]grid_region=gsouth -6: + wave_grib2_sbs.sh[36]grid_res=0p25 -6: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551 -6: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_gsh_15m -6: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_gsh_15m -6: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_gsh_15m -6: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_gsh_15m -6: ++ wave_grib2_sbs.sh[47]printf %03i 36 -6: + wave_grib2_sbs.sh[47]FH3=036 -6: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -6: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.gsouth.0p25.f036.grib2 -6: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f036.grib2 ]] -6: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mod_def.gsh_15m ./mod_def.ww3 -6: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.gsh_15m ./out_grd.ww3 -6: + wave_grib2_sbs.sh[73]ngrib=1 -6: + wave_grib2_sbs.sh[74]dtgrib=3600 -6: + wave_grib2_sbs.sh[75]tstart='20210325 000000' -6: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 000000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gsh_15m.inp.tmpl -6: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -6: $ WAVEWATCH-III gridded output input file -6: $ ---------------------------------------- -6: 20210325 000000 3600 1 -6: N -6: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -6: $ -6: 20210325 000000 7 11 255 0 0 -6: $ -6: $ end of input file -6: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[88]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.232246 ']' -6: ++ prep_step[4]echo gfs_ww3_grib.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[90]export err=0 -6: + wave_grib2_sbs.sh[90]err=0 -6: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[95]cat grib2_gsouth_036.out -6: -6: *** WAVEWATCH III GRIB output postp. *** -6: ============================================== -6: -6: Comment character is '$' -6: -6: Grid name : GFSv16-wave S Hemisphere 1/4 d -6: -6: LINEIN: -6: 20210325 000000 3600 1 -6: -6: 20210325000000 3600 1 -6: GEN_PRO -99999 -6: -6: Output time data : -6: ----------------------------------------------------- -6: First time : 2021/03/25 00:00:00 UTC -6: Interval : 01:00:00 -6: Number of requests : 1 -6: Fields : Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: Requested output fields not yet available: -6: ----------------------------------------------------- -6: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -6: -6: Successfully requested output fields : -6: ----------------------------------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: -6: Additional GRIB parameters : -6: ----------------------------------------------------- -6: Run time : 2021/03/25 00:00:00 UTC -6: GRIB center ID : 7 -6: GRIB gen. proc. ID : 11 -6: GRIB grid ID : 255 -6: GRIB GDS parameter : 0 -6: Fields in file : -6: -------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: CHOSEN GRID TYPE: : LLRECTILINEAR -6: -6: -6: -6: Generating file -6: ----------------------------------------------------- -6: Data for 2021/03/25 00:00:00 UTC 0H forecast. -6: -6: End of program -6: ========================================= -6: WAVEWATCH III GRIB output -6: -6: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -6: + wave_grib2_sbs.sh[102][[ 36 -gt 0 ]] -6: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '36 hour fcst' -grib gfs.wave.t12z.gsouth.0p25.f036.grib2 -6: 1:0:d=2021032312:SPC:surface:36 hour fcst: -6: 2:118368:d=2021032312:DIRC:surface:36 hour fcst: -6: 3:503507:d=2021032312:UOGRD:surface:36 hour fcst: -6: 4:616867:d=2021032312:VOGRD:surface:36 hour fcst: -6: 5:739531:d=2021032312:WIND:surface:36 hour fcst: -6: 6:983479:d=2021032312:WDIR:surface:36 hour fcst: -6: 7:1364701:d=2021032312:UGRD:surface:36 hour fcst: -6: 8:1601583:d=2021032312:VGRD:surface:36 hour fcst: -6: 9:1844688:d=2021032312:ICEC:surface:36 hour fcst: -6: 10:1905887:d=2021032312:HTSGW:surface:36 hour fcst: -6: 11:2062943:d=2021032312:IMWF:surface:36 hour fcst: -6: 12:2232542:d=2021032312:MWSPER:surface:36 hour fcst: -6: 13:2403830:d=2021032312:PERPW:surface:36 hour fcst: -6: 14:2596727:d=2021032312:WWSDIR:surface:36 hour fcst: -6: 15:2949063:d=2021032312:DIRPW:surface:36 hour fcst: -6: 16:3325113:d=2021032312:WVHGT:surface:36 hour fcst: -6: 17:3493318:d=2021032312:SWELL:1 in sequence:36 hour fcst: -6: 18:3633153:d=2021032312:SWELL:2 in sequence:36 hour fcst: -6: 19:3718043:d=2021032312:SWELL:3 in sequence:36 hour fcst: -6: 20:3773070:d=2021032312:WVPER:surface:36 hour fcst: -6: 21:3967363:d=2021032312:SWPER:1 in sequence:36 hour fcst: -6: 22:4127156:d=2021032312:SWPER:2 in sequence:36 hour fcst: -6: 23:4228363:d=2021032312:SWPER:3 in sequence:36 hour fcst: -6: 24:4287421:d=2021032312:WVDIR:surface:36 hour fcst: -6: 25:4611857:d=2021032312:SWDIR:1 in sequence:36 hour fcst: -6: 26:4896890:d=2021032312:SWDIR:2 in sequence:36 hour fcst: -6: 27:5056028:d=2021032312:SWDIR:3 in sequence:36 hour fcst: -6: + wave_grib2_sbs.sh[104]err=0 -6: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.gsouth.0p25.f036.grib2 -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f036.grib2 ]] -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f036.grib2.idx ]] -6: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.gsouth.0p25.f036.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f036.grib2 -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f036.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f036.grib2 = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f036.grib2 ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f036.grib2 -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f036.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f036.grib2.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f036.grib2.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f036.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f036.grib2 -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.gsouth.0p25.f036.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f036.grib2.idx -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f036.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f036.grib2.idx = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f036.grib2.idx ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f036.grib2.idx -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f036.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f036.grib2.idx.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f036.grib2.idx.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f036.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f036.grib2.idx -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.gsouth.0p25.f036.grib2 and gfs.wave.t12z.gsouth.0p25.f036.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_gsh_15m to COM' -6: INFO: Copied gfs.wave.t12z.gsouth.0p25.f036.grib2 and gfs.wave.t12z.gsouth.0p25.f036.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_gsh_15m to COM -6: + wave_grib2_sbs.sh[130][[ gsh_15m == '' ]] -6: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -6: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.gsouth.0p25.f036.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -6: INFO: gfs.wave.t12z.gsouth.0p25.f036.grib2 is global.0p50 or SENDDBN is NO, no alert sent -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + cmdfile.6[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032500 3600. 9999 -5: + cmdfile.6[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grid_interp_sbs.sh[25]grdID=gnh_10m -5: + wave_grid_interp_sbs.sh[26]valid_time=2021032500 -5: + wave_grid_interp_sbs.sh[27]dt=3600. -5: + wave_grid_interp_sbs.sh[28]nst=9999 -5: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551 -5: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.uglo_100km ./out_grd.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mod_def.uglo_100km ./mod_def.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mod_def.gnh_10m ./mod_def.gnh_10m -5: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m'\''' -5: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m' -5: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[51]weights_found=1 -5: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 000000' -5: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 000000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gnh_10m/g ww3_gint.inp.tmpl -5: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -5: $ Input file for interpolation of uglo_100km to gnh_10m -5: $------------------------------------------------ -5: $ Start Time 3600. NSteps -5: 20210325 000000 3600. 9999 -5: $ Total number of grids -5: 2 -5: $ Grid extensions -5: 'uglo_100km' -5: 'gnh_10m' -5: $ -5: 0 -5: $ -5: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[70]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.232246 ']' -5: ++ prep_step[4]echo gfs_ww3_gint.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -5: INFO: Executing 'gfs_ww3_gint.x' -5: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[73]cat grid_interp.gnh_10m.out -5: -5: *** WAVEWATCH III Grid interpolation *** -5: =============================================== -5: -5: Comment character is '$' -5: -5: Time Information : -5: --------------------------------------------- -5: Starting Time : 2021/03/25 00:00:00 UTC -5: Interval (in sec) : 3600.00 -5: Number of requests : 9999 -5: --------------------------------------------- -5: Number of grids (including output grid) = 2 -5: -5: -5: Extension for grid 1 is --> uglo_100km -5: -5: Grid Particulars are : -5: Dimensions = 45166 1 -5: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Extension for grid 2 is --> gnh_10m -5: -5: Grid Particulars are : -5: Dimensions = 2160 406 -5: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -5: -5: -5: Preparing interpolation weights for output grid -5: Total number of wet points for interpolation 571209 -5: -5: -5: Variable: Grid Interpolation Map Units: 0.100E+01 -5: -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 406 | | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: 355 | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 304 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 253 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 202 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 151 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 100 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 49 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: -5: -5: Interpolating fields .... -5: -5: Output group 1 -5: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -5: Output group 2 -5: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -5: Output group 3 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 4 -5: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -5: Output group 5 -5: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -5: Output group 6 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 7 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 8 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 9 -5: Output variables skipped -5: Output group 10 -5: Output variables skipped -5: ------------------------------------------------ -5: 1Current vel. -5: 1Wind speed -5: 1Ice concentration -5: 2Wave height -5: 2Mean wave period(+2) -5: 2Mean wave period(+1) -5: 2Peak frequency -5: 2Mean wave dir. a1b1 -5: 2Peak direction -5: 4Part. wave height -5: 4Part. peak period -5: 4Part. mean direction -5: 5Charnock parameter -5: ------------------------------------------------ -5: OUTPUT TIME : 2021/03/25 00:00:00 UTC -5: -5: End of file reached -5: -5: -5: *** End of Grid interpolation Routine *** -5: =============================================== -5: -5: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -5: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -5: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grid_interp_gnh_10m/out_grd.gnh_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.gnh_10m -5: + cmdfile.6[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032500 36 global 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + cmdfile.6[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib2_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grib2_sbs.sh[30]grdID=gnh_10m -5: + wave_grib2_sbs.sh[31]GRIDNR=255 -5: + wave_grib2_sbs.sh[32]MODNR=11 -5: + wave_grib2_sbs.sh[33]valid_time=2021032500 -5: + wave_grib2_sbs.sh[34]fhr=36 -5: + wave_grib2_sbs.sh[35]grid_region=global -5: + wave_grib2_sbs.sh[36]grid_res=0p16 -5: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551 -5: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_gnh_10m -5: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_gnh_10m -5: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_gnh_10m -5: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_gnh_10m -5: ++ wave_grib2_sbs.sh[47]printf %03i 36 -5: + wave_grib2_sbs.sh[47]FH3=036 -5: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p16 -5: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p16.f036.grib2 -5: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f036.grib2 ]] -5: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/mod_def.gnh_10m ./mod_def.ww3 -5: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/out_grd.gnh_10m ./out_grd.ww3 -5: + wave_grib2_sbs.sh[73]ngrib=1 -5: + wave_grib2_sbs.sh[74]dtgrib=3600 -5: + wave_grib2_sbs.sh[75]tstart='20210325 000000' -5: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 000000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gnh_10m.inp.tmpl -5: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -5: $ WAVEWATCH-III gridded output input file -5: $ ---------------------------------------- -5: 20210325 000000 3600 1 -5: N -5: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -5: $ -5: 20210325 000000 7 11 255 0 0 -5: $ -5: $ end of input file -5: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[88]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.232246 ']' -5: ++ prep_step[4]echo gfs_ww3_grib.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[90]export err=0 -5: + wave_grib2_sbs.sh[90]err=0 -5: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[95]cat grib2_global_036.out -5: -5: *** WAVEWATCH III GRIB output postp. *** -5: ============================================== -5: -5: Comment character is '$' -5: -5: Grid name : GFSv16-wave N Hemisphere 1/6 d -5: -5: LINEIN: -5: 20210325 000000 3600 1 -5: -5: 20210325000000 3600 1 -5: GEN_PRO -99999 -5: -5: Output time data : -5: ----------------------------------------------------- -5: First time : 2021/03/25 00:00:00 UTC -5: Interval : 01:00:00 -5: Number of requests : 1 -5: Fields : Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: Requested output fields not yet available: -5: ----------------------------------------------------- -5: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -5: -5: Successfully requested output fields : -5: ----------------------------------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: -5: Additional GRIB parameters : -5: ----------------------------------------------------- -5: Run time : 2021/03/25 00:00:00 UTC -5: GRIB center ID : 7 -5: GRIB gen. proc. ID : 11 -5: GRIB grid ID : 255 -5: GRIB GDS parameter : 0 -5: Fields in file : -5: -------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: CHOSEN GRID TYPE: : LLRECTILINEAR -5: -5: -5: -5: Generating file -5: ----------------------------------------------------- -5: Data for 2021/03/25 00:00:00 UTC 0H forecast. -5: -5: End of program -5: ========================================= -5: WAVEWATCH III GRIB output -5: -5: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -5: + wave_grib2_sbs.sh[102][[ 36 -gt 0 ]] -5: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '36 hour fcst' -grib gfs.wave.t12z.global.0p16.f036.grib2 -5: 1:0:d=2021032312:SPC:surface:36 hour fcst: -5: 2:235509:d=2021032312:DIRC:surface:36 hour fcst: -5: 3:860728:d=2021032312:UOGRD:surface:36 hour fcst: -5: 4:1091092:d=2021032312:VOGRD:surface:36 hour fcst: -5: 5:1328984:d=2021032312:WIND:surface:36 hour fcst: -5: 6:1730691:d=2021032312:WDIR:surface:36 hour fcst: -5: 7:2379943:d=2021032312:UGRD:surface:36 hour fcst: -5: 8:2778532:d=2021032312:VGRD:surface:36 hour fcst: -5: 9:3179592:d=2021032312:ICEC:surface:36 hour fcst: -5: 10:3294316:d=2021032312:HTSGW:surface:36 hour fcst: -5: 11:3575950:d=2021032312:IMWF:surface:36 hour fcst: -5: 12:3885003:d=2021032312:MWSPER:surface:36 hour fcst: -5: 13:4196496:d=2021032312:PERPW:surface:36 hour fcst: -5: 14:4533097:d=2021032312:WWSDIR:surface:36 hour fcst: -5: 15:5145946:d=2021032312:DIRPW:surface:36 hour fcst: -5: 16:5800479:d=2021032312:WVHGT:surface:36 hour fcst: -5: 17:6084079:d=2021032312:SWELL:1 in sequence:36 hour fcst: -5: 18:6303206:d=2021032312:SWELL:2 in sequence:36 hour fcst: -5: 19:6450169:d=2021032312:SWELL:3 in sequence:36 hour fcst: -5: 20:6565717:d=2021032312:WVPER:surface:36 hour fcst: -5: 21:6893272:d=2021032312:SWPER:1 in sequence:36 hour fcst: -5: 22:7145360:d=2021032312:SWPER:2 in sequence:36 hour fcst: -5: 23:7312322:d=2021032312:SWPER:3 in sequence:36 hour fcst: -5: 24:7432753:d=2021032312:WVDIR:surface:36 hour fcst: -5: 25:7986386:d=2021032312:SWDIR:1 in sequence:36 hour fcst: -5: 26:8413235:d=2021032312:SWDIR:2 in sequence:36 hour fcst: -5: 27:8653499:d=2021032312:SWDIR:3 in sequence:36 hour fcst: -5: + wave_grib2_sbs.sh[104]err=0 -5: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p16.f036.grib2 -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f036.grib2 ]] -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f036.grib2.idx ]] -5: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p16.f036.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f036.grib2 -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f036.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f036.grib2 = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f036.grib2 ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f036.grib2 -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f036.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f036.grib2.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f036.grib2.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f036.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f036.grib2 -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p16.f036.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f036.grib2.idx -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f036.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f036.grib2.idx = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f036.grib2.idx ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f036.grib2.idx -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f036.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f036.grib2.idx.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f036.grib2.idx.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f036.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f036.grib2.idx -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p16.f036.grib2 and gfs.wave.t12z.global.0p16.f036.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_gnh_10m to COM' -5: INFO: Copied gfs.wave.t12z.global.0p16.f036.grib2 and gfs.wave.t12z.global.0p16.f036.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551/grib_gnh_10m to COM -5: + wave_grib2_sbs.sh[130][[ gnh_10m == '' ]] -5: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -5: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p16.f036.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -5: INFO: gfs.wave.t12z.global.0p16.f036.grib2 is global.0p50 or SENDDBN is NO, no alert sent -+ run_mpmd.sh[113]exit 0 -+ run_mpmd.sh[1]postamble run_mpmd.sh 1753758278 0 -+ preamble.sh[62]set +x -End run_mpmd.sh at 03:04:48 with error code 0 (time elapsed: 00:00:10) -+ exgfs_wave_post_gridded_sbs.sh[122]true -+ exgfs_wave_post_gridded_sbs.sh[123]export err=0 -+ exgfs_wave_post_gridded_sbs.sh[123]err=0 -+ exgfs_wave_post_gridded_sbs.sh[124][[ 0 -ne 0 ]] -+ exgfs_wave_post_gridded_sbs.sh[130]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ exgfs_wave_post_gridded_sbs.sh[131]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ exgfs_wave_post_gridded_sbs.sh[132]gribchk=gfs.wave.t12z.gsouth.0p25.f036.grib2 -+ exgfs_wave_post_gridded_sbs.sh[133][[ ! -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f036.grib2 ]] -+ exgfs_wave_post_gridded_sbs.sh[138]exit 0 -+ JGLOBAL_WAVE_POST_SBS[28]true -+ JGLOBAL_WAVE_POST_SBS[29]export err=0 -+ JGLOBAL_WAVE_POST_SBS[29]err=0 -+ JGLOBAL_WAVE_POST_SBS[30][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[37]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312 -+ JGLOBAL_WAVE_POST_SBS[38][[ NO != \Y\E\S ]] -+ JGLOBAL_WAVE_POST_SBS[39]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f036.231551 -+ JGLOBAL_WAVE_POST_SBS[42]exit 0 -+ JGLOBAL_WAVE_POST_SBS[1]postamble /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS 1753758266 0 -+ preamble.sh[62]set +x -End /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at 03:04:49 with error code 0 (time elapsed: 00:00:23) -Begin /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at Tue Jul 29 03:04:49 UTC 2025 -++ jjob_header.sh[46]OPTIND=1 -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[50]env_job=wavepostsbs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[49]read -ra configs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[61]shift 4 -++ jjob_header.sh[63][[ -z wavepostsbs ]] -++ jjob_header.sh[71]export DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551 -++ jjob_header.sh[71]DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551 -++ jjob_header.sh[72][[ YES == \Y\E\S ]] -++ jjob_header.sh[73]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551 -++ jjob_header.sh[75]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551 -++ jjob_header.sh[76]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551 -++ jjob_header.sh[85]export pid=235235 -++ jjob_header.sh[85]pid=235235 -++ jjob_header.sh[86]export pgmout=OUTPUT.235235 -++ jjob_header.sh[86]pgmout=OUTPUT.235235 -++ jjob_header.sh[87]export pgmerr=errfile -++ jjob_header.sh[87]pgmerr=errfile -++ jjob_header.sh[90]export pgm= -++ jjob_header.sh[90]pgm= -++ jjob_header.sh[96]export cycle=t12z -++ jjob_header.sh[96]cycle=t12z -++ jjob_header.sh[97]setpdy.sh -+ setpdy.sh[20]'[' /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551 == /home/mterry ']' -+ setpdy.sh[25][[ ! t12z =~ t??z ]] -+ setpdy.sh[30]case $# in -+ setpdy.sh[31]dates_before_PDY=7 -+ setpdy.sh[32]dates_after_PDY=7 -+ setpdy.sh[50]COMDATEROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+ setpdy.sh[53]'[' -z 20210323 ']' -+ setpdy.sh[57]sed 's/[0-9]\{8\}/20210323/' /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z -sed: can't read /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z: No such file or directory -++ jjob_header.sh[97]true -++ jjob_header.sh[98]source ./PDY -/work2/noaa/global/mterry/global-workflow_forked/ush/jjob_header.sh: line 98: ./PDY: No such file or directory -++ jjob_header.sh[98]true -++ jjob_header.sh[104]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[104]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.base -+++ config.base[6]echo 'BEGIN: config.base' -BEGIN: config.base -+++ config.base[9]export machine=HERCULES -+++ config.base[9]machine=HERCULES -+++ config.base[12]export RUN_ENVIR=emc -+++ config.base[12]RUN_ENVIR=emc -+++ config.base[15]export ACCOUNT=fv3-cpu -+++ config.base[15]ACCOUNT=fv3-cpu -+++ config.base[16]export QUEUE=batch -+++ config.base[16]QUEUE=batch -+++ config.base[17]export QUEUE_SERVICE=batch -+++ config.base[17]QUEUE_SERVICE=batch -+++ config.base[18]export QUEUE_DTN=batch -+++ config.base[18]QUEUE_DTN=batch -+++ config.base[19]export PARTITION_BATCH=hercules -+++ config.base[19]PARTITION_BATCH=hercules -+++ config.base[20]export PARTITION_SERVICE=service -+++ config.base[20]PARTITION_SERVICE=service -+++ config.base[21]export PARTITION_DTN= -+++ config.base[21]PARTITION_DTN= -+++ config.base[22]export RESERVATION= -+++ config.base[22]RESERVATION= -+++ config.base[23]export CLUSTERS= -+++ config.base[23]CLUSTERS= -+++ config.base[24]export CLUSTERS_SERVICE= -+++ config.base[24]CLUSTERS_SERVICE= -+++ config.base[25]export CLUSTERS_DTN= -+++ config.base[25]CLUSTERS_DTN= -+++ config.base[28]export HPSS_PROJECT=emc-global -+++ config.base[28]HPSS_PROJECT=emc-global -+++ config.base[31]export HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[31]HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[32]export EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[32]EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[33]export FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[33]FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[34]export PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[34]PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[35]export SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[35]SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[36]export USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[36]USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[38]export FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[38]FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[39]export FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[39]FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[40]export FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[40]FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[41]export FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[41]FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[42]export FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[42]FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[43]export FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[43]FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[44]export FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[44]FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[45]export FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[45]FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[50]export PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[50]PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[51]export COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[51]COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[52]export COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[52]COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[53]export DMPDIR=/work/noaa/rstprod/dump -+++ config.base[53]DMPDIR=/work/noaa/rstprod/dump -+++ config.base[57]export COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[57]COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[58]export COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[58]COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[59]export COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[59]COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[62]export HOMEDIR=/work2/noaa/global/mterry -+++ config.base[62]HOMEDIR=/work2/noaa/global/mterry -+++ config.base[63]export STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[63]STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]export PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[65]export NOSCRUB=/work2/noaa/global/mterry -+++ config.base[65]NOSCRUB=/work2/noaa/global/mterry -+++ config.base[68]export BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[68]BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[71]export BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[71]BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[74]export DO_PREP_SFC=NO -+++ config.base[74]DO_PREP_SFC=NO -+++ config.base[77]export DO_GOES=NO -+++ config.base[77]DO_GOES=NO -+++ config.base[78]export DO_BUFRSND=NO -+++ config.base[78]DO_BUFRSND=NO -+++ config.base[79]export DO_GEMPAK=NO -+++ config.base[79]DO_GEMPAK=NO -+++ config.base[80]export DO_AWIPS=NO -+++ config.base[80]DO_AWIPS=NO -+++ config.base[81]export DO_NPOESS=NO -+++ config.base[81]DO_NPOESS=NO -+++ config.base[82]export DO_TRACKER=YES -+++ config.base[82]DO_TRACKER=YES -+++ config.base[83]export DO_GENESIS=YES -+++ config.base[83]DO_GENESIS=YES -+++ config.base[84]export DO_GENESIS_FSU=NO -+++ config.base[84]DO_GENESIS_FSU=NO -+++ config.base[85]export DO_VERFOZN=YES -+++ config.base[85]DO_VERFOZN=YES -+++ config.base[86]export DO_VERFRAD=YES -+++ config.base[86]DO_VERFRAD=YES -+++ config.base[87]export DO_VMINMON=YES -+++ config.base[87]DO_VMINMON=YES -+++ config.base[88]export DO_ANLSTAT=NO -+++ config.base[88]DO_ANLSTAT=NO -+++ config.base[91]export MODE=forecast-only -+++ config.base[91]MODE=forecast-only -+++ config.base[92]export DO_TEST_MODE=YES -+++ config.base[92]DO_TEST_MODE=YES -+++ config.base[101]export FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[101]FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[102]export HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[102]HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[103]export HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[103]HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[106]export NMV=/bin/mv -+++ config.base[106]NMV=/bin/mv -+++ config.base[107]export 'NLN=/bin/ln -sf' -+++ config.base[107]NLN='/bin/ln -sf' -+++ config.base[108]export VERBOSE=YES -+++ config.base[108]VERBOSE=YES -+++ config.base[109]export KEEPDATA=NO -+++ config.base[109]KEEPDATA=NO -+++ config.base[110]export DEBUG_POSTSCRIPT=NO -+++ config.base[110]DEBUG_POSTSCRIPT=NO -+++ config.base[111]export CHGRP_RSTPROD=YES -+++ config.base[111]CHGRP_RSTPROD=YES -+++ config.base[112]export 'CHGRP_CMD=chgrp rstprod' -+++ config.base[112]CHGRP_CMD='chgrp rstprod' -+++ config.base[113]export NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[113]NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[114]export NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[114]NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[117]export BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[117]BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[120]export SDATE=2021032312 -+++ config.base[120]SDATE=2021032312 -+++ config.base[121]export EDATE=2021032312 -+++ config.base[121]EDATE=2021032312 -+++ config.base[122]export EXP_WARM_START=.false. -+++ config.base[122]EXP_WARM_START=.false. -+++ config.base[123]export assim_freq=6 -+++ config.base[123]assim_freq=6 -+++ config.base[124]export PSLOT=C48_S2SW -+++ config.base[124]PSLOT=C48_S2SW -+++ config.base[125]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[125]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[126]export ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[126]ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[127]export DUMP_SUFFIX= -+++ config.base[127]DUMP_SUFFIX= -+++ config.base[128][[ 2021032312 -ge 2019092100 ]] -+++ config.base[128][[ 2021032312 -le 2019110700 ]] -+++ config.base[131]export ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[131]ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[132]export ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[132]ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[133]export FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[133]FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[136]export envir=prod -+++ config.base[136]envir=prod -+++ config.base[137]export NET=gfs -+++ config.base[137]NET=gfs -+++ config.base[138]export RUN=gfs -+++ config.base[138]RUN=gfs -+++ config.base[141]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.com -++++ config.com[4]echo 'BEGIN: config.com' -BEGIN: config.com -++++ config.com[38][[ emc == \n\c\o ]] -++++ config.com[43]COM_OBSPROC_TMPL='${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos' -++++ config.com[44]COM_RTOFS_TMPL='${DMPDIR}' -++++ config.com[45]COM_TCVITAL_TMPL='${DMPDIR}/${RUN}.${YMD}/${HH}/atmos' -++++ config.com[47]declare -rx 'COM_OBS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/obs' -++++ config.com[48]declare -rx COM_OBSPROC_TMPL COM_RTOFS_TMPL -++++ config.com[50]COM_BASE='${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}' -++++ config.com[52]declare -rx 'COM_TOP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}' -++++ config.com[54]declare -rx 'COM_CONF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf' -++++ config.com[55]declare -rx 'COM_OBS_JEDI=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi' -++++ config.com[57]declare -rx 'COM_ATMOS_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input' -++++ config.com[58]declare -rx 'COM_ATMOS_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart' -++++ config.com[59]declare -rx 'COM_ATMOS_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos' -++++ config.com[60]declare -rx 'COM_SNOW_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow' -++++ config.com[61]declare -rx 'COM_SNOW_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon' -++++ config.com[62]declare -rx 'COM_ATMOS_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history' -++++ config.com[63]declare -rx 'COM_ATMOS_MASTER_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master' -++++ config.com[64]declare -rx 'COM_ATMOS_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2' -++++ config.com[65]declare -rx 'COM_ATMOS_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}' -++++ config.com[66]declare -rx 'COM_ATMOS_BUFR_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr' -++++ config.com[67]declare -rx 'COM_ATMOS_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}' -++++ config.com[68]declare -rx 'COM_ATMOS_GENESIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital' -++++ config.com[69]declare -rx 'COM_ATMOS_TRACK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks' -++++ config.com[70]declare -rx 'COM_ATMOS_GOES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim' -++++ config.com[71]declare -rx 'COM_ATMOS_IMAGERY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery' -++++ config.com[72]declare -rx 'COM_ATMOS_OZNMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon' -++++ config.com[73]declare -rx 'COM_ATMOS_RADMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon' -++++ config.com[74]declare -rx 'COM_ATMOS_MINMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon' -++++ config.com[75]declare -rx 'COM_ATMOS_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon' -++++ config.com[76]declare -rx 'COM_ATMOS_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo' -++++ config.com[78]declare -rx 'COM_WAVE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart' -++++ config.com[79]declare -rx 'COM_WAVE_PREP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep' -++++ config.com[80]declare -rx 'COM_WAVE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history' -++++ config.com[81]declare -rx 'COM_WAVE_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded' -++++ config.com[82]declare -rx 'COM_WAVE_GRID_RES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}' -++++ config.com[83]declare -rx 'COM_WAVE_STATION_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station' -++++ config.com[84]declare -rx 'COM_WAVE_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak' -++++ config.com[85]declare -rx 'COM_WAVE_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo' -++++ config.com[87]declare -rx 'COM_OCEAN_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history' -++++ config.com[88]declare -rx 'COM_OCEAN_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart' -++++ config.com[89]declare -rx 'COM_OCEAN_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input' -++++ config.com[90]declare -rx 'COM_OCEAN_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean' -++++ config.com[91]declare -rx 'COM_OCEAN_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon' -++++ config.com[92]declare -rx 'COM_OCEAN_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf' -++++ config.com[93]declare -rx 'COM_OCEAN_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean' -++++ config.com[94]declare -rx 'COM_OCEAN_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf' -++++ config.com[95]declare -rx 'COM_OCEAN_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2' -++++ config.com[96]declare -rx 'COM_OCEAN_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}' -++++ config.com[98]declare -rx 'COM_ICE_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice' -++++ config.com[99]declare -rx 'COM_ICE_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf' -++++ config.com[100]declare -rx 'COM_ICE_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon' -++++ config.com[101]declare -rx 'COM_ICE_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice' -++++ config.com[102]declare -rx 'COM_ICE_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input' -++++ config.com[103]declare -rx 'COM_ICE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history' -++++ config.com[104]declare -rx 'COM_ICE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart' -++++ config.com[105]declare -rx 'COM_ICE_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf' -++++ config.com[106]declare -rx 'COM_ICE_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2' -++++ config.com[107]declare -rx 'COM_ICE_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}' -++++ config.com[109]declare -rx 'COM_CHEM_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history' -++++ config.com[110]declare -rx 'COM_CHEM_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem' -++++ config.com[111]declare -rx 'COM_CHEM_BMAT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix' -++++ config.com[112]declare -rx 'COM_CHEM_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon' -++++ config.com[114]declare -rx 'COM_MED_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart' -+++ config.base[143]export LOGSCRIPT= -+++ config.base[143]LOGSCRIPT= -+++ config.base[145]export 'REDOUT=1>' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wave -+++ config.wave[6]echo 'BEGIN: config.wave' -BEGIN: config.wave -+++ config.wave[13]export RUNRSTwave=gdas -+++ config.wave[13]RUNRSTwave=gdas -+++ config.wave[16]export MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[16]MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[19]case "${waveGRD}" in -+++ config.wave[64]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[64]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[65]export wavepostGRD= -+++ config.wave[65]wavepostGRD= -+++ config.wave[66]export waveuoutpGRD=uglo_100km -+++ config.wave[66]waveuoutpGRD=uglo_100km -+++ config.wave[75]export WAVEWND_DID= -+++ config.wave[75]WAVEWND_DID= -+++ config.wave[76]export WAVEWND_FID= -+++ config.wave[76]WAVEWND_FID= -+++ config.wave[79][[ gfs == \g\f\s ]] -+++ config.wave[80]export FHMAX_WAV=120 -+++ config.wave[80]FHMAX_WAV=120 -+++ config.wave[82]export WAVHINDH=0 -+++ config.wave[82]WAVHINDH=0 -+++ config.wave[83]export FHMAX_WAV_IBP=180 -+++ config.wave[83]FHMAX_WAV_IBP=180 -+++ config.wave[84](( FHMAX_WAV < FHMAX_WAV_IBP )) -+++ config.wave[84]export FHMAX_WAV_IBP=120 -+++ config.wave[84]FHMAX_WAV_IBP=120 -+++ config.wave[87]export DTFLD_WAV=3600 -+++ config.wave[87]DTFLD_WAV=3600 -+++ config.wave[88]export DTPNT_WAV=3600 -+++ config.wave[88]DTPNT_WAV=3600 -+++ config.wave[89]export FHINCP_WAV=1 -+++ config.wave[89]FHINCP_WAV=1 -+++ config.wave[92]export 'OUTPARS_WAV=WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[92]OUTPARS_WAV='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[95][[ gfs == \g\d\a\s ]] -+++ config.wave[99](( INTERVAL_GFS > 0 )) -+++ config.wave[100]export WAVHCYC=6 -+++ config.wave[100]WAVHCYC=6 -+++ config.wave[101]export FHMAX_WAV_CUR=192 -+++ config.wave[101]FHMAX_WAV_CUR=192 -+++ config.wave[109]export RSTTYPE_WAV=T -+++ config.wave[109]RSTTYPE_WAV=T -+++ config.wave[110][[ gfs != gfs ]] -+++ config.wave[115]rst_dt_gfs=43200 -+++ config.wave[116][[ 43200 -gt 0 ]] -+++ config.wave[117]export DT_1_RST_WAV=0 -+++ config.wave[117]DT_1_RST_WAV=0 -+++ config.wave[120]export DT_2_RST_WAV=43200 -+++ config.wave[120]DT_2_RST_WAV=43200 -+++ config.wave[126]export RSTIOFF_WAV=0 -+++ config.wave[126]RSTIOFF_WAV=0 -+++ config.wave[131]export RUNMEM=-1 -+++ config.wave[131]RUNMEM=-1 -+++ config.wave[134](( RUNMEM == -1 )) -+++ config.wave[136]export waveMEMB= -+++ config.wave[136]waveMEMB= -+++ config.wave[143]export WW3ATMINP=CPL -+++ config.wave[143]WW3ATMINP=CPL -+++ config.wave[144][[ YES == \Y\E\S ]] -+++ config.wave[145]export WW3ICEINP=CPL -+++ config.wave[145]WW3ICEINP=CPL -+++ config.wave[146]export WAVEICE_FID= -+++ config.wave[146]WAVEICE_FID= -+++ config.wave[152][[ YES == \Y\E\S ]] -+++ config.wave[153]export WW3CURINP=CPL -+++ config.wave[153]WW3CURINP=CPL -+++ config.wave[154]export WAVECUR_FID= -+++ config.wave[154]WAVECUR_FID= -+++ config.wave[161]export WW3ATMIENS=F -+++ config.wave[161]WW3ATMIENS=F -+++ config.wave[162]export WW3ICEIENS=F -+++ config.wave[162]WW3ICEIENS=F -+++ config.wave[163]export WW3CURIENS=F -+++ config.wave[163]WW3CURIENS=F -+++ config.wave[165]export GOFILETYPE=1 -+++ config.wave[165]GOFILETYPE=1 -+++ config.wave[166]export POFILETYPE=1 -+++ config.wave[166]POFILETYPE=1 -+++ config.wave[170]export FUNIPNT=T -+++ config.wave[170]FUNIPNT=T -+++ config.wave[172]export IOSRV=1 -+++ config.wave[172]IOSRV=1 -+++ config.wave[174]export FPNTPROC=T -+++ config.wave[174]FPNTPROC=T -+++ config.wave[176]export FGRDPROC=F -+++ config.wave[176]FGRDPROC=F -+++ config.wave[178]export FLAGMASKCOMP=F -+++ config.wave[178]FLAGMASKCOMP=F -+++ config.wave[180]export FLAGMASKOUT=F -+++ config.wave[180]FLAGMASKOUT=F -+++ config.wave[182]echo 'END: config.wave' -END: config.wave -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wavepostsbs -+++ config.wavepostsbs[6]echo 'BEGIN: config.wavepostsbs' -BEGIN: config.wavepostsbs -+++ config.wavepostsbs[9]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources wavepostsbs -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=wavepostsbs -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[156]ntasks=8 -++++ config.resources[157]threads_per_task=1 -++++ config.resources[158]tasks_per_node=80 -++++ config.resources[159]NTASKS=8 -++++ config.resources[160]memory=20GB -++++ config.resources[161]walltime=00:15:00 -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export NTASKS -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export memory -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.wavepostsbs[12]export MAX_TASKS=25 -+++ config.wavepostsbs[12]MAX_TASKS=25 -+++ config.wavepostsbs[15]export WAV_SUBGRBSRC= -+++ config.wavepostsbs[15]WAV_SUBGRBSRC= -+++ config.wavepostsbs[16]export WAV_SUBGRB= -+++ config.wavepostsbs[16]WAV_SUBGRB= -+++ config.wavepostsbs[19]export DOFLD_WAV=YES -+++ config.wavepostsbs[19]DOFLD_WAV=YES -+++ config.wavepostsbs[20]export DOPNT_WAV=YES -+++ config.wavepostsbs[20]DOPNT_WAV=YES -+++ config.wavepostsbs[21]export DOGRB_WAV=YES -+++ config.wavepostsbs[21]DOGRB_WAV=YES -+++ config.wavepostsbs[22][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+++ config.wavepostsbs[23]export DOGRI_WAV=YES -+++ config.wavepostsbs[23]DOGRI_WAV=YES -+++ config.wavepostsbs[27]export DOSPC_WAV=YES -+++ config.wavepostsbs[27]DOSPC_WAV=YES -+++ config.wavepostsbs[28]export DOBLL_WAV=YES -+++ config.wavepostsbs[28]DOBLL_WAV=YES -+++ config.wavepostsbs[30]echo 'END: config.wavepostsbs' -END: config.wavepostsbs -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env wavepostsbs -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=wavepostsbs -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 20971520 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 8 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[34]max_threads_per_task=1 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 1 ]] -+++ HERCULES.env[40][[ 1 -gt 1 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 8' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[63]export USE_CFP=YES -+++ HERCULES.env[63]USE_CFP=YES -+++ HERCULES.env[64][[ wavepostsbs == \w\a\v\e\p\r\e\p ]] -+++ HERCULES.env[65]export 'wavempexec=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[65]wavempexec='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[66]export 'wave_mpmd=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[66]wave_mpmd='--multi-prog --output=mpmd.%j.%t.out' -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[4]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ JGLOBAL_WAVE_POST_SBS[7]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[7]HH=12 -+ JGLOBAL_WAVE_POST_SBS[7]declare_from_tmpl -rx COMIN_WAVE_PREP:COM_WAVE_PREP_TMPL COMIN_WAVE_HISTORY:COM_WAVE_HISTORY_TMPL COMOUT_WAVE_GRID:COM_WAVE_GRID_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_WAVE_PREP=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep -declare_from_tmpl :: COMIN_WAVE_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history -declare_from_tmpl :: COMOUT_WAVE_GRID=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded -+ JGLOBAL_WAVE_POST_SBS[12][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n '' ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_atlocn_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_atlocn_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_atlocn_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_epacif_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_epacif_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_epacif_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_wcoast_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_wcoast_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_wcoast_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p50 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p50:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p50=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p50 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=arctic.9km -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_arctic_9km:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_arctic_9km=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_arctic_9km -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_gsouth_0p25:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_gsouth_0p25=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 ]] -+ JGLOBAL_WAVE_POST_SBS[28]/work2/noaa/global/mterry/global-workflow_forked/scripts/exgfs_wave_post_gridded_sbs.sh -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ exgfs_wave_post_gridded_sbs.sh[24]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ exgfs_wave_post_gridded_sbs.sh[26]DOGRI_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[27]DOGRB_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[29]export waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[29]waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[30]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[30]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[31]export wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[31]wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[33]cat - INFO: Grid information: - INFO: Native wave grids: uglo_100km - INFO: Interpolated grids: at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m - INFO: Post-process grids: -++ exgfs_wave_post_gridded_sbs.sh[40]printf %03i 37 -+ exgfs_wave_post_gridded_sbs.sh[40]fhr3=037 -++ exgfs_wave_post_gridded_sbs.sh[41]date -u -d '20210323 12 + 37 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[41]valid_time=2021032501 -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[49]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f037.bin ./out_grd.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f037.bin ./out_grd.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[52][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[62]cat - INFO: Summary: - INFO: Grid interp: DOGRI_WAV="YES" - INFO: Grib files: DOGRB_WAV="YES" - INFO: Fields to be included in grib files: - INFO: OUTPARS_WAV="WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA" -+ exgfs_wave_post_gridded_sbs.sh[70][[ YES == \N\O ]] -+ exgfs_wave_post_gridded_sbs.sh[76]rm -f 'cmdfile.*' cmdfile -+ exgfs_wave_post_gridded_sbs.sh[77]count=0 -+ exgfs_wave_post_gridded_sbs.sh[80][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[81]dt_int=3600. -+ exgfs_wave_post_gridded_sbs.sh[82]n_int=9999 -++ exgfs_wave_post_gridded_sbs.sh[83]date -u -d '20210325 01 - 0 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[83]ymdh_int=2021032501 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=1 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032501 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032501 37 atlocn 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib2_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib2_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=2 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032501 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032501 37 epacif 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib2_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib2_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=3 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032501 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032501 37 wcoast 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib2_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib2_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=4 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032501 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032501 37 global 0p50 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib2_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib2_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=5 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032501 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032501 37 arctic 9km '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib2_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib2_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=6 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032501 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032501 37 global 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib2_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib2_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=7 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032501 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032501 37 gsouth 0p25 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib2_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib2_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[100][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[113][[ 8 -lt 7 ]] -+ exgfs_wave_post_gridded_sbs.sh[121]echo 'INFO: Running MPMD job with 7 commands' -INFO: Running MPMD job with 7 commands -+ exgfs_wave_post_gridded_sbs.sh[122]/work2/noaa/global/mterry/global-workflow_forked/ush/run_mpmd.sh /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/cmdfile -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ run_mpmd.sh[31]source /work2/noaa/global/mterry/global-workflow_forked/ush/preamble.sh -++ preamble.sh[20]set +x -Begin run_mpmd.sh at Tue Jul 29 03:05:00 UTC 2025 -+ run_mpmd.sh[33]cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/cmdfile -+ run_mpmd.sh[36][[ YES != \Y\E\S ]] -+ run_mpmd.sh[46]export OMP_NUM_THREADS=1 -+ run_mpmd.sh[46]OMP_NUM_THREADS=1 -++ run_mpmd.sh[49]wc -l -+ run_mpmd.sh[49]nprocs=7 -+ run_mpmd.sh[52]mpmd_cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mpmd_cmdfile -+ run_mpmd.sh[53][[ -s /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mpmd_cmdfile ]] -+ run_mpmd.sh[55]cat - INFO: Executing MPMD job, STDOUT redirected for each process separately - INFO: On failure, logs for each job will be available in /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mpmd.proc_num.out - INFO: The proc_num corresponds to the line in '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mpmd_cmdfile' -+ run_mpmd.sh[61][[ srun -l --export=ALL --hint=nomultithread =~ ^srun.* ]] -+ run_mpmd.sh[65]nm=0 -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '0 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/cmdfile.1' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '1 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/cmdfile.2' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '2 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/cmdfile.3' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '3 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/cmdfile.4' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '4 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/cmdfile.5' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '5 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/cmdfile.6' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '6 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/cmdfile.7' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[72]set +e -+ run_mpmd.sh[74]srun -l --export=ALL --hint=nomultithread --multi-prog --output=mpmd.%j.%t.out -n 7 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mpmd_cmdfile -+ run_mpmd.sh[75]err=0 -+ run_mpmd.sh[76]set_strict -+ preamble.sh[35][[ YES == \Y\E\S ]] -+ preamble.sh[37]set -eu -+ run_mpmd.sh[103][[ 0 -eq 0 ]] -+ run_mpmd.sh[104]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mpmd_cmdfile -++ run_mpmd.sh[105]find . -name 'mpmd.*.out' -+ run_mpmd.sh[105]out_files='./mpmd.5951789.1.out -./mpmd.5951789.0.out -./mpmd.5951789.3.out -./mpmd.5951789.2.out -./mpmd.5951789.4.out -./mpmd.5951789.6.out -./mpmd.5951789.5.out' -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951789.1.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951789.1.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951789.0.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951789.0.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951789.3.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951789.3.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951789.2.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951789.2.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951789.4.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951789.4.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951789.6.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951789.6.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951789.5.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951789.5.out -+ run_mpmd.sh[110]cat mpmd.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + cmdfile.2[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032501 3600. 9999 -1: + cmdfile.2[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grid_interp_sbs.sh[25]grdID=ep_10m -1: + wave_grid_interp_sbs.sh[26]valid_time=2021032501 -1: + wave_grid_interp_sbs.sh[27]dt=3600. -1: + wave_grid_interp_sbs.sh[28]nst=9999 -1: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551 -1: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.uglo_100km ./out_grd.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mod_def.uglo_100km ./mod_def.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mod_def.ep_10m ./mod_def.ep_10m -1: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ]] -1: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m'\''' -1: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m' -1: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[51]weights_found=1 -1: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 010000' -1: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 010000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/ep_10m/g ww3_gint.inp.tmpl -1: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -1: $ Input file for interpolation of uglo_100km to ep_10m -1: $------------------------------------------------ -1: $ Start Time 3600. NSteps -1: 20210325 010000 3600. 9999 -1: $ Total number of grids -1: 2 -1: $ Grid extensions -1: 'uglo_100km' -1: 'ep_10m' -1: $ -1: 0 -1: $ -1: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[70]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.235235 ']' -1: ++ prep_step[4]echo gfs_ww3_gint.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -1: INFO: Executing 'gfs_ww3_gint.x' -1: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[73]cat grid_interp.ep_10m.out -1: -1: *** WAVEWATCH III Grid interpolation *** -1: =============================================== -1: -1: Comment character is '$' -1: -1: Time Information : -1: --------------------------------------------- -1: Starting Time : 2021/03/25 01:00:00 UTC -1: Interval (in sec) : 3600.00 -1: Number of requests : 9999 -1: --------------------------------------------- -1: Number of grids (including output grid) = 2 -1: -1: -1: Extension for grid 1 is --> uglo_100km -1: -1: Grid Particulars are : -1: Dimensions = 45166 1 -1: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Extension for grid 2 is --> ep_10m -1: -1: Grid Particulars are : -1: Dimensions = 511 301 -1: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -1: -1: -1: Preparing interpolation weights for output grid -1: Total number of wet points for interpolation 7439 -1: -1: -1: Variable: Grid Interpolation Map Units: 0.100E+01 -1: -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 301 | | -1: | | -1: | 0 0 | -1: 262 | 0 0 0 0 | -1: | 0 0 0 0 | -1: | 0 0 0 0 | -1: 223 | 0 0 0 | -1: | 0 | -1: | 0 | -1: 184 | | -1: | 0 0 0 | -1: | 0 0 0 0 | -1: 145 | 0 | -1: | | -1: | | -1: 106 | | -1: | | -1: | | -1: 67 | | -1: | 0 | -1: | | -1: 28 | 0 | -1: | | -1: | | -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: -1: -1: Interpolating fields .... -1: -1: Output group 1 -1: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -1: Output group 2 -1: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -1: Output group 3 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 4 -1: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -1: Output group 5 -1: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -1: Output group 6 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 7 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 8 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 9 -1: Output variables skipped -1: Output group 10 -1: Output variables skipped -1: ------------------------------------------------ -1: 1Current vel. -1: 1Wind speed -1: 1Ice concentration -1: 2Wave height -1: 2Mean wave period(+2) -1: 2Mean wave period(+1) -1: 2Peak frequency -1: 2Mean wave dir. a1b1 -1: 2Peak direction -1: 4Part. wave height -1: 4Part. peak period -1: 4Part. mean direction -1: 5Charnock parameter -1: ------------------------------------------------ -1: OUTPUT TIME : 2021/03/25 01:00:00 UTC -1: -1: End of file reached -1: -1: -1: *** End of Grid interpolation Routine *** -1: =============================================== -1: -1: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -1: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -1: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_ep_10m/out_grd.ep_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.ep_10m -1: + cmdfile.2[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032501 37 epacif 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + cmdfile.2[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib2_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grib2_sbs.sh[30]grdID=ep_10m -1: + wave_grib2_sbs.sh[31]GRIDNR=255 -1: + wave_grib2_sbs.sh[32]MODNR=11 -1: + wave_grib2_sbs.sh[33]valid_time=2021032501 -1: + wave_grib2_sbs.sh[34]fhr=37 -1: + wave_grib2_sbs.sh[35]grid_region=epacif -1: + wave_grib2_sbs.sh[36]grid_res=0p16 -1: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551 -1: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_ep_10m -1: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_ep_10m -1: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_ep_10m -1: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_ep_10m -1: ++ wave_grib2_sbs.sh[47]printf %03i 37 -1: + wave_grib2_sbs.sh[47]FH3=037 -1: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_epacif_0p16 -1: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.epacif.0p16.f037.grib2 -1: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f037.grib2 ]] -1: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mod_def.ep_10m ./mod_def.ww3 -1: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.ep_10m ./out_grd.ww3 -1: + wave_grib2_sbs.sh[73]ngrib=1 -1: + wave_grib2_sbs.sh[74]dtgrib=3600 -1: + wave_grib2_sbs.sh[75]tstart='20210325 010000' -1: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 010000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.ep_10m.inp.tmpl -1: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -1: $ WAVEWATCH-III gridded output input file -1: $ ---------------------------------------- -1: 20210325 010000 3600 1 -1: N -1: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -1: $ -1: 20210325 010000 7 11 255 0 0 -1: $ -1: $ end of input file -1: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[88]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.235235 ']' -1: ++ prep_step[4]echo gfs_ww3_grib.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[90]export err=0 -1: + wave_grib2_sbs.sh[90]err=0 -1: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[95]cat grib2_epacif_037.out -1: -1: *** WAVEWATCH III GRIB output postp. *** -1: ============================================== -1: -1: Comment character is '$' -1: -1: Grid name : East Pacific 10 min wave grid -1: -1: LINEIN: -1: 20210325 010000 3600 1 -1: -1: 20210325010000 3600 1 -1: GEN_PRO -99999 -1: -1: Output time data : -1: ----------------------------------------------------- -1: First time : 2021/03/25 01:00:00 UTC -1: Interval : 01:00:00 -1: Number of requests : 1 -1: Fields : Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: Requested output fields not yet available: -1: ----------------------------------------------------- -1: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -1: -1: Successfully requested output fields : -1: ----------------------------------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: -1: Additional GRIB parameters : -1: ----------------------------------------------------- -1: Run time : 2021/03/25 01:00:00 UTC -1: GRIB center ID : 7 -1: GRIB gen. proc. ID : 11 -1: GRIB grid ID : 255 -1: GRIB GDS parameter : 0 -1: Fields in file : -1: -------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: CHOSEN GRID TYPE: : LLRECTILINEAR -1: -1: -1: -1: Generating file -1: ----------------------------------------------------- -1: Data for 2021/03/25 01:00:00 UTC 0H forecast. -1: Warning: bitmask off everywhere. -1: Pretend one point in jpcpack to avoid crash. -1: Warning: bitmask off everywhere. -1: Pretend one point in jpcpack to avoid crash. -1: Warning: bitmask off everywhere. -1: Pretend one point in jpcpack to avoid crash. -1: -1: End of program -1: ========================================= -1: WAVEWATCH III GRIB output -1: -1: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -1: + wave_grib2_sbs.sh[102][[ 37 -gt 0 ]] -1: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '37 hour fcst' -grib gfs.wave.t12z.epacif.0p16.f037.grib2 -1: 1:0:d=2021032312:SPC:surface:37 hour fcst: -1: 2:21816:d=2021032312:DIRC:surface:37 hour fcst: -1: 3:49624:d=2021032312:UOGRD:surface:37 hour fcst: -1: 4:71540:d=2021032312:VOGRD:surface:37 hour fcst: -1: 5:93364:d=2021032312:WIND:surface:37 hour fcst: -1: 6:117843:d=2021032312:WDIR:surface:37 hour fcst: -1: 7:145196:d=2021032312:UGRD:surface:37 hour fcst: -1: 8:169845:d=2021032312:VGRD:surface:37 hour fcst: -1: 9:194297:d=2021032312:ICEC:surface:37 hour fcst: -1: 10:213705:d=2021032312:HTSGW:surface:37 hour fcst: -1: 11:236603:d=2021032312:IMWF:surface:37 hour fcst: -1: 12:259770:d=2021032312:MWSPER:surface:37 hour fcst: -1: 13:282972:d=2021032312:PERPW:surface:37 hour fcst: -1: 14:306502:d=2021032312:WWSDIR:surface:37 hour fcst: -1: 15:334036:d=2021032312:DIRPW:surface:37 hour fcst: -1: 16:362240:d=2021032312:WVHGT:surface:37 hour fcst: -1: 17:385144:d=2021032312:SWELL:1 in sequence:37 hour fcst: -1: 18:406543:d=2021032312:SWELL:2 in sequence:37 hour fcst: -1: 19:427053:d=2021032312:SWELL:3 in sequence:37 hour fcst: -1: 20:446461:d=2021032312:WVPER:surface:37 hour fcst: -1: 21:470093:d=2021032312:SWPER:1 in sequence:37 hour fcst: -1: 22:491899:d=2021032312:SWPER:2 in sequence:37 hour fcst: -1: 23:512756:d=2021032312:SWPER:3 in sequence:37 hour fcst: -1: 24:532164:d=2021032312:WVDIR:surface:37 hour fcst: -1: 25:558432:d=2021032312:SWDIR:1 in sequence:37 hour fcst: -1: 26:582995:d=2021032312:SWDIR:2 in sequence:37 hour fcst: -1: 27:605600:d=2021032312:SWDIR:3 in sequence:37 hour fcst: -1: + wave_grib2_sbs.sh[104]err=0 -1: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.epacif.0p16.f037.grib2 -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f037.grib2 ]] -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f037.grib2.idx ]] -1: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.epacif.0p16.f037.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f037.grib2 -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f037.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f037.grib2 = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f037.grib2 ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f037.grib2 -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f037.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f037.grib2.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f037.grib2.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f037.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f037.grib2 -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.epacif.0p16.f037.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f037.grib2.idx -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f037.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f037.grib2.idx = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f037.grib2.idx ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f037.grib2.idx -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f037.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f037.grib2.idx.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f037.grib2.idx.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f037.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f037.grib2.idx -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.epacif.0p16.f037.grib2 and gfs.wave.t12z.epacif.0p16.f037.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_ep_10m to COM' -1: INFO: Copied gfs.wave.t12z.epacif.0p16.f037.grib2 and gfs.wave.t12z.epacif.0p16.f037.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_ep_10m to COM -1: + wave_grib2_sbs.sh[130][[ ep_10m == '' ]] -1: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -1: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.epacif.0p16.f037.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -1: INFO: gfs.wave.t12z.epacif.0p16.f037.grib2 is global.0p50 or SENDDBN is NO, no alert sent -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + cmdfile.1[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032501 3600. 9999 -0: + cmdfile.1[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grid_interp_sbs.sh[25]grdID=at_10m -0: + wave_grid_interp_sbs.sh[26]valid_time=2021032501 -0: + wave_grid_interp_sbs.sh[27]dt=3600. -0: + wave_grid_interp_sbs.sh[28]nst=9999 -0: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551 -0: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.uglo_100km ./out_grd.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mod_def.uglo_100km ./mod_def.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mod_def.at_10m ./mod_def.at_10m -0: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ]] -0: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m'\''' -0: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m' -0: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[51]weights_found=1 -0: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 010000' -0: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 010000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/at_10m/g ww3_gint.inp.tmpl -0: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -0: $ Input file for interpolation of uglo_100km to at_10m -0: $------------------------------------------------ -0: $ Start Time 3600. NSteps -0: 20210325 010000 3600. 9999 -0: $ Total number of grids -0: 2 -0: $ Grid extensions -0: 'uglo_100km' -0: 'at_10m' -0: $ -0: 0 -0: $ -0: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[70]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.235235 ']' -0: ++ prep_step[4]echo gfs_ww3_gint.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -0: INFO: Executing 'gfs_ww3_gint.x' -0: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[73]cat grid_interp.at_10m.out -0: -0: *** WAVEWATCH III Grid interpolation *** -0: =============================================== -0: -0: Comment character is '$' -0: -0: Time Information : -0: --------------------------------------------- -0: Starting Time : 2021/03/25 01:00:00 UTC -0: Interval (in sec) : 3600.00 -0: Number of requests : 9999 -0: --------------------------------------------- -0: Number of grids (including output grid) = 2 -0: -0: -0: Extension for grid 1 is --> uglo_100km -0: -0: Grid Particulars are : -0: Dimensions = 45166 1 -0: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Extension for grid 2 is --> at_10m -0: -0: Grid Particulars are : -0: Dimensions = 301 331 -0: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -0: -0: -0: Preparing interpolation weights for output grid -0: Total number of wet points for interpolation 29591 -0: -0: -0: Variable: Grid Interpolation Map Units: 0.100E+01 -0: -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 331 | | -0: | | -0: | | -0: 289 | | -0: | | -0: | 0 0 0 | -0: 247 | 0 0 0 0 | -0: | 0 0 0 0 0 | -0: | 0 0 0 0 | -0: 205 | 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 163 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 121 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 79 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 | -0: | 0 0 | -0: 37 | 0 | -0: | | -0: | | -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: -0: -0: Interpolating fields .... -0: -0: Output group 1 -0: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -0: Output group 2 -0: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -0: Output group 3 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 4 -0: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -0: Output group 5 -0: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -0: Output group 6 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 7 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 8 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 9 -0: Output variables skipped -0: Output group 10 -0: Output variables skipped -0: ------------------------------------------------ -0: 1Current vel. -0: 1Wind speed -0: 1Ice concentration -0: 2Wave height -0: 2Mean wave period(+2) -0: 2Mean wave period(+1) -0: 2Peak frequency -0: 2Mean wave dir. a1b1 -0: 2Peak direction -0: 4Part. wave height -0: 4Part. peak period -0: 4Part. mean direction -0: 5Charnock parameter -0: ------------------------------------------------ -0: OUTPUT TIME : 2021/03/25 01:00:00 UTC -0: -0: End of file reached -0: -0: -0: *** End of Grid interpolation Routine *** -0: =============================================== -0: -0: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -0: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -0: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_at_10m/out_grd.at_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.at_10m -0: + cmdfile.1[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032501 37 atlocn 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + cmdfile.1[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib2_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grib2_sbs.sh[30]grdID=at_10m -0: + wave_grib2_sbs.sh[31]GRIDNR=255 -0: + wave_grib2_sbs.sh[32]MODNR=11 -0: + wave_grib2_sbs.sh[33]valid_time=2021032501 -0: + wave_grib2_sbs.sh[34]fhr=37 -0: + wave_grib2_sbs.sh[35]grid_region=atlocn -0: + wave_grib2_sbs.sh[36]grid_res=0p16 -0: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551 -0: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_at_10m -0: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_at_10m -0: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_at_10m -0: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_at_10m -0: ++ wave_grib2_sbs.sh[47]printf %03i 37 -0: + wave_grib2_sbs.sh[47]FH3=037 -0: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_atlocn_0p16 -0: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.atlocn.0p16.f037.grib2 -0: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f037.grib2 ]] -0: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mod_def.at_10m ./mod_def.ww3 -0: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.at_10m ./out_grd.ww3 -0: + wave_grib2_sbs.sh[73]ngrib=1 -0: + wave_grib2_sbs.sh[74]dtgrib=3600 -0: + wave_grib2_sbs.sh[75]tstart='20210325 010000' -0: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 010000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.at_10m.inp.tmpl -0: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -0: $ WAVEWATCH-III gridded output input file -0: $ ---------------------------------------- -0: 20210325 010000 3600 1 -0: N -0: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -0: $ -0: 20210325 010000 7 11 255 0 0 -0: $ -0: $ end of input file -0: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[88]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.235235 ']' -0: ++ prep_step[4]echo gfs_ww3_grib.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[90]export err=0 -0: + wave_grib2_sbs.sh[90]err=0 -0: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[95]cat grib2_atlocn_037.out -0: -0: *** WAVEWATCH III GRIB output postp. *** -0: ============================================== -0: -0: Comment character is '$' -0: -0: Grid name : NW Atlantic 10 min wave grid -0: -0: LINEIN: -0: 20210325 010000 3600 1 -0: -0: 20210325010000 3600 1 -0: GEN_PRO -99999 -0: -0: Output time data : -0: ----------------------------------------------------- -0: First time : 2021/03/25 01:00:00 UTC -0: Interval : 01:00:00 -0: Number of requests : 1 -0: Fields : Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: Requested output fields not yet available: -0: ----------------------------------------------------- -0: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -0: -0: Successfully requested output fields : -0: ----------------------------------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: -0: Additional GRIB parameters : -0: ----------------------------------------------------- -0: Run time : 2021/03/25 01:00:00 UTC -0: GRIB center ID : 7 -0: GRIB gen. proc. ID : 11 -0: GRIB grid ID : 255 -0: GRIB GDS parameter : 0 -0: Fields in file : -0: -------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: CHOSEN GRID TYPE: : LLRECTILINEAR -0: -0: -0: -0: Generating file -0: ----------------------------------------------------- -0: Data for 2021/03/25 01:00:00 UTC 0H forecast. -0: -0: End of program -0: ========================================= -0: WAVEWATCH III GRIB output -0: -0: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -0: + wave_grib2_sbs.sh[102][[ 37 -gt 0 ]] -0: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '37 hour fcst' -grib gfs.wave.t12z.atlocn.0p16.f037.grib2 -0: 1:0:d=2021032312:SPC:surface:37 hour fcst: -0: 2:20011:d=2021032312:DIRC:surface:37 hour fcst: -0: 3:59206:d=2021032312:UOGRD:surface:37 hour fcst: -0: 4:79192:d=2021032312:VOGRD:surface:37 hour fcst: -0: 5:98989:d=2021032312:WIND:surface:37 hour fcst: -0: 6:129655:d=2021032312:WDIR:surface:37 hour fcst: -0: 7:172364:d=2021032312:UGRD:surface:37 hour fcst: -0: 8:202659:d=2021032312:VGRD:surface:37 hour fcst: -0: 9:233341:d=2021032312:ICEC:surface:37 hour fcst: -0: 10:245976:d=2021032312:HTSGW:surface:37 hour fcst: -0: 11:270225:d=2021032312:IMWF:surface:37 hour fcst: -0: 12:296276:d=2021032312:MWSPER:surface:37 hour fcst: -0: 13:322636:d=2021032312:PERPW:surface:37 hour fcst: -0: 14:350412:d=2021032312:WWSDIR:surface:37 hour fcst: -0: 15:391909:d=2021032312:DIRPW:surface:37 hour fcst: -0: 16:434400:d=2021032312:WVHGT:surface:37 hour fcst: -0: 17:458248:d=2021032312:SWELL:1 in sequence:37 hour fcst: -0: 18:476859:d=2021032312:SWELL:2 in sequence:37 hour fcst: -0: 19:491920:d=2021032312:SWELL:3 in sequence:37 hour fcst: -0: 20:505252:d=2021032312:WVPER:surface:37 hour fcst: -0: 21:532034:d=2021032312:SWPER:1 in sequence:37 hour fcst: -0: 22:552666:d=2021032312:SWPER:2 in sequence:37 hour fcst: -0: 23:569574:d=2021032312:SWPER:3 in sequence:37 hour fcst: -0: 24:583783:d=2021032312:WVDIR:surface:37 hour fcst: -0: 25:622249:d=2021032312:SWDIR:1 in sequence:37 hour fcst: -0: 26:650967:d=2021032312:SWDIR:2 in sequence:37 hour fcst: -0: 27:672915:d=2021032312:SWDIR:3 in sequence:37 hour fcst: -0: + wave_grib2_sbs.sh[104]err=0 -0: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.atlocn.0p16.f037.grib2 -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f037.grib2 ]] -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f037.grib2.idx ]] -0: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.atlocn.0p16.f037.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f037.grib2 -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f037.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f037.grib2 = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f037.grib2 ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f037.grib2 -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f037.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f037.grib2.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f037.grib2.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f037.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f037.grib2 -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.atlocn.0p16.f037.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f037.grib2.idx -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f037.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f037.grib2.idx = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f037.grib2.idx ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f037.grib2.idx -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f037.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f037.grib2.idx.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f037.grib2.idx.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f037.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f037.grib2.idx -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.atlocn.0p16.f037.grib2 and gfs.wave.t12z.atlocn.0p16.f037.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_at_10m to COM' -0: INFO: Copied gfs.wave.t12z.atlocn.0p16.f037.grib2 and gfs.wave.t12z.atlocn.0p16.f037.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_at_10m to COM -0: + wave_grib2_sbs.sh[130][[ at_10m == '' ]] -0: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -0: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.atlocn.0p16.f037.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -0: INFO: gfs.wave.t12z.atlocn.0p16.f037.grib2 is global.0p50 or SENDDBN is NO, no alert sent -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + cmdfile.4[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032501 3600. 9999 -3: + cmdfile.4[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grid_interp_sbs.sh[25]grdID=glo_30m -3: + wave_grid_interp_sbs.sh[26]valid_time=2021032501 -3: + wave_grid_interp_sbs.sh[27]dt=3600. -3: + wave_grid_interp_sbs.sh[28]nst=9999 -3: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551 -3: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.uglo_100km ./out_grd.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mod_def.uglo_100km ./mod_def.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mod_def.glo_30m ./mod_def.glo_30m -3: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ]] -3: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m'\''' -3: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m' -3: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[51]weights_found=1 -3: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 010000' -3: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 010000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/glo_30m/g ww3_gint.inp.tmpl -3: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -3: $ Input file for interpolation of uglo_100km to glo_30m -3: $------------------------------------------------ -3: $ Start Time 3600. NSteps -3: 20210325 010000 3600. 9999 -3: $ Total number of grids -3: 2 -3: $ Grid extensions -3: 'uglo_100km' -3: 'glo_30m' -3: $ -3: 0 -3: $ -3: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[70]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.235235 ']' -3: ++ prep_step[4]echo gfs_ww3_gint.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -3: INFO: Executing 'gfs_ww3_gint.x' -3: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[73]cat grid_interp.glo_30m.out -3: -3: *** WAVEWATCH III Grid interpolation *** -3: =============================================== -3: -3: Comment character is '$' -3: -3: Time Information : -3: --------------------------------------------- -3: Starting Time : 2021/03/25 01:00:00 UTC -3: Interval (in sec) : 3600.00 -3: Number of requests : 9999 -3: --------------------------------------------- -3: Number of grids (including output grid) = 2 -3: -3: -3: Extension for grid 1 is --> uglo_100km -3: -3: Grid Particulars are : -3: Dimensions = 45166 1 -3: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Extension for grid 2 is --> glo_30m -3: -3: Grid Particulars are : -3: Dimensions = 720 336 -3: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -3: -3: -3: Preparing interpolation weights for output grid -3: Total number of wet points for interpolation 167619 -3: -3: -3: Variable: Grid Interpolation Map Units: 0.100E+01 -3: -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 336 | | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 291 | 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 | -3: 246 | 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 201 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 156 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 111 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 66 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 21 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 | -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: -3: -3: Interpolating fields .... -3: -3: Output group 1 -3: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -3: Output group 2 -3: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -3: Output group 3 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 4 -3: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -3: Output group 5 -3: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -3: Output group 6 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 7 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 8 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 9 -3: Output variables skipped -3: Output group 10 -3: Output variables skipped -3: ------------------------------------------------ -3: 1Current vel. -3: 1Wind speed -3: 1Ice concentration -3: 2Wave height -3: 2Mean wave period(+2) -3: 2Mean wave period(+1) -3: 2Peak frequency -3: 2Mean wave dir. a1b1 -3: 2Peak direction -3: 4Part. wave height -3: 4Part. peak period -3: 4Part. mean direction -3: 5Charnock parameter -3: ------------------------------------------------ -3: OUTPUT TIME : 2021/03/25 01:00:00 UTC -3: -3: End of file reached -3: -3: -3: *** End of Grid interpolation Routine *** -3: =============================================== -3: -3: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -3: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -3: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_glo_30m/out_grd.glo_30m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.glo_30m -3: + cmdfile.4[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032501 37 global 0p50 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + cmdfile.4[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib2_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grib2_sbs.sh[30]grdID=glo_30m -3: + wave_grib2_sbs.sh[31]GRIDNR=255 -3: + wave_grib2_sbs.sh[32]MODNR=11 -3: + wave_grib2_sbs.sh[33]valid_time=2021032501 -3: + wave_grib2_sbs.sh[34]fhr=37 -3: + wave_grib2_sbs.sh[35]grid_region=global -3: + wave_grib2_sbs.sh[36]grid_res=0p50 -3: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551 -3: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_glo_30m -3: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_glo_30m -3: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_glo_30m -3: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_glo_30m -3: ++ wave_grib2_sbs.sh[47]printf %03i 37 -3: + wave_grib2_sbs.sh[47]FH3=037 -3: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p50 -3: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p50.f037.grib2 -3: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f037.grib2 ]] -3: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mod_def.glo_30m ./mod_def.ww3 -3: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.glo_30m ./out_grd.ww3 -3: + wave_grib2_sbs.sh[73]ngrib=1 -3: + wave_grib2_sbs.sh[74]dtgrib=3600 -3: + wave_grib2_sbs.sh[75]tstart='20210325 010000' -3: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 010000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.glo_30m.inp.tmpl -3: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -3: $ WAVEWATCH-III gridded output input file -3: $ ---------------------------------------- -3: 20210325 010000 3600 1 -3: N -3: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -3: $ -3: 20210325 010000 7 11 255 0 0 -3: $ -3: $ end of input file -3: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[88]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.235235 ']' -3: ++ prep_step[4]echo gfs_ww3_grib.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[90]export err=0 -3: + wave_grib2_sbs.sh[90]err=0 -3: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[95]cat grib2_global_037.out -3: -3: *** WAVEWATCH III GRIB output postp. *** -3: ============================================== -3: -3: Comment character is '$' -3: -3: Grid name : Global 30 min wave grid -3: -3: LINEIN: -3: 20210325 010000 3600 1 -3: -3: 20210325010000 3600 1 -3: GEN_PRO -99999 -3: -3: Output time data : -3: ----------------------------------------------------- -3: First time : 2021/03/25 01:00:00 UTC -3: Interval : 01:00:00 -3: Number of requests : 1 -3: Fields : Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: Requested output fields not yet available: -3: ----------------------------------------------------- -3: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -3: -3: Successfully requested output fields : -3: ----------------------------------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: -3: Additional GRIB parameters : -3: ----------------------------------------------------- -3: Run time : 2021/03/25 01:00:00 UTC -3: GRIB center ID : 7 -3: GRIB gen. proc. ID : 11 -3: GRIB grid ID : 255 -3: GRIB GDS parameter : 0 -3: Fields in file : -3: -------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: CHOSEN GRID TYPE: : LLRECTILINEAR -3: -3: -3: -3: Generating file -3: ----------------------------------------------------- -3: Data for 2021/03/25 01:00:00 UTC 0H forecast. -3: -3: End of program -3: ========================================= -3: WAVEWATCH III GRIB output -3: -3: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -3: + wave_grib2_sbs.sh[102][[ 37 -gt 0 ]] -3: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '37 hour fcst' -grib gfs.wave.t12z.global.0p50.f037.grib2 -3: 1:0:d=2021032312:SPC:surface:37 hour fcst: -3: 2:77856:d=2021032312:DIRC:surface:37 hour fcst: -3: 3:319809:d=2021032312:UOGRD:surface:37 hour fcst: -3: 4:394963:d=2021032312:VOGRD:surface:37 hour fcst: -3: 5:474667:d=2021032312:WIND:surface:37 hour fcst: -3: 6:631915:d=2021032312:WDIR:surface:37 hour fcst: -3: 7:868612:d=2021032312:UGRD:surface:37 hour fcst: -3: 8:1022933:d=2021032312:VGRD:surface:37 hour fcst: -3: 9:1180106:d=2021032312:ICEC:surface:37 hour fcst: -3: 10:1220491:d=2021032312:HTSGW:surface:37 hour fcst: -3: 11:1319146:d=2021032312:IMWF:surface:37 hour fcst: -3: 12:1427643:d=2021032312:MWSPER:surface:37 hour fcst: -3: 13:1537390:d=2021032312:PERPW:surface:37 hour fcst: -3: 14:1659009:d=2021032312:WWSDIR:surface:37 hour fcst: -3: 15:1867698:d=2021032312:DIRPW:surface:37 hour fcst: -3: 16:2088578:d=2021032312:WVHGT:surface:37 hour fcst: -3: 17:2189909:d=2021032312:SWELL:1 in sequence:37 hour fcst: -3: 18:2271727:d=2021032312:SWELL:2 in sequence:37 hour fcst: -3: 19:2320989:d=2021032312:SWELL:3 in sequence:37 hour fcst: -3: 20:2354395:d=2021032312:WVPER:surface:37 hour fcst: -3: 21:2472016:d=2021032312:SWPER:1 in sequence:37 hour fcst: -3: 22:2565739:d=2021032312:SWPER:2 in sequence:37 hour fcst: -3: 23:2623193:d=2021032312:SWPER:3 in sequence:37 hour fcst: -3: 24:2658761:d=2021032312:WVDIR:surface:37 hour fcst: -3: 25:2847454:d=2021032312:SWDIR:1 in sequence:37 hour fcst: -3: 26:3003330:d=2021032312:SWDIR:2 in sequence:37 hour fcst: -3: 27:3087232:d=2021032312:SWDIR:3 in sequence:37 hour fcst: -3: + wave_grib2_sbs.sh[104]err=0 -3: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p50.f037.grib2 -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f037.grib2 ]] -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f037.grib2.idx ]] -3: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p50.f037.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f037.grib2 -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f037.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f037.grib2 = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f037.grib2 ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f037.grib2 -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f037.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f037.grib2.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f037.grib2.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f037.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f037.grib2 -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p50.f037.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f037.grib2.idx -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f037.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f037.grib2.idx = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f037.grib2.idx ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f037.grib2.idx -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f037.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f037.grib2.idx.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f037.grib2.idx.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f037.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f037.grib2.idx -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p50.f037.grib2 and gfs.wave.t12z.global.0p50.f037.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_glo_30m to COM' -3: INFO: Copied gfs.wave.t12z.global.0p50.f037.grib2 and gfs.wave.t12z.global.0p50.f037.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_glo_30m to COM -3: + wave_grib2_sbs.sh[130][[ glo_30m == '' ]] -3: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -3: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p50.f037.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -3: INFO: gfs.wave.t12z.global.0p50.f037.grib2 is global.0p50 or SENDDBN is NO, no alert sent -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + cmdfile.3[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032501 3600. 9999 -2: + cmdfile.3[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grid_interp_sbs.sh[25]grdID=wc_10m -2: + wave_grid_interp_sbs.sh[26]valid_time=2021032501 -2: + wave_grid_interp_sbs.sh[27]dt=3600. -2: + wave_grid_interp_sbs.sh[28]nst=9999 -2: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551 -2: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.uglo_100km ./out_grd.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mod_def.uglo_100km ./mod_def.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mod_def.wc_10m ./mod_def.wc_10m -2: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ]] -2: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m'\''' -2: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m' -2: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[51]weights_found=1 -2: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 010000' -2: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 010000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/wc_10m/g ww3_gint.inp.tmpl -2: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -2: $ Input file for interpolation of uglo_100km to wc_10m -2: $------------------------------------------------ -2: $ Start Time 3600. NSteps -2: 20210325 010000 3600. 9999 -2: $ Total number of grids -2: 2 -2: $ Grid extensions -2: 'uglo_100km' -2: 'wc_10m' -2: $ -2: 0 -2: $ -2: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[70]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.235235 ']' -2: ++ prep_step[4]echo gfs_ww3_gint.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -2: INFO: Executing 'gfs_ww3_gint.x' -2: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[73]cat grid_interp.wc_10m.out -2: -2: *** WAVEWATCH III Grid interpolation *** -2: =============================================== -2: -2: Comment character is '$' -2: -2: Time Information : -2: --------------------------------------------- -2: Starting Time : 2021/03/25 01:00:00 UTC -2: Interval (in sec) : 3600.00 -2: Number of requests : 9999 -2: --------------------------------------------- -2: Number of grids (including output grid) = 2 -2: -2: -2: Extension for grid 1 is --> uglo_100km -2: -2: Grid Particulars are : -2: Dimensions = 45166 1 -2: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Extension for grid 2 is --> wc_10m -2: -2: Grid Particulars are : -2: Dimensions = 241 151 -2: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -2: -2: -2: Preparing interpolation weights for output grid -2: Total number of wet points for interpolation 11044 -2: -2: -2: Variable: Grid Interpolation Map Units: 0.100E+01 -2: -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: +---------------------------------------------------------------------------------------------------------------+ -2: 151 | 0 0 0 0 | -2: | 0 0 0 0 0 | -2: | 0 0 0 0 0 | -2: 130 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 109 | 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 88 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 | -2: 67 | 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 | -2: 46 | 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: 25 | 0 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 0 | -2: | | -2: 4 | | -2: +---------------------------------------------------------------------------------------------------------------+ -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: -2: -2: Interpolating fields .... -2: -2: Output group 1 -2: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -2: Output group 2 -2: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -2: Output group 3 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 4 -2: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -2: Output group 5 -2: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -2: Output group 6 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 7 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 8 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 9 -2: Output variables skipped -2: Output group 10 -2: Output variables skipped -2: ------------------------------------------------ -2: 1Current vel. -2: 1Wind speed -2: 1Ice concentration -2: 2Wave height -2: 2Mean wave period(+2) -2: 2Mean wave period(+1) -2: 2Peak frequency -2: 2Mean wave dir. a1b1 -2: 2Peak direction -2: 4Part. wave height -2: 4Part. peak period -2: 4Part. mean direction -2: 5Charnock parameter -2: ------------------------------------------------ -2: OUTPUT TIME : 2021/03/25 01:00:00 UTC -2: -2: End of file reached -2: -2: -2: *** End of Grid interpolation Routine *** -2: =============================================== -2: -2: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -2: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -2: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_wc_10m/out_grd.wc_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.wc_10m -2: + cmdfile.3[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032501 37 wcoast 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + cmdfile.3[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib2_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grib2_sbs.sh[30]grdID=wc_10m -2: + wave_grib2_sbs.sh[31]GRIDNR=255 -2: + wave_grib2_sbs.sh[32]MODNR=11 -2: + wave_grib2_sbs.sh[33]valid_time=2021032501 -2: + wave_grib2_sbs.sh[34]fhr=37 -2: + wave_grib2_sbs.sh[35]grid_region=wcoast -2: + wave_grib2_sbs.sh[36]grid_res=0p16 -2: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551 -2: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_wc_10m -2: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_wc_10m -2: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_wc_10m -2: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_wc_10m -2: ++ wave_grib2_sbs.sh[47]printf %03i 37 -2: + wave_grib2_sbs.sh[47]FH3=037 -2: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_wcoast_0p16 -2: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.wcoast.0p16.f037.grib2 -2: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f037.grib2 ]] -2: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mod_def.wc_10m ./mod_def.ww3 -2: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.wc_10m ./out_grd.ww3 -2: + wave_grib2_sbs.sh[73]ngrib=1 -2: + wave_grib2_sbs.sh[74]dtgrib=3600 -2: + wave_grib2_sbs.sh[75]tstart='20210325 010000' -2: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 010000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.wc_10m.inp.tmpl -2: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -2: $ WAVEWATCH-III gridded output input file -2: $ ---------------------------------------- -2: 20210325 010000 3600 1 -2: N -2: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -2: $ -2: 20210325 010000 7 11 255 0 0 -2: $ -2: $ end of input file -2: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[88]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.235235 ']' -2: ++ prep_step[4]echo gfs_ww3_grib.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[90]export err=0 -2: + wave_grib2_sbs.sh[90]err=0 -2: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[95]cat grib2_wcoast_037.out -2: -2: *** WAVEWATCH III GRIB output postp. *** -2: ============================================== -2: -2: Comment character is '$' -2: -2: Grid name : West Coast 10 min wave grid -2: -2: LINEIN: -2: 20210325 010000 3600 1 -2: -2: 20210325010000 3600 1 -2: GEN_PRO -99999 -2: -2: Output time data : -2: ----------------------------------------------------- -2: First time : 2021/03/25 01:00:00 UTC -2: Interval : 01:00:00 -2: Number of requests : 1 -2: Fields : Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: Requested output fields not yet available: -2: ----------------------------------------------------- -2: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -2: -2: Successfully requested output fields : -2: ----------------------------------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: -2: Additional GRIB parameters : -2: ----------------------------------------------------- -2: Run time : 2021/03/25 01:00:00 UTC -2: GRIB center ID : 7 -2: GRIB gen. proc. ID : 11 -2: GRIB grid ID : 255 -2: GRIB GDS parameter : 0 -2: Fields in file : -2: -------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: CHOSEN GRID TYPE: : LLRECTILINEAR -2: -2: -2: -2: Generating file -2: ----------------------------------------------------- -2: Data for 2021/03/25 01:00:00 UTC 0H forecast. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: -2: End of program -2: ========================================= -2: WAVEWATCH III GRIB output -2: -2: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -2: + wave_grib2_sbs.sh[102][[ 37 -gt 0 ]] -2: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '37 hour fcst' -grib gfs.wave.t12z.wcoast.0p16.f037.grib2 -2: 1:0:d=2021032312:SPC:surface:37 hour fcst: -2: 2:6917:d=2021032312:DIRC:surface:37 hour fcst: -2: 3:23525:d=2021032312:UOGRD:surface:37 hour fcst: -2: 4:30628:d=2021032312:VOGRD:surface:37 hour fcst: -2: 5:37978:d=2021032312:WIND:surface:37 hour fcst: -2: 6:49345:d=2021032312:WDIR:surface:37 hour fcst: -2: 7:65043:d=2021032312:UGRD:surface:37 hour fcst: -2: 8:76412:d=2021032312:VGRD:surface:37 hour fcst: -2: 9:87915:d=2021032312:ICEC:surface:37 hour fcst: -2: 10:92645:d=2021032312:HTSGW:surface:37 hour fcst: -2: 11:102057:d=2021032312:IMWF:surface:37 hour fcst: -2: 12:111635:d=2021032312:MWSPER:surface:37 hour fcst: -2: 13:121313:d=2021032312:PERPW:surface:37 hour fcst: -2: 14:131195:d=2021032312:WWSDIR:surface:37 hour fcst: -2: 15:146354:d=2021032312:DIRPW:surface:37 hour fcst: -2: 16:162070:d=2021032312:WVHGT:surface:37 hour fcst: -2: 17:171451:d=2021032312:SWELL:1 in sequence:37 hour fcst: -2: 18:177385:d=2021032312:SWELL:2 in sequence:37 hour fcst: -2: 19:182115:d=2021032312:SWELL:3 in sequence:37 hour fcst: -2: 20:186845:d=2021032312:WVPER:surface:37 hour fcst: -2: 21:196630:d=2021032312:SWPER:1 in sequence:37 hour fcst: -2: 22:202667:d=2021032312:SWPER:2 in sequence:37 hour fcst: -2: 23:207397:d=2021032312:SWPER:3 in sequence:37 hour fcst: -2: 24:212127:d=2021032312:WVDIR:surface:37 hour fcst: -2: 25:227140:d=2021032312:SWDIR:1 in sequence:37 hour fcst: -2: 26:234610:d=2021032312:SWDIR:2 in sequence:37 hour fcst: -2: 27:239340:d=2021032312:SWDIR:3 in sequence:37 hour fcst: -2: + wave_grib2_sbs.sh[104]err=0 -2: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.wcoast.0p16.f037.grib2 -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f037.grib2 ]] -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f037.grib2.idx ]] -2: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.wcoast.0p16.f037.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f037.grib2 -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f037.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f037.grib2 = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f037.grib2 ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f037.grib2 -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f037.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f037.grib2.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f037.grib2.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f037.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f037.grib2 -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.wcoast.0p16.f037.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f037.grib2.idx -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f037.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f037.grib2.idx = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f037.grib2.idx ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f037.grib2.idx -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f037.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f037.grib2.idx.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f037.grib2.idx.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f037.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f037.grib2.idx -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.wcoast.0p16.f037.grib2 and gfs.wave.t12z.wcoast.0p16.f037.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_wc_10m to COM' -2: INFO: Copied gfs.wave.t12z.wcoast.0p16.f037.grib2 and gfs.wave.t12z.wcoast.0p16.f037.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_wc_10m to COM -2: + wave_grib2_sbs.sh[130][[ wc_10m == '' ]] -2: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -2: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.wcoast.0p16.f037.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -2: INFO: gfs.wave.t12z.wcoast.0p16.f037.grib2 is global.0p50 or SENDDBN is NO, no alert sent -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + cmdfile.5[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032501 3600. 9999 -4: + cmdfile.5[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grid_interp_sbs.sh[25]grdID=aoc_9km -4: + wave_grid_interp_sbs.sh[26]valid_time=2021032501 -4: + wave_grid_interp_sbs.sh[27]dt=3600. -4: + wave_grid_interp_sbs.sh[28]nst=9999 -4: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551 -4: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.uglo_100km ./out_grd.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mod_def.uglo_100km ./mod_def.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mod_def.aoc_9km ./mod_def.aoc_9km -4: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km'\''' -4: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km' -4: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[51]weights_found=1 -4: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 010000' -4: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 010000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/aoc_9km/g ww3_gint.inp.tmpl -4: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -4: $ Input file for interpolation of uglo_100km to aoc_9km -4: $------------------------------------------------ -4: $ Start Time 3600. NSteps -4: 20210325 010000 3600. 9999 -4: $ Total number of grids -4: 2 -4: $ Grid extensions -4: 'uglo_100km' -4: 'aoc_9km' -4: $ -4: 0 -4: $ -4: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[70]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.235235 ']' -4: ++ prep_step[4]echo gfs_ww3_gint.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -4: INFO: Executing 'gfs_ww3_gint.x' -4: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[73]cat grid_interp.aoc_9km.out -4: -4: *** WAVEWATCH III Grid interpolation *** -4: =============================================== -4: -4: Comment character is '$' -4: -4: Time Information : -4: --------------------------------------------- -4: Starting Time : 2021/03/25 01:00:00 UTC -4: Interval (in sec) : 3600.00 -4: Number of requests : 9999 -4: --------------------------------------------- -4: Number of grids (including output grid) = 2 -4: -4: -4: Extension for grid 1 is --> uglo_100km -4: -4: Grid Particulars are : -4: Dimensions = 45166 1 -4: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Extension for grid 2 is --> aoc_9km -4: -4: Grid Particulars are : -4: Dimensions = 1006 1006 -4: Grid Type = 2 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -4: -4: -4: Preparing interpolation weights for output grid -4: Total number of wet points for interpolation 360052 -4: -4: -4: Variable: Grid Interpolation Map Units: 0.100E+01 -4: -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: *** | | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: 880 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 | -4: 754 | 0 0 0 | -4: | 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 628 | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 | -4: 502 | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 376 | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 250 | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 124 | 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 | -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: -4: -4: Interpolating fields .... -4: -4: Output group 1 -4: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -4: Output group 2 -4: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -4: Output group 3 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 4 -4: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -4: Output group 5 -4: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -4: Output group 6 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 7 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 8 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 9 -4: Output variables skipped -4: Output group 10 -4: Output variables skipped -4: ------------------------------------------------ -4: 1Current vel. -4: 1Wind speed -4: 1Ice concentration -4: 2Wave height -4: 2Mean wave period(+2) -4: 2Mean wave period(+1) -4: 2Peak frequency -4: 2Mean wave dir. a1b1 -4: 2Peak direction -4: 4Part. wave height -4: 4Part. peak period -4: 4Part. mean direction -4: 5Charnock parameter -4: ------------------------------------------------ -4: OUTPUT TIME : 2021/03/25 01:00:00 UTC -4: -4: End of file reached -4: -4: -4: *** End of Grid interpolation Routine *** -4: =============================================== -4: -4: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -4: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -4: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_aoc_9km/out_grd.aoc_9km /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.aoc_9km -4: + cmdfile.5[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032501 37 arctic 9km 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + cmdfile.5[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib2_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grib2_sbs.sh[30]grdID=aoc_9km -4: + wave_grib2_sbs.sh[31]GRIDNR=255 -4: + wave_grib2_sbs.sh[32]MODNR=11 -4: + wave_grib2_sbs.sh[33]valid_time=2021032501 -4: + wave_grib2_sbs.sh[34]fhr=37 -4: + wave_grib2_sbs.sh[35]grid_region=arctic -4: + wave_grib2_sbs.sh[36]grid_res=9km -4: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551 -4: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_aoc_9km -4: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_aoc_9km -4: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_aoc_9km -4: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_aoc_9km -4: ++ wave_grib2_sbs.sh[47]printf %03i 37 -4: + wave_grib2_sbs.sh[47]FH3=037 -4: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_arctic_9km -4: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.arctic.9km.f037.grib2 -4: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f037.grib2 ]] -4: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mod_def.aoc_9km ./mod_def.ww3 -4: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.aoc_9km ./out_grd.ww3 -4: + wave_grib2_sbs.sh[73]ngrib=1 -4: + wave_grib2_sbs.sh[74]dtgrib=3600 -4: + wave_grib2_sbs.sh[75]tstart='20210325 010000' -4: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 010000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.aoc_9km.inp.tmpl -4: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -4: $ WAVEWATCH-III gridded output input file -4: $ ---------------------------------------- -4: 20210325 010000 3600 1 -4: N -4: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -4: $ -4: 20210325 010000 7 11 255 0 20 -4: $ -4: 70 0 9.0 9.0 64 -4: $ 60 0 8.64919046313 8.64919046313 64 -4: $ end of input file -4: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[88]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.235235 ']' -4: ++ prep_step[4]echo gfs_ww3_grib.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[90]export err=0 -4: + wave_grib2_sbs.sh[90]err=0 -4: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[95]cat grib2_arctic_037.out -4: -4: *** WAVEWATCH III GRIB output postp. *** -4: ============================================== -4: -4: Comment character is '$' -4: -4: Grid name : Arctic Ocean PolarStereo 9km -4: -4: LINEIN: -4: 20210325 010000 3600 1 -4: -4: 20210325010000 3600 1 -4: GEN_PRO -99999 -4: -4: Output time data : -4: ----------------------------------------------------- -4: First time : 2021/03/25 01:00:00 UTC -4: Interval : 01:00:00 -4: Number of requests : 1 -4: Fields : Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: Requested output fields not yet available: -4: ----------------------------------------------------- -4: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -4: -4: Successfully requested output fields : -4: ----------------------------------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: -4: Additional GRIB parameters : -4: ----------------------------------------------------- -4: Run time : 2021/03/25 01:00:00 UTC -4: GRIB center ID : 7 -4: GRIB gen. proc. ID : 11 -4: GRIB grid ID : 255 -4: GRIB GDS parameter : 0 -4: Fields in file : -4: -------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: CHOSEN GRID TYPE: : POLARSTEREO -4: -4: -4: -4: Generating file -4: ----------------------------------------------------- -4: Data for 2021/03/25 01:00:00 UTC 0H forecast. -4: -4: End of program -4: ========================================= -4: WAVEWATCH III GRIB output -4: -4: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -4: + wave_grib2_sbs.sh[102][[ 37 -gt 0 ]] -4: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '37 hour fcst' -grib gfs.wave.t12z.arctic.9km.f037.grib2 -4: 1:0:d=2021032312:SPC:surface:37 hour fcst: -4: 2:191357:d=2021032312:DIRC:surface:37 hour fcst: -4: 3:662106:d=2021032312:UOGRD:surface:37 hour fcst: -4: 4:850476:d=2021032312:VOGRD:surface:37 hour fcst: -4: 5:1044595:d=2021032312:WIND:surface:37 hour fcst: -4: 6:1343295:d=2021032312:WDIR:surface:37 hour fcst: -4: 7:1766283:d=2021032312:UGRD:surface:37 hour fcst: -4: 8:2061962:d=2021032312:VGRD:surface:37 hour fcst: -4: 9:2358937:d=2021032312:ICEC:surface:37 hour fcst: -4: 10:2530396:d=2021032312:HTSGW:surface:37 hour fcst: -4: 11:2740183:d=2021032312:IMWF:surface:37 hour fcst: -4: 12:2953569:d=2021032312:MWSPER:surface:37 hour fcst: -4: 13:3168348:d=2021032312:PERPW:surface:37 hour fcst: -4: 14:3393124:d=2021032312:WWSDIR:surface:37 hour fcst: -4: 15:3690900:d=2021032312:DIRPW:surface:37 hour fcst: -4: 16:4000039:d=2021032312:WVHGT:surface:37 hour fcst: -4: 17:4207063:d=2021032312:SWELL:1 in sequence:37 hour fcst: -4: 18:4396527:d=2021032312:SWELL:2 in sequence:37 hour fcst: -4: 19:4542766:d=2021032312:SWELL:3 in sequence:37 hour fcst: -4: 20:4672713:d=2021032312:WVPER:surface:37 hour fcst: -4: 21:4891112:d=2021032312:SWPER:1 in sequence:37 hour fcst: -4: 22:5089305:d=2021032312:SWPER:2 in sequence:37 hour fcst: -4: 23:5242419:d=2021032312:SWPER:3 in sequence:37 hour fcst: -4: 24:5374266:d=2021032312:WVDIR:surface:37 hour fcst: -4: 25:5657146:d=2021032312:SWDIR:1 in sequence:37 hour fcst: -4: 26:5922187:d=2021032312:SWDIR:2 in sequence:37 hour fcst: -4: 27:6099163:d=2021032312:SWDIR:3 in sequence:37 hour fcst: -4: + wave_grib2_sbs.sh[104]err=0 -4: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.arctic.9km.f037.grib2 -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f037.grib2 ]] -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f037.grib2.idx ]] -4: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.arctic.9km.f037.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f037.grib2 -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f037.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f037.grib2 = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f037.grib2 ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f037.grib2 -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f037.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f037.grib2.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f037.grib2.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f037.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f037.grib2 -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.arctic.9km.f037.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f037.grib2.idx -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f037.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f037.grib2.idx = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f037.grib2.idx ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f037.grib2.idx -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f037.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f037.grib2.idx.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f037.grib2.idx.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f037.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f037.grib2.idx -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.arctic.9km.f037.grib2 and gfs.wave.t12z.arctic.9km.f037.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_aoc_9km to COM' -4: INFO: Copied gfs.wave.t12z.arctic.9km.f037.grib2 and gfs.wave.t12z.arctic.9km.f037.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_aoc_9km to COM -4: + wave_grib2_sbs.sh[130][[ aoc_9km == '' ]] -4: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -4: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.arctic.9km.f037.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -4: INFO: gfs.wave.t12z.arctic.9km.f037.grib2 is global.0p50 or SENDDBN is NO, no alert sent -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + cmdfile.7[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032501 3600. 9999 -6: + cmdfile.7[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grid_interp_sbs.sh[25]grdID=gsh_15m -6: + wave_grid_interp_sbs.sh[26]valid_time=2021032501 -6: + wave_grid_interp_sbs.sh[27]dt=3600. -6: + wave_grid_interp_sbs.sh[28]nst=9999 -6: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551 -6: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.uglo_100km ./out_grd.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mod_def.uglo_100km ./mod_def.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mod_def.gsh_15m ./mod_def.gsh_15m -6: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m'\''' -6: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m' -6: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[51]weights_found=1 -6: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 010000' -6: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 010000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gsh_15m/g ww3_gint.inp.tmpl -6: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -6: $ Input file for interpolation of uglo_100km to gsh_15m -6: $------------------------------------------------ -6: $ Start Time 3600. NSteps -6: 20210325 010000 3600. 9999 -6: $ Total number of grids -6: 2 -6: $ Grid extensions -6: 'uglo_100km' -6: 'gsh_15m' -6: $ -6: 0 -6: $ -6: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[70]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.235235 ']' -6: ++ prep_step[4]echo gfs_ww3_gint.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -6: INFO: Executing 'gfs_ww3_gint.x' -6: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[73]cat grid_interp.gsh_15m.out -6: -6: *** WAVEWATCH III Grid interpolation *** -6: =============================================== -6: -6: Comment character is '$' -6: -6: Time Information : -6: --------------------------------------------- -6: Starting Time : 2021/03/25 01:00:00 UTC -6: Interval (in sec) : 3600.00 -6: Number of requests : 9999 -6: --------------------------------------------- -6: Number of grids (including output grid) = 2 -6: -6: -6: Extension for grid 1 is --> uglo_100km -6: -6: Grid Particulars are : -6: Dimensions = 45166 1 -6: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Extension for grid 2 is --> gsh_15m -6: -6: Grid Particulars are : -6: Dimensions = 1440 277 -6: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -6: -6: -6: Preparing interpolation weights for output grid -6: Total number of wet points for interpolation 317192 -6: -6: -6: Variable: Grid Interpolation Map Units: 0.100E+01 -6: -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 277 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 241 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 205 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 169 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 133 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 97 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 61 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 | -6: 25 | 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 | -6: | | -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: -6: -6: Interpolating fields .... -6: -6: Output group 1 -6: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -6: Output group 2 -6: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -6: Output group 3 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 4 -6: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -6: Output group 5 -6: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -6: Output group 6 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 7 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 8 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 9 -6: Output variables skipped -6: Output group 10 -6: Output variables skipped -6: ------------------------------------------------ -6: 1Current vel. -6: 1Wind speed -6: 1Ice concentration -6: 2Wave height -6: 2Mean wave period(+2) -6: 2Mean wave period(+1) -6: 2Peak frequency -6: 2Mean wave dir. a1b1 -6: 2Peak direction -6: 4Part. wave height -6: 4Part. peak period -6: 4Part. mean direction -6: 5Charnock parameter -6: ------------------------------------------------ -6: OUTPUT TIME : 2021/03/25 01:00:00 UTC -6: -6: End of file reached -6: -6: -6: *** End of Grid interpolation Routine *** -6: =============================================== -6: -6: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -6: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -6: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_gsh_15m/out_grd.gsh_15m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.gsh_15m -6: + cmdfile.7[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032501 37 gsouth 0p25 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + cmdfile.7[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib2_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grib2_sbs.sh[30]grdID=gsh_15m -6: + wave_grib2_sbs.sh[31]GRIDNR=255 -6: + wave_grib2_sbs.sh[32]MODNR=11 -6: + wave_grib2_sbs.sh[33]valid_time=2021032501 -6: + wave_grib2_sbs.sh[34]fhr=37 -6: + wave_grib2_sbs.sh[35]grid_region=gsouth -6: + wave_grib2_sbs.sh[36]grid_res=0p25 -6: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551 -6: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_gsh_15m -6: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_gsh_15m -6: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_gsh_15m -6: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_gsh_15m -6: ++ wave_grib2_sbs.sh[47]printf %03i 37 -6: + wave_grib2_sbs.sh[47]FH3=037 -6: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -6: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.gsouth.0p25.f037.grib2 -6: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f037.grib2 ]] -6: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mod_def.gsh_15m ./mod_def.ww3 -6: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.gsh_15m ./out_grd.ww3 -6: + wave_grib2_sbs.sh[73]ngrib=1 -6: + wave_grib2_sbs.sh[74]dtgrib=3600 -6: + wave_grib2_sbs.sh[75]tstart='20210325 010000' -6: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 010000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gsh_15m.inp.tmpl -6: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -6: $ WAVEWATCH-III gridded output input file -6: $ ---------------------------------------- -6: 20210325 010000 3600 1 -6: N -6: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -6: $ -6: 20210325 010000 7 11 255 0 0 -6: $ -6: $ end of input file -6: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[88]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.235235 ']' -6: ++ prep_step[4]echo gfs_ww3_grib.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[90]export err=0 -6: + wave_grib2_sbs.sh[90]err=0 -6: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[95]cat grib2_gsouth_037.out -6: -6: *** WAVEWATCH III GRIB output postp. *** -6: ============================================== -6: -6: Comment character is '$' -6: -6: Grid name : GFSv16-wave S Hemisphere 1/4 d -6: -6: LINEIN: -6: 20210325 010000 3600 1 -6: -6: 20210325010000 3600 1 -6: GEN_PRO -99999 -6: -6: Output time data : -6: ----------------------------------------------------- -6: First time : 2021/03/25 01:00:00 UTC -6: Interval : 01:00:00 -6: Number of requests : 1 -6: Fields : Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: Requested output fields not yet available: -6: ----------------------------------------------------- -6: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -6: -6: Successfully requested output fields : -6: ----------------------------------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: -6: Additional GRIB parameters : -6: ----------------------------------------------------- -6: Run time : 2021/03/25 01:00:00 UTC -6: GRIB center ID : 7 -6: GRIB gen. proc. ID : 11 -6: GRIB grid ID : 255 -6: GRIB GDS parameter : 0 -6: Fields in file : -6: -------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: CHOSEN GRID TYPE: : LLRECTILINEAR -6: -6: -6: -6: Generating file -6: ----------------------------------------------------- -6: Data for 2021/03/25 01:00:00 UTC 0H forecast. -6: -6: End of program -6: ========================================= -6: WAVEWATCH III GRIB output -6: -6: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -6: + wave_grib2_sbs.sh[102][[ 37 -gt 0 ]] -6: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '37 hour fcst' -grib gfs.wave.t12z.gsouth.0p25.f037.grib2 -6: 1:0:d=2021032312:SPC:surface:37 hour fcst: -6: 2:118279:d=2021032312:DIRC:surface:37 hour fcst: -6: 3:503134:d=2021032312:UOGRD:surface:37 hour fcst: -6: 4:616543:d=2021032312:VOGRD:surface:37 hour fcst: -6: 5:739282:d=2021032312:WIND:surface:37 hour fcst: -6: 6:983423:d=2021032312:WDIR:surface:37 hour fcst: -6: 7:1364246:d=2021032312:UGRD:surface:37 hour fcst: -6: 8:1601106:d=2021032312:VGRD:surface:37 hour fcst: -6: 9:1844068:d=2021032312:ICEC:surface:37 hour fcst: -6: 10:1905291:d=2021032312:HTSGW:surface:37 hour fcst: -6: 11:2062591:d=2021032312:IMWF:surface:37 hour fcst: -6: 12:2232264:d=2021032312:MWSPER:surface:37 hour fcst: -6: 13:2403563:d=2021032312:PERPW:surface:37 hour fcst: -6: 14:2596335:d=2021032312:WWSDIR:surface:37 hour fcst: -6: 15:2947904:d=2021032312:DIRPW:surface:37 hour fcst: -6: 16:3322551:d=2021032312:WVHGT:surface:37 hour fcst: -6: 17:3491195:d=2021032312:SWELL:1 in sequence:37 hour fcst: -6: 18:3632812:d=2021032312:SWELL:2 in sequence:37 hour fcst: -6: 19:3718025:d=2021032312:SWELL:3 in sequence:37 hour fcst: -6: 20:3773415:d=2021032312:WVPER:surface:37 hour fcst: -6: 21:3967320:d=2021032312:SWPER:1 in sequence:37 hour fcst: -6: 22:4129105:d=2021032312:SWPER:2 in sequence:37 hour fcst: -6: 23:4230710:d=2021032312:SWPER:3 in sequence:37 hour fcst: -6: 24:4290374:d=2021032312:WVDIR:surface:37 hour fcst: -6: 25:4614398:d=2021032312:SWDIR:1 in sequence:37 hour fcst: -6: 26:4902806:d=2021032312:SWDIR:2 in sequence:37 hour fcst: -6: 27:5062778:d=2021032312:SWDIR:3 in sequence:37 hour fcst: -6: + wave_grib2_sbs.sh[104]err=0 -6: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.gsouth.0p25.f037.grib2 -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f037.grib2 ]] -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f037.grib2.idx ]] -6: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.gsouth.0p25.f037.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f037.grib2 -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f037.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f037.grib2 = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f037.grib2 ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f037.grib2 -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f037.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f037.grib2.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f037.grib2.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f037.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f037.grib2 -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.gsouth.0p25.f037.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f037.grib2.idx -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f037.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f037.grib2.idx = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f037.grib2.idx ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f037.grib2.idx -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f037.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f037.grib2.idx.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f037.grib2.idx.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f037.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f037.grib2.idx -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.gsouth.0p25.f037.grib2 and gfs.wave.t12z.gsouth.0p25.f037.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_gsh_15m to COM' -6: INFO: Copied gfs.wave.t12z.gsouth.0p25.f037.grib2 and gfs.wave.t12z.gsouth.0p25.f037.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_gsh_15m to COM -6: + wave_grib2_sbs.sh[130][[ gsh_15m == '' ]] -6: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -6: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.gsouth.0p25.f037.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -6: INFO: gfs.wave.t12z.gsouth.0p25.f037.grib2 is global.0p50 or SENDDBN is NO, no alert sent -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + cmdfile.6[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032501 3600. 9999 -5: + cmdfile.6[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grid_interp_sbs.sh[25]grdID=gnh_10m -5: + wave_grid_interp_sbs.sh[26]valid_time=2021032501 -5: + wave_grid_interp_sbs.sh[27]dt=3600. -5: + wave_grid_interp_sbs.sh[28]nst=9999 -5: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551 -5: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.uglo_100km ./out_grd.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mod_def.uglo_100km ./mod_def.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mod_def.gnh_10m ./mod_def.gnh_10m -5: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m'\''' -5: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m' -5: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[51]weights_found=1 -5: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 010000' -5: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 010000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gnh_10m/g ww3_gint.inp.tmpl -5: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -5: $ Input file for interpolation of uglo_100km to gnh_10m -5: $------------------------------------------------ -5: $ Start Time 3600. NSteps -5: 20210325 010000 3600. 9999 -5: $ Total number of grids -5: 2 -5: $ Grid extensions -5: 'uglo_100km' -5: 'gnh_10m' -5: $ -5: 0 -5: $ -5: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[70]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.235235 ']' -5: ++ prep_step[4]echo gfs_ww3_gint.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -5: INFO: Executing 'gfs_ww3_gint.x' -5: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[73]cat grid_interp.gnh_10m.out -5: -5: *** WAVEWATCH III Grid interpolation *** -5: =============================================== -5: -5: Comment character is '$' -5: -5: Time Information : -5: --------------------------------------------- -5: Starting Time : 2021/03/25 01:00:00 UTC -5: Interval (in sec) : 3600.00 -5: Number of requests : 9999 -5: --------------------------------------------- -5: Number of grids (including output grid) = 2 -5: -5: -5: Extension for grid 1 is --> uglo_100km -5: -5: Grid Particulars are : -5: Dimensions = 45166 1 -5: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Extension for grid 2 is --> gnh_10m -5: -5: Grid Particulars are : -5: Dimensions = 2160 406 -5: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -5: -5: -5: Preparing interpolation weights for output grid -5: Total number of wet points for interpolation 571209 -5: -5: -5: Variable: Grid Interpolation Map Units: 0.100E+01 -5: -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 406 | | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: 355 | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 304 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 253 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 202 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 151 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 100 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 49 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: -5: -5: Interpolating fields .... -5: -5: Output group 1 -5: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -5: Output group 2 -5: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -5: Output group 3 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 4 -5: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -5: Output group 5 -5: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -5: Output group 6 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 7 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 8 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 9 -5: Output variables skipped -5: Output group 10 -5: Output variables skipped -5: ------------------------------------------------ -5: 1Current vel. -5: 1Wind speed -5: 1Ice concentration -5: 2Wave height -5: 2Mean wave period(+2) -5: 2Mean wave period(+1) -5: 2Peak frequency -5: 2Mean wave dir. a1b1 -5: 2Peak direction -5: 4Part. wave height -5: 4Part. peak period -5: 4Part. mean direction -5: 5Charnock parameter -5: ------------------------------------------------ -5: OUTPUT TIME : 2021/03/25 01:00:00 UTC -5: -5: End of file reached -5: -5: -5: *** End of Grid interpolation Routine *** -5: =============================================== -5: -5: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -5: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -5: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grid_interp_gnh_10m/out_grd.gnh_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.gnh_10m -5: + cmdfile.6[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032501 37 global 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + cmdfile.6[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib2_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grib2_sbs.sh[30]grdID=gnh_10m -5: + wave_grib2_sbs.sh[31]GRIDNR=255 -5: + wave_grib2_sbs.sh[32]MODNR=11 -5: + wave_grib2_sbs.sh[33]valid_time=2021032501 -5: + wave_grib2_sbs.sh[34]fhr=37 -5: + wave_grib2_sbs.sh[35]grid_region=global -5: + wave_grib2_sbs.sh[36]grid_res=0p16 -5: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551 -5: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_gnh_10m -5: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_gnh_10m -5: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_gnh_10m -5: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_gnh_10m -5: ++ wave_grib2_sbs.sh[47]printf %03i 37 -5: + wave_grib2_sbs.sh[47]FH3=037 -5: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p16 -5: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p16.f037.grib2 -5: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f037.grib2 ]] -5: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/mod_def.gnh_10m ./mod_def.ww3 -5: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/out_grd.gnh_10m ./out_grd.ww3 -5: + wave_grib2_sbs.sh[73]ngrib=1 -5: + wave_grib2_sbs.sh[74]dtgrib=3600 -5: + wave_grib2_sbs.sh[75]tstart='20210325 010000' -5: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 010000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gnh_10m.inp.tmpl -5: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -5: $ WAVEWATCH-III gridded output input file -5: $ ---------------------------------------- -5: 20210325 010000 3600 1 -5: N -5: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -5: $ -5: 20210325 010000 7 11 255 0 0 -5: $ -5: $ end of input file -5: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[88]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.235235 ']' -5: ++ prep_step[4]echo gfs_ww3_grib.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[90]export err=0 -5: + wave_grib2_sbs.sh[90]err=0 -5: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[95]cat grib2_global_037.out -5: -5: *** WAVEWATCH III GRIB output postp. *** -5: ============================================== -5: -5: Comment character is '$' -5: -5: Grid name : GFSv16-wave N Hemisphere 1/6 d -5: -5: LINEIN: -5: 20210325 010000 3600 1 -5: -5: 20210325010000 3600 1 -5: GEN_PRO -99999 -5: -5: Output time data : -5: ----------------------------------------------------- -5: First time : 2021/03/25 01:00:00 UTC -5: Interval : 01:00:00 -5: Number of requests : 1 -5: Fields : Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: Requested output fields not yet available: -5: ----------------------------------------------------- -5: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -5: -5: Successfully requested output fields : -5: ----------------------------------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: -5: Additional GRIB parameters : -5: ----------------------------------------------------- -5: Run time : 2021/03/25 01:00:00 UTC -5: GRIB center ID : 7 -5: GRIB gen. proc. ID : 11 -5: GRIB grid ID : 255 -5: GRIB GDS parameter : 0 -5: Fields in file : -5: -------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: CHOSEN GRID TYPE: : LLRECTILINEAR -5: -5: -5: -5: Generating file -5: ----------------------------------------------------- -5: Data for 2021/03/25 01:00:00 UTC 0H forecast. -5: -5: End of program -5: ========================================= -5: WAVEWATCH III GRIB output -5: -5: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -5: + wave_grib2_sbs.sh[102][[ 37 -gt 0 ]] -5: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '37 hour fcst' -grib gfs.wave.t12z.global.0p16.f037.grib2 -5: 1:0:d=2021032312:SPC:surface:37 hour fcst: -5: 2:235254:d=2021032312:DIRC:surface:37 hour fcst: -5: 3:860171:d=2021032312:UOGRD:surface:37 hour fcst: -5: 4:1090435:d=2021032312:VOGRD:surface:37 hour fcst: -5: 5:1328271:d=2021032312:WIND:surface:37 hour fcst: -5: 6:1729178:d=2021032312:WDIR:surface:37 hour fcst: -5: 7:2377741:d=2021032312:UGRD:surface:37 hour fcst: -5: 8:2775369:d=2021032312:VGRD:surface:37 hour fcst: -5: 9:3176593:d=2021032312:ICEC:surface:37 hour fcst: -5: 10:3291336:d=2021032312:HTSGW:surface:37 hour fcst: -5: 11:3573104:d=2021032312:IMWF:surface:37 hour fcst: -5: 12:3882021:d=2021032312:MWSPER:surface:37 hour fcst: -5: 13:4193375:d=2021032312:PERPW:surface:37 hour fcst: -5: 14:4529815:d=2021032312:WWSDIR:surface:37 hour fcst: -5: 15:5141205:d=2021032312:DIRPW:surface:37 hour fcst: -5: 16:5793471:d=2021032312:WVHGT:surface:37 hour fcst: -5: 17:6077453:d=2021032312:SWELL:1 in sequence:37 hour fcst: -5: 18:6299444:d=2021032312:SWELL:2 in sequence:37 hour fcst: -5: 19:6447964:d=2021032312:SWELL:3 in sequence:37 hour fcst: -5: 20:6563557:d=2021032312:WVPER:surface:37 hour fcst: -5: 21:6890650:d=2021032312:SWPER:1 in sequence:37 hour fcst: -5: 22:7145546:d=2021032312:SWPER:2 in sequence:37 hour fcst: -5: 23:7314528:d=2021032312:SWPER:3 in sequence:37 hour fcst: -5: 24:7435012:d=2021032312:WVDIR:surface:37 hour fcst: -5: 25:7987681:d=2021032312:SWDIR:1 in sequence:37 hour fcst: -5: 26:8421477:d=2021032312:SWDIR:2 in sequence:37 hour fcst: -5: 27:8665457:d=2021032312:SWDIR:3 in sequence:37 hour fcst: -5: + wave_grib2_sbs.sh[104]err=0 -5: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p16.f037.grib2 -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f037.grib2 ]] -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f037.grib2.idx ]] -5: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p16.f037.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f037.grib2 -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f037.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f037.grib2 = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f037.grib2 ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f037.grib2 -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f037.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f037.grib2.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f037.grib2.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f037.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f037.grib2 -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p16.f037.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f037.grib2.idx -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f037.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f037.grib2.idx = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f037.grib2.idx ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f037.grib2.idx -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f037.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f037.grib2.idx.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f037.grib2.idx.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f037.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f037.grib2.idx -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p16.f037.grib2 and gfs.wave.t12z.global.0p16.f037.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_gnh_10m to COM' -5: INFO: Copied gfs.wave.t12z.global.0p16.f037.grib2 and gfs.wave.t12z.global.0p16.f037.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551/grib_gnh_10m to COM -5: + wave_grib2_sbs.sh[130][[ gnh_10m == '' ]] -5: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -5: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p16.f037.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -5: INFO: gfs.wave.t12z.global.0p16.f037.grib2 is global.0p50 or SENDDBN is NO, no alert sent -+ run_mpmd.sh[113]exit 0 -+ run_mpmd.sh[1]postamble run_mpmd.sh 1753758300 0 -+ preamble.sh[62]set +x -End run_mpmd.sh at 03:05:10 with error code 0 (time elapsed: 00:00:10) -+ exgfs_wave_post_gridded_sbs.sh[122]true -+ exgfs_wave_post_gridded_sbs.sh[123]export err=0 -+ exgfs_wave_post_gridded_sbs.sh[123]err=0 -+ exgfs_wave_post_gridded_sbs.sh[124][[ 0 -ne 0 ]] -+ exgfs_wave_post_gridded_sbs.sh[130]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ exgfs_wave_post_gridded_sbs.sh[131]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ exgfs_wave_post_gridded_sbs.sh[132]gribchk=gfs.wave.t12z.gsouth.0p25.f037.grib2 -+ exgfs_wave_post_gridded_sbs.sh[133][[ ! -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f037.grib2 ]] -+ exgfs_wave_post_gridded_sbs.sh[138]exit 0 -+ JGLOBAL_WAVE_POST_SBS[28]true -+ JGLOBAL_WAVE_POST_SBS[29]export err=0 -+ JGLOBAL_WAVE_POST_SBS[29]err=0 -+ JGLOBAL_WAVE_POST_SBS[30][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[37]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312 -+ JGLOBAL_WAVE_POST_SBS[38][[ NO != \Y\E\S ]] -+ JGLOBAL_WAVE_POST_SBS[39]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f037.231551 -+ JGLOBAL_WAVE_POST_SBS[42]exit 0 -+ JGLOBAL_WAVE_POST_SBS[1]postamble /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS 1753758289 0 -+ preamble.sh[62]set +x -End /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at 03:05:11 with error code 0 (time elapsed: 00:00:22) -Begin /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at Tue Jul 29 03:05:11 UTC 2025 -++ jjob_header.sh[46]OPTIND=1 -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[50]env_job=wavepostsbs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[49]read -ra configs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[61]shift 4 -++ jjob_header.sh[63][[ -z wavepostsbs ]] -++ jjob_header.sh[71]export DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551 -++ jjob_header.sh[71]DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551 -++ jjob_header.sh[72][[ YES == \Y\E\S ]] -++ jjob_header.sh[73]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551 -++ jjob_header.sh[75]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551 -++ jjob_header.sh[76]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551 -++ jjob_header.sh[85]export pid=238224 -++ jjob_header.sh[85]pid=238224 -++ jjob_header.sh[86]export pgmout=OUTPUT.238224 -++ jjob_header.sh[86]pgmout=OUTPUT.238224 -++ jjob_header.sh[87]export pgmerr=errfile -++ jjob_header.sh[87]pgmerr=errfile -++ jjob_header.sh[90]export pgm= -++ jjob_header.sh[90]pgm= -++ jjob_header.sh[96]export cycle=t12z -++ jjob_header.sh[96]cycle=t12z -++ jjob_header.sh[97]setpdy.sh -+ setpdy.sh[20]'[' /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551 == /home/mterry ']' -+ setpdy.sh[25][[ ! t12z =~ t??z ]] -+ setpdy.sh[30]case $# in -+ setpdy.sh[31]dates_before_PDY=7 -+ setpdy.sh[32]dates_after_PDY=7 -+ setpdy.sh[50]COMDATEROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+ setpdy.sh[53]'[' -z 20210323 ']' -+ setpdy.sh[57]sed 's/[0-9]\{8\}/20210323/' /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z -sed: can't read /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z: No such file or directory -++ jjob_header.sh[97]true -++ jjob_header.sh[98]source ./PDY -/work2/noaa/global/mterry/global-workflow_forked/ush/jjob_header.sh: line 98: ./PDY: No such file or directory -++ jjob_header.sh[98]true -++ jjob_header.sh[104]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[104]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.base -+++ config.base[6]echo 'BEGIN: config.base' -BEGIN: config.base -+++ config.base[9]export machine=HERCULES -+++ config.base[9]machine=HERCULES -+++ config.base[12]export RUN_ENVIR=emc -+++ config.base[12]RUN_ENVIR=emc -+++ config.base[15]export ACCOUNT=fv3-cpu -+++ config.base[15]ACCOUNT=fv3-cpu -+++ config.base[16]export QUEUE=batch -+++ config.base[16]QUEUE=batch -+++ config.base[17]export QUEUE_SERVICE=batch -+++ config.base[17]QUEUE_SERVICE=batch -+++ config.base[18]export QUEUE_DTN=batch -+++ config.base[18]QUEUE_DTN=batch -+++ config.base[19]export PARTITION_BATCH=hercules -+++ config.base[19]PARTITION_BATCH=hercules -+++ config.base[20]export PARTITION_SERVICE=service -+++ config.base[20]PARTITION_SERVICE=service -+++ config.base[21]export PARTITION_DTN= -+++ config.base[21]PARTITION_DTN= -+++ config.base[22]export RESERVATION= -+++ config.base[22]RESERVATION= -+++ config.base[23]export CLUSTERS= -+++ config.base[23]CLUSTERS= -+++ config.base[24]export CLUSTERS_SERVICE= -+++ config.base[24]CLUSTERS_SERVICE= -+++ config.base[25]export CLUSTERS_DTN= -+++ config.base[25]CLUSTERS_DTN= -+++ config.base[28]export HPSS_PROJECT=emc-global -+++ config.base[28]HPSS_PROJECT=emc-global -+++ config.base[31]export HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[31]HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[32]export EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[32]EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[33]export FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[33]FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[34]export PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[34]PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[35]export SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[35]SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[36]export USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[36]USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[38]export FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[38]FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[39]export FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[39]FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[40]export FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[40]FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[41]export FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[41]FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[42]export FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[42]FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[43]export FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[43]FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[44]export FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[44]FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[45]export FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[45]FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[50]export PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[50]PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[51]export COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[51]COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[52]export COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[52]COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[53]export DMPDIR=/work/noaa/rstprod/dump -+++ config.base[53]DMPDIR=/work/noaa/rstprod/dump -+++ config.base[57]export COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[57]COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[58]export COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[58]COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[59]export COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[59]COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[62]export HOMEDIR=/work2/noaa/global/mterry -+++ config.base[62]HOMEDIR=/work2/noaa/global/mterry -+++ config.base[63]export STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[63]STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]export PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[65]export NOSCRUB=/work2/noaa/global/mterry -+++ config.base[65]NOSCRUB=/work2/noaa/global/mterry -+++ config.base[68]export BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[68]BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[71]export BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[71]BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[74]export DO_PREP_SFC=NO -+++ config.base[74]DO_PREP_SFC=NO -+++ config.base[77]export DO_GOES=NO -+++ config.base[77]DO_GOES=NO -+++ config.base[78]export DO_BUFRSND=NO -+++ config.base[78]DO_BUFRSND=NO -+++ config.base[79]export DO_GEMPAK=NO -+++ config.base[79]DO_GEMPAK=NO -+++ config.base[80]export DO_AWIPS=NO -+++ config.base[80]DO_AWIPS=NO -+++ config.base[81]export DO_NPOESS=NO -+++ config.base[81]DO_NPOESS=NO -+++ config.base[82]export DO_TRACKER=YES -+++ config.base[82]DO_TRACKER=YES -+++ config.base[83]export DO_GENESIS=YES -+++ config.base[83]DO_GENESIS=YES -+++ config.base[84]export DO_GENESIS_FSU=NO -+++ config.base[84]DO_GENESIS_FSU=NO -+++ config.base[85]export DO_VERFOZN=YES -+++ config.base[85]DO_VERFOZN=YES -+++ config.base[86]export DO_VERFRAD=YES -+++ config.base[86]DO_VERFRAD=YES -+++ config.base[87]export DO_VMINMON=YES -+++ config.base[87]DO_VMINMON=YES -+++ config.base[88]export DO_ANLSTAT=NO -+++ config.base[88]DO_ANLSTAT=NO -+++ config.base[91]export MODE=forecast-only -+++ config.base[91]MODE=forecast-only -+++ config.base[92]export DO_TEST_MODE=YES -+++ config.base[92]DO_TEST_MODE=YES -+++ config.base[101]export FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[101]FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[102]export HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[102]HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[103]export HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[103]HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[106]export NMV=/bin/mv -+++ config.base[106]NMV=/bin/mv -+++ config.base[107]export 'NLN=/bin/ln -sf' -+++ config.base[107]NLN='/bin/ln -sf' -+++ config.base[108]export VERBOSE=YES -+++ config.base[108]VERBOSE=YES -+++ config.base[109]export KEEPDATA=NO -+++ config.base[109]KEEPDATA=NO -+++ config.base[110]export DEBUG_POSTSCRIPT=NO -+++ config.base[110]DEBUG_POSTSCRIPT=NO -+++ config.base[111]export CHGRP_RSTPROD=YES -+++ config.base[111]CHGRP_RSTPROD=YES -+++ config.base[112]export 'CHGRP_CMD=chgrp rstprod' -+++ config.base[112]CHGRP_CMD='chgrp rstprod' -+++ config.base[113]export NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[113]NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[114]export NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[114]NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[117]export BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[117]BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[120]export SDATE=2021032312 -+++ config.base[120]SDATE=2021032312 -+++ config.base[121]export EDATE=2021032312 -+++ config.base[121]EDATE=2021032312 -+++ config.base[122]export EXP_WARM_START=.false. -+++ config.base[122]EXP_WARM_START=.false. -+++ config.base[123]export assim_freq=6 -+++ config.base[123]assim_freq=6 -+++ config.base[124]export PSLOT=C48_S2SW -+++ config.base[124]PSLOT=C48_S2SW -+++ config.base[125]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[125]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[126]export ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[126]ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[127]export DUMP_SUFFIX= -+++ config.base[127]DUMP_SUFFIX= -+++ config.base[128][[ 2021032312 -ge 2019092100 ]] -+++ config.base[128][[ 2021032312 -le 2019110700 ]] -+++ config.base[131]export ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[131]ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[132]export ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[132]ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[133]export FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[133]FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[136]export envir=prod -+++ config.base[136]envir=prod -+++ config.base[137]export NET=gfs -+++ config.base[137]NET=gfs -+++ config.base[138]export RUN=gfs -+++ config.base[138]RUN=gfs -+++ config.base[141]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.com -++++ config.com[4]echo 'BEGIN: config.com' -BEGIN: config.com -++++ config.com[38][[ emc == \n\c\o ]] -++++ config.com[43]COM_OBSPROC_TMPL='${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos' -++++ config.com[44]COM_RTOFS_TMPL='${DMPDIR}' -++++ config.com[45]COM_TCVITAL_TMPL='${DMPDIR}/${RUN}.${YMD}/${HH}/atmos' -++++ config.com[47]declare -rx 'COM_OBS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/obs' -++++ config.com[48]declare -rx COM_OBSPROC_TMPL COM_RTOFS_TMPL -++++ config.com[50]COM_BASE='${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}' -++++ config.com[52]declare -rx 'COM_TOP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}' -++++ config.com[54]declare -rx 'COM_CONF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf' -++++ config.com[55]declare -rx 'COM_OBS_JEDI=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi' -++++ config.com[57]declare -rx 'COM_ATMOS_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input' -++++ config.com[58]declare -rx 'COM_ATMOS_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart' -++++ config.com[59]declare -rx 'COM_ATMOS_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos' -++++ config.com[60]declare -rx 'COM_SNOW_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow' -++++ config.com[61]declare -rx 'COM_SNOW_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon' -++++ config.com[62]declare -rx 'COM_ATMOS_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history' -++++ config.com[63]declare -rx 'COM_ATMOS_MASTER_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master' -++++ config.com[64]declare -rx 'COM_ATMOS_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2' -++++ config.com[65]declare -rx 'COM_ATMOS_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}' -++++ config.com[66]declare -rx 'COM_ATMOS_BUFR_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr' -++++ config.com[67]declare -rx 'COM_ATMOS_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}' -++++ config.com[68]declare -rx 'COM_ATMOS_GENESIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital' -++++ config.com[69]declare -rx 'COM_ATMOS_TRACK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks' -++++ config.com[70]declare -rx 'COM_ATMOS_GOES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim' -++++ config.com[71]declare -rx 'COM_ATMOS_IMAGERY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery' -++++ config.com[72]declare -rx 'COM_ATMOS_OZNMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon' -++++ config.com[73]declare -rx 'COM_ATMOS_RADMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon' -++++ config.com[74]declare -rx 'COM_ATMOS_MINMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon' -++++ config.com[75]declare -rx 'COM_ATMOS_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon' -++++ config.com[76]declare -rx 'COM_ATMOS_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo' -++++ config.com[78]declare -rx 'COM_WAVE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart' -++++ config.com[79]declare -rx 'COM_WAVE_PREP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep' -++++ config.com[80]declare -rx 'COM_WAVE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history' -++++ config.com[81]declare -rx 'COM_WAVE_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded' -++++ config.com[82]declare -rx 'COM_WAVE_GRID_RES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}' -++++ config.com[83]declare -rx 'COM_WAVE_STATION_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station' -++++ config.com[84]declare -rx 'COM_WAVE_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak' -++++ config.com[85]declare -rx 'COM_WAVE_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo' -++++ config.com[87]declare -rx 'COM_OCEAN_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history' -++++ config.com[88]declare -rx 'COM_OCEAN_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart' -++++ config.com[89]declare -rx 'COM_OCEAN_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input' -++++ config.com[90]declare -rx 'COM_OCEAN_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean' -++++ config.com[91]declare -rx 'COM_OCEAN_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon' -++++ config.com[92]declare -rx 'COM_OCEAN_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf' -++++ config.com[93]declare -rx 'COM_OCEAN_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean' -++++ config.com[94]declare -rx 'COM_OCEAN_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf' -++++ config.com[95]declare -rx 'COM_OCEAN_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2' -++++ config.com[96]declare -rx 'COM_OCEAN_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}' -++++ config.com[98]declare -rx 'COM_ICE_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice' -++++ config.com[99]declare -rx 'COM_ICE_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf' -++++ config.com[100]declare -rx 'COM_ICE_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon' -++++ config.com[101]declare -rx 'COM_ICE_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice' -++++ config.com[102]declare -rx 'COM_ICE_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input' -++++ config.com[103]declare -rx 'COM_ICE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history' -++++ config.com[104]declare -rx 'COM_ICE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart' -++++ config.com[105]declare -rx 'COM_ICE_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf' -++++ config.com[106]declare -rx 'COM_ICE_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2' -++++ config.com[107]declare -rx 'COM_ICE_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}' -++++ config.com[109]declare -rx 'COM_CHEM_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history' -++++ config.com[110]declare -rx 'COM_CHEM_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem' -++++ config.com[111]declare -rx 'COM_CHEM_BMAT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix' -++++ config.com[112]declare -rx 'COM_CHEM_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon' -++++ config.com[114]declare -rx 'COM_MED_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart' -+++ config.base[143]export LOGSCRIPT= -+++ config.base[143]LOGSCRIPT= -+++ config.base[145]export 'REDOUT=1>' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wave -+++ config.wave[6]echo 'BEGIN: config.wave' -BEGIN: config.wave -+++ config.wave[13]export RUNRSTwave=gdas -+++ config.wave[13]RUNRSTwave=gdas -+++ config.wave[16]export MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[16]MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[19]case "${waveGRD}" in -+++ config.wave[64]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[64]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[65]export wavepostGRD= -+++ config.wave[65]wavepostGRD= -+++ config.wave[66]export waveuoutpGRD=uglo_100km -+++ config.wave[66]waveuoutpGRD=uglo_100km -+++ config.wave[75]export WAVEWND_DID= -+++ config.wave[75]WAVEWND_DID= -+++ config.wave[76]export WAVEWND_FID= -+++ config.wave[76]WAVEWND_FID= -+++ config.wave[79][[ gfs == \g\f\s ]] -+++ config.wave[80]export FHMAX_WAV=120 -+++ config.wave[80]FHMAX_WAV=120 -+++ config.wave[82]export WAVHINDH=0 -+++ config.wave[82]WAVHINDH=0 -+++ config.wave[83]export FHMAX_WAV_IBP=180 -+++ config.wave[83]FHMAX_WAV_IBP=180 -+++ config.wave[84](( FHMAX_WAV < FHMAX_WAV_IBP )) -+++ config.wave[84]export FHMAX_WAV_IBP=120 -+++ config.wave[84]FHMAX_WAV_IBP=120 -+++ config.wave[87]export DTFLD_WAV=3600 -+++ config.wave[87]DTFLD_WAV=3600 -+++ config.wave[88]export DTPNT_WAV=3600 -+++ config.wave[88]DTPNT_WAV=3600 -+++ config.wave[89]export FHINCP_WAV=1 -+++ config.wave[89]FHINCP_WAV=1 -+++ config.wave[92]export 'OUTPARS_WAV=WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[92]OUTPARS_WAV='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[95][[ gfs == \g\d\a\s ]] -+++ config.wave[99](( INTERVAL_GFS > 0 )) -+++ config.wave[100]export WAVHCYC=6 -+++ config.wave[100]WAVHCYC=6 -+++ config.wave[101]export FHMAX_WAV_CUR=192 -+++ config.wave[101]FHMAX_WAV_CUR=192 -+++ config.wave[109]export RSTTYPE_WAV=T -+++ config.wave[109]RSTTYPE_WAV=T -+++ config.wave[110][[ gfs != gfs ]] -+++ config.wave[115]rst_dt_gfs=43200 -+++ config.wave[116][[ 43200 -gt 0 ]] -+++ config.wave[117]export DT_1_RST_WAV=0 -+++ config.wave[117]DT_1_RST_WAV=0 -+++ config.wave[120]export DT_2_RST_WAV=43200 -+++ config.wave[120]DT_2_RST_WAV=43200 -+++ config.wave[126]export RSTIOFF_WAV=0 -+++ config.wave[126]RSTIOFF_WAV=0 -+++ config.wave[131]export RUNMEM=-1 -+++ config.wave[131]RUNMEM=-1 -+++ config.wave[134](( RUNMEM == -1 )) -+++ config.wave[136]export waveMEMB= -+++ config.wave[136]waveMEMB= -+++ config.wave[143]export WW3ATMINP=CPL -+++ config.wave[143]WW3ATMINP=CPL -+++ config.wave[144][[ YES == \Y\E\S ]] -+++ config.wave[145]export WW3ICEINP=CPL -+++ config.wave[145]WW3ICEINP=CPL -+++ config.wave[146]export WAVEICE_FID= -+++ config.wave[146]WAVEICE_FID= -+++ config.wave[152][[ YES == \Y\E\S ]] -+++ config.wave[153]export WW3CURINP=CPL -+++ config.wave[153]WW3CURINP=CPL -+++ config.wave[154]export WAVECUR_FID= -+++ config.wave[154]WAVECUR_FID= -+++ config.wave[161]export WW3ATMIENS=F -+++ config.wave[161]WW3ATMIENS=F -+++ config.wave[162]export WW3ICEIENS=F -+++ config.wave[162]WW3ICEIENS=F -+++ config.wave[163]export WW3CURIENS=F -+++ config.wave[163]WW3CURIENS=F -+++ config.wave[165]export GOFILETYPE=1 -+++ config.wave[165]GOFILETYPE=1 -+++ config.wave[166]export POFILETYPE=1 -+++ config.wave[166]POFILETYPE=1 -+++ config.wave[170]export FUNIPNT=T -+++ config.wave[170]FUNIPNT=T -+++ config.wave[172]export IOSRV=1 -+++ config.wave[172]IOSRV=1 -+++ config.wave[174]export FPNTPROC=T -+++ config.wave[174]FPNTPROC=T -+++ config.wave[176]export FGRDPROC=F -+++ config.wave[176]FGRDPROC=F -+++ config.wave[178]export FLAGMASKCOMP=F -+++ config.wave[178]FLAGMASKCOMP=F -+++ config.wave[180]export FLAGMASKOUT=F -+++ config.wave[180]FLAGMASKOUT=F -+++ config.wave[182]echo 'END: config.wave' -END: config.wave -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wavepostsbs -+++ config.wavepostsbs[6]echo 'BEGIN: config.wavepostsbs' -BEGIN: config.wavepostsbs -+++ config.wavepostsbs[9]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources wavepostsbs -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=wavepostsbs -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[156]ntasks=8 -++++ config.resources[157]threads_per_task=1 -++++ config.resources[158]tasks_per_node=80 -++++ config.resources[159]NTASKS=8 -++++ config.resources[160]memory=20GB -++++ config.resources[161]walltime=00:15:00 -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export NTASKS -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export memory -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.wavepostsbs[12]export MAX_TASKS=25 -+++ config.wavepostsbs[12]MAX_TASKS=25 -+++ config.wavepostsbs[15]export WAV_SUBGRBSRC= -+++ config.wavepostsbs[15]WAV_SUBGRBSRC= -+++ config.wavepostsbs[16]export WAV_SUBGRB= -+++ config.wavepostsbs[16]WAV_SUBGRB= -+++ config.wavepostsbs[19]export DOFLD_WAV=YES -+++ config.wavepostsbs[19]DOFLD_WAV=YES -+++ config.wavepostsbs[20]export DOPNT_WAV=YES -+++ config.wavepostsbs[20]DOPNT_WAV=YES -+++ config.wavepostsbs[21]export DOGRB_WAV=YES -+++ config.wavepostsbs[21]DOGRB_WAV=YES -+++ config.wavepostsbs[22][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+++ config.wavepostsbs[23]export DOGRI_WAV=YES -+++ config.wavepostsbs[23]DOGRI_WAV=YES -+++ config.wavepostsbs[27]export DOSPC_WAV=YES -+++ config.wavepostsbs[27]DOSPC_WAV=YES -+++ config.wavepostsbs[28]export DOBLL_WAV=YES -+++ config.wavepostsbs[28]DOBLL_WAV=YES -+++ config.wavepostsbs[30]echo 'END: config.wavepostsbs' -END: config.wavepostsbs -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env wavepostsbs -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=wavepostsbs -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 20971520 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 8 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[34]max_threads_per_task=1 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 1 ]] -+++ HERCULES.env[40][[ 1 -gt 1 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 8' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[63]export USE_CFP=YES -+++ HERCULES.env[63]USE_CFP=YES -+++ HERCULES.env[64][[ wavepostsbs == \w\a\v\e\p\r\e\p ]] -+++ HERCULES.env[65]export 'wavempexec=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[65]wavempexec='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[66]export 'wave_mpmd=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[66]wave_mpmd='--multi-prog --output=mpmd.%j.%t.out' -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[4]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ JGLOBAL_WAVE_POST_SBS[7]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[7]HH=12 -+ JGLOBAL_WAVE_POST_SBS[7]declare_from_tmpl -rx COMIN_WAVE_PREP:COM_WAVE_PREP_TMPL COMIN_WAVE_HISTORY:COM_WAVE_HISTORY_TMPL COMOUT_WAVE_GRID:COM_WAVE_GRID_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_WAVE_PREP=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep -declare_from_tmpl :: COMIN_WAVE_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history -declare_from_tmpl :: COMOUT_WAVE_GRID=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded -+ JGLOBAL_WAVE_POST_SBS[12][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n '' ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_atlocn_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_atlocn_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_atlocn_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_epacif_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_epacif_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_epacif_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_wcoast_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_wcoast_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_wcoast_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p50 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p50:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p50=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p50 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=arctic.9km -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_arctic_9km:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_arctic_9km=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_arctic_9km -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_gsouth_0p25:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_gsouth_0p25=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 ]] -+ JGLOBAL_WAVE_POST_SBS[28]/work2/noaa/global/mterry/global-workflow_forked/scripts/exgfs_wave_post_gridded_sbs.sh -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ exgfs_wave_post_gridded_sbs.sh[24]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ exgfs_wave_post_gridded_sbs.sh[26]DOGRI_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[27]DOGRB_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[29]export waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[29]waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[30]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[30]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[31]export wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[31]wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[33]cat - INFO: Grid information: - INFO: Native wave grids: uglo_100km - INFO: Interpolated grids: at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m - INFO: Post-process grids: -++ exgfs_wave_post_gridded_sbs.sh[40]printf %03i 38 -+ exgfs_wave_post_gridded_sbs.sh[40]fhr3=038 -++ exgfs_wave_post_gridded_sbs.sh[41]date -u -d '20210323 12 + 38 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[41]valid_time=2021032502 -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[49]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f038.bin ./out_grd.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f038.bin ./out_grd.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[52][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[62]cat - INFO: Summary: - INFO: Grid interp: DOGRI_WAV="YES" - INFO: Grib files: DOGRB_WAV="YES" - INFO: Fields to be included in grib files: - INFO: OUTPARS_WAV="WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA" -+ exgfs_wave_post_gridded_sbs.sh[70][[ YES == \N\O ]] -+ exgfs_wave_post_gridded_sbs.sh[76]rm -f 'cmdfile.*' cmdfile -+ exgfs_wave_post_gridded_sbs.sh[77]count=0 -+ exgfs_wave_post_gridded_sbs.sh[80][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[81]dt_int=3600. -+ exgfs_wave_post_gridded_sbs.sh[82]n_int=9999 -++ exgfs_wave_post_gridded_sbs.sh[83]date -u -d '20210325 02 - 0 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[83]ymdh_int=2021032502 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=1 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032502 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032502 38 atlocn 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib2_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib2_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=2 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032502 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032502 38 epacif 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib2_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib2_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=3 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032502 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032502 38 wcoast 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib2_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib2_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=4 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032502 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032502 38 global 0p50 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib2_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib2_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=5 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032502 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032502 38 arctic 9km '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib2_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib2_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=6 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032502 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032502 38 global 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib2_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib2_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=7 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032502 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032502 38 gsouth 0p25 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib2_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib2_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[100][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[113][[ 8 -lt 7 ]] -+ exgfs_wave_post_gridded_sbs.sh[121]echo 'INFO: Running MPMD job with 7 commands' -INFO: Running MPMD job with 7 commands -+ exgfs_wave_post_gridded_sbs.sh[122]/work2/noaa/global/mterry/global-workflow_forked/ush/run_mpmd.sh /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/cmdfile -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ run_mpmd.sh[31]source /work2/noaa/global/mterry/global-workflow_forked/ush/preamble.sh -++ preamble.sh[20]set +x -Begin run_mpmd.sh at Tue Jul 29 03:05:22 UTC 2025 -+ run_mpmd.sh[33]cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/cmdfile -+ run_mpmd.sh[36][[ YES != \Y\E\S ]] -+ run_mpmd.sh[46]export OMP_NUM_THREADS=1 -+ run_mpmd.sh[46]OMP_NUM_THREADS=1 -++ run_mpmd.sh[49]wc -l -+ run_mpmd.sh[49]nprocs=7 -+ run_mpmd.sh[52]mpmd_cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mpmd_cmdfile -+ run_mpmd.sh[53][[ -s /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mpmd_cmdfile ]] -+ run_mpmd.sh[55]cat - INFO: Executing MPMD job, STDOUT redirected for each process separately - INFO: On failure, logs for each job will be available in /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mpmd.proc_num.out - INFO: The proc_num corresponds to the line in '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mpmd_cmdfile' -+ run_mpmd.sh[61][[ srun -l --export=ALL --hint=nomultithread =~ ^srun.* ]] -+ run_mpmd.sh[65]nm=0 -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '0 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/cmdfile.1' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '1 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/cmdfile.2' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '2 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/cmdfile.3' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '3 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/cmdfile.4' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '4 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/cmdfile.5' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '5 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/cmdfile.6' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '6 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/cmdfile.7' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[72]set +e -+ run_mpmd.sh[74]srun -l --export=ALL --hint=nomultithread --multi-prog --output=mpmd.%j.%t.out -n 7 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mpmd_cmdfile -+ run_mpmd.sh[75]err=0 -+ run_mpmd.sh[76]set_strict -+ preamble.sh[35][[ YES == \Y\E\S ]] -+ preamble.sh[37]set -eu -+ run_mpmd.sh[103][[ 0 -eq 0 ]] -+ run_mpmd.sh[104]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mpmd_cmdfile -++ run_mpmd.sh[105]find . -name 'mpmd.*.out' -+ run_mpmd.sh[105]out_files='./mpmd.5951789.1.out -./mpmd.5951789.0.out -./mpmd.5951789.3.out -./mpmd.5951789.2.out -./mpmd.5951789.4.out -./mpmd.5951789.6.out -./mpmd.5951789.5.out' -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951789.1.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951789.1.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951789.0.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951789.0.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951789.3.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951789.3.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951789.2.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951789.2.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951789.4.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951789.4.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951789.6.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951789.6.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951789.5.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951789.5.out -+ run_mpmd.sh[110]cat mpmd.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + cmdfile.2[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032502 3600. 9999 -1: + cmdfile.2[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grid_interp_sbs.sh[25]grdID=ep_10m -1: + wave_grid_interp_sbs.sh[26]valid_time=2021032502 -1: + wave_grid_interp_sbs.sh[27]dt=3600. -1: + wave_grid_interp_sbs.sh[28]nst=9999 -1: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551 -1: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.uglo_100km ./out_grd.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mod_def.uglo_100km ./mod_def.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mod_def.ep_10m ./mod_def.ep_10m -1: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ]] -1: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m'\''' -1: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m' -1: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[51]weights_found=1 -1: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 020000' -1: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 020000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/ep_10m/g ww3_gint.inp.tmpl -1: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -1: $ Input file for interpolation of uglo_100km to ep_10m -1: $------------------------------------------------ -1: $ Start Time 3600. NSteps -1: 20210325 020000 3600. 9999 -1: $ Total number of grids -1: 2 -1: $ Grid extensions -1: 'uglo_100km' -1: 'ep_10m' -1: $ -1: 0 -1: $ -1: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[70]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.238224 ']' -1: ++ prep_step[4]echo gfs_ww3_gint.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -1: INFO: Executing 'gfs_ww3_gint.x' -1: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[73]cat grid_interp.ep_10m.out -1: -1: *** WAVEWATCH III Grid interpolation *** -1: =============================================== -1: -1: Comment character is '$' -1: -1: Time Information : -1: --------------------------------------------- -1: Starting Time : 2021/03/25 02:00:00 UTC -1: Interval (in sec) : 3600.00 -1: Number of requests : 9999 -1: --------------------------------------------- -1: Number of grids (including output grid) = 2 -1: -1: -1: Extension for grid 1 is --> uglo_100km -1: -1: Grid Particulars are : -1: Dimensions = 45166 1 -1: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Extension for grid 2 is --> ep_10m -1: -1: Grid Particulars are : -1: Dimensions = 511 301 -1: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -1: -1: -1: Preparing interpolation weights for output grid -1: Total number of wet points for interpolation 7439 -1: -1: -1: Variable: Grid Interpolation Map Units: 0.100E+01 -1: -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 301 | | -1: | | -1: | 0 0 | -1: 262 | 0 0 0 0 | -1: | 0 0 0 0 | -1: | 0 0 0 0 | -1: 223 | 0 0 0 | -1: | 0 | -1: | 0 | -1: 184 | | -1: | 0 0 0 | -1: | 0 0 0 0 | -1: 145 | 0 | -1: | | -1: | | -1: 106 | | -1: | | -1: | | -1: 67 | | -1: | 0 | -1: | | -1: 28 | 0 | -1: | | -1: | | -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: -1: -1: Interpolating fields .... -1: -1: Output group 1 -1: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -1: Output group 2 -1: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -1: Output group 3 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 4 -1: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -1: Output group 5 -1: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -1: Output group 6 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 7 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 8 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 9 -1: Output variables skipped -1: Output group 10 -1: Output variables skipped -1: ------------------------------------------------ -1: 1Current vel. -1: 1Wind speed -1: 1Ice concentration -1: 2Wave height -1: 2Mean wave period(+2) -1: 2Mean wave period(+1) -1: 2Peak frequency -1: 2Mean wave dir. a1b1 -1: 2Peak direction -1: 4Part. wave height -1: 4Part. peak period -1: 4Part. mean direction -1: 5Charnock parameter -1: ------------------------------------------------ -1: OUTPUT TIME : 2021/03/25 02:00:00 UTC -1: -1: End of file reached -1: -1: -1: *** End of Grid interpolation Routine *** -1: =============================================== -1: -1: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -1: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -1: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_ep_10m/out_grd.ep_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.ep_10m -1: + cmdfile.2[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032502 38 epacif 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + cmdfile.2[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib2_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grib2_sbs.sh[30]grdID=ep_10m -1: + wave_grib2_sbs.sh[31]GRIDNR=255 -1: + wave_grib2_sbs.sh[32]MODNR=11 -1: + wave_grib2_sbs.sh[33]valid_time=2021032502 -1: + wave_grib2_sbs.sh[34]fhr=38 -1: + wave_grib2_sbs.sh[35]grid_region=epacif -1: + wave_grib2_sbs.sh[36]grid_res=0p16 -1: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551 -1: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_ep_10m -1: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_ep_10m -1: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_ep_10m -1: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_ep_10m -1: ++ wave_grib2_sbs.sh[47]printf %03i 38 -1: + wave_grib2_sbs.sh[47]FH3=038 -1: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_epacif_0p16 -1: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.epacif.0p16.f038.grib2 -1: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f038.grib2 ]] -1: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mod_def.ep_10m ./mod_def.ww3 -1: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.ep_10m ./out_grd.ww3 -1: + wave_grib2_sbs.sh[73]ngrib=1 -1: + wave_grib2_sbs.sh[74]dtgrib=3600 -1: + wave_grib2_sbs.sh[75]tstart='20210325 020000' -1: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 020000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.ep_10m.inp.tmpl -1: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -1: $ WAVEWATCH-III gridded output input file -1: $ ---------------------------------------- -1: 20210325 020000 3600 1 -1: N -1: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -1: $ -1: 20210325 020000 7 11 255 0 0 -1: $ -1: $ end of input file -1: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[88]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.238224 ']' -1: ++ prep_step[4]echo gfs_ww3_grib.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[90]export err=0 -1: + wave_grib2_sbs.sh[90]err=0 -1: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[95]cat grib2_epacif_038.out -1: -1: *** WAVEWATCH III GRIB output postp. *** -1: ============================================== -1: -1: Comment character is '$' -1: -1: Grid name : East Pacific 10 min wave grid -1: -1: LINEIN: -1: 20210325 020000 3600 1 -1: -1: 20210325020000 3600 1 -1: GEN_PRO -99999 -1: -1: Output time data : -1: ----------------------------------------------------- -1: First time : 2021/03/25 02:00:00 UTC -1: Interval : 01:00:00 -1: Number of requests : 1 -1: Fields : Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: Requested output fields not yet available: -1: ----------------------------------------------------- -1: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -1: -1: Successfully requested output fields : -1: ----------------------------------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: -1: Additional GRIB parameters : -1: ----------------------------------------------------- -1: Run time : 2021/03/25 02:00:00 UTC -1: GRIB center ID : 7 -1: GRIB gen. proc. ID : 11 -1: GRIB grid ID : 255 -1: GRIB GDS parameter : 0 -1: Fields in file : -1: -------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: CHOSEN GRID TYPE: : LLRECTILINEAR -1: -1: -1: -1: Generating file -1: ----------------------------------------------------- -1: Data for 2021/03/25 02:00:00 UTC 0H forecast. -1: -1: End of program -1: ========================================= -1: WAVEWATCH III GRIB output -1: -1: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -1: + wave_grib2_sbs.sh[102][[ 38 -gt 0 ]] -1: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '38 hour fcst' -grib gfs.wave.t12z.epacif.0p16.f038.grib2 -1: 1:0:d=2021032312:SPC:surface:38 hour fcst: -1: 2:21808:d=2021032312:DIRC:surface:38 hour fcst: -1: 3:49475:d=2021032312:UOGRD:surface:38 hour fcst: -1: 4:71450:d=2021032312:VOGRD:surface:38 hour fcst: -1: 5:93293:d=2021032312:WIND:surface:38 hour fcst: -1: 6:117746:d=2021032312:WDIR:surface:38 hour fcst: -1: 7:145113:d=2021032312:UGRD:surface:38 hour fcst: -1: 8:169767:d=2021032312:VGRD:surface:38 hour fcst: -1: 9:194215:d=2021032312:ICEC:surface:38 hour fcst: -1: 10:213623:d=2021032312:HTSGW:surface:38 hour fcst: -1: 11:236528:d=2021032312:IMWF:surface:38 hour fcst: -1: 12:259688:d=2021032312:MWSPER:surface:38 hour fcst: -1: 13:282900:d=2021032312:PERPW:surface:38 hour fcst: -1: 14:306436:d=2021032312:WWSDIR:surface:38 hour fcst: -1: 15:333954:d=2021032312:DIRPW:surface:38 hour fcst: -1: 16:362100:d=2021032312:WVHGT:surface:38 hour fcst: -1: 17:385022:d=2021032312:SWELL:1 in sequence:38 hour fcst: -1: 18:406477:d=2021032312:SWELL:2 in sequence:38 hour fcst: -1: 19:427036:d=2021032312:SWELL:3 in sequence:38 hour fcst: -1: 20:446627:d=2021032312:WVPER:surface:38 hour fcst: -1: 21:470221:d=2021032312:SWPER:1 in sequence:38 hour fcst: -1: 22:492063:d=2021032312:SWPER:2 in sequence:38 hour fcst: -1: 23:512974:d=2021032312:SWPER:3 in sequence:38 hour fcst: -1: 24:532581:d=2021032312:WVDIR:surface:38 hour fcst: -1: 25:558860:d=2021032312:SWDIR:1 in sequence:38 hour fcst: -1: 26:583603:d=2021032312:SWDIR:2 in sequence:38 hour fcst: -1: 27:606214:d=2021032312:SWDIR:3 in sequence:38 hour fcst: -1: + wave_grib2_sbs.sh[104]err=0 -1: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.epacif.0p16.f038.grib2 -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f038.grib2 ]] -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f038.grib2.idx ]] -1: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.epacif.0p16.f038.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f038.grib2 -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f038.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f038.grib2 = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f038.grib2 ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f038.grib2 -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f038.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f038.grib2.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f038.grib2.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f038.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f038.grib2 -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.epacif.0p16.f038.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f038.grib2.idx -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f038.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f038.grib2.idx = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f038.grib2.idx ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f038.grib2.idx -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f038.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f038.grib2.idx.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f038.grib2.idx.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f038.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f038.grib2.idx -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.epacif.0p16.f038.grib2 and gfs.wave.t12z.epacif.0p16.f038.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_ep_10m to COM' -1: INFO: Copied gfs.wave.t12z.epacif.0p16.f038.grib2 and gfs.wave.t12z.epacif.0p16.f038.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_ep_10m to COM -1: + wave_grib2_sbs.sh[130][[ ep_10m == '' ]] -1: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -1: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.epacif.0p16.f038.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -1: INFO: gfs.wave.t12z.epacif.0p16.f038.grib2 is global.0p50 or SENDDBN is NO, no alert sent -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + cmdfile.1[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032502 3600. 9999 -0: + cmdfile.1[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grid_interp_sbs.sh[25]grdID=at_10m -0: + wave_grid_interp_sbs.sh[26]valid_time=2021032502 -0: + wave_grid_interp_sbs.sh[27]dt=3600. -0: + wave_grid_interp_sbs.sh[28]nst=9999 -0: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551 -0: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.uglo_100km ./out_grd.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mod_def.uglo_100km ./mod_def.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mod_def.at_10m ./mod_def.at_10m -0: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ]] -0: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m'\''' -0: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m' -0: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[51]weights_found=1 -0: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 020000' -0: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 020000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/at_10m/g ww3_gint.inp.tmpl -0: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -0: $ Input file for interpolation of uglo_100km to at_10m -0: $------------------------------------------------ -0: $ Start Time 3600. NSteps -0: 20210325 020000 3600. 9999 -0: $ Total number of grids -0: 2 -0: $ Grid extensions -0: 'uglo_100km' -0: 'at_10m' -0: $ -0: 0 -0: $ -0: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[70]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.238224 ']' -0: ++ prep_step[4]echo gfs_ww3_gint.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -0: INFO: Executing 'gfs_ww3_gint.x' -0: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[73]cat grid_interp.at_10m.out -0: -0: *** WAVEWATCH III Grid interpolation *** -0: =============================================== -0: -0: Comment character is '$' -0: -0: Time Information : -0: --------------------------------------------- -0: Starting Time : 2021/03/25 02:00:00 UTC -0: Interval (in sec) : 3600.00 -0: Number of requests : 9999 -0: --------------------------------------------- -0: Number of grids (including output grid) = 2 -0: -0: -0: Extension for grid 1 is --> uglo_100km -0: -0: Grid Particulars are : -0: Dimensions = 45166 1 -0: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Extension for grid 2 is --> at_10m -0: -0: Grid Particulars are : -0: Dimensions = 301 331 -0: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -0: -0: -0: Preparing interpolation weights for output grid -0: Total number of wet points for interpolation 29591 -0: -0: -0: Variable: Grid Interpolation Map Units: 0.100E+01 -0: -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 331 | | -0: | | -0: | | -0: 289 | | -0: | | -0: | 0 0 0 | -0: 247 | 0 0 0 0 | -0: | 0 0 0 0 0 | -0: | 0 0 0 0 | -0: 205 | 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 163 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 121 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 79 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 | -0: | 0 0 | -0: 37 | 0 | -0: | | -0: | | -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: -0: -0: Interpolating fields .... -0: -0: Output group 1 -0: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -0: Output group 2 -0: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -0: Output group 3 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 4 -0: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -0: Output group 5 -0: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -0: Output group 6 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 7 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 8 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 9 -0: Output variables skipped -0: Output group 10 -0: Output variables skipped -0: ------------------------------------------------ -0: 1Current vel. -0: 1Wind speed -0: 1Ice concentration -0: 2Wave height -0: 2Mean wave period(+2) -0: 2Mean wave period(+1) -0: 2Peak frequency -0: 2Mean wave dir. a1b1 -0: 2Peak direction -0: 4Part. wave height -0: 4Part. peak period -0: 4Part. mean direction -0: 5Charnock parameter -0: ------------------------------------------------ -0: OUTPUT TIME : 2021/03/25 02:00:00 UTC -0: -0: End of file reached -0: -0: -0: *** End of Grid interpolation Routine *** -0: =============================================== -0: -0: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -0: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -0: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_at_10m/out_grd.at_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.at_10m -0: + cmdfile.1[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032502 38 atlocn 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + cmdfile.1[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib2_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grib2_sbs.sh[30]grdID=at_10m -0: + wave_grib2_sbs.sh[31]GRIDNR=255 -0: + wave_grib2_sbs.sh[32]MODNR=11 -0: + wave_grib2_sbs.sh[33]valid_time=2021032502 -0: + wave_grib2_sbs.sh[34]fhr=38 -0: + wave_grib2_sbs.sh[35]grid_region=atlocn -0: + wave_grib2_sbs.sh[36]grid_res=0p16 -0: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551 -0: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_at_10m -0: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_at_10m -0: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_at_10m -0: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_at_10m -0: ++ wave_grib2_sbs.sh[47]printf %03i 38 -0: + wave_grib2_sbs.sh[47]FH3=038 -0: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_atlocn_0p16 -0: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.atlocn.0p16.f038.grib2 -0: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f038.grib2 ]] -0: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mod_def.at_10m ./mod_def.ww3 -0: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.at_10m ./out_grd.ww3 -0: + wave_grib2_sbs.sh[73]ngrib=1 -0: + wave_grib2_sbs.sh[74]dtgrib=3600 -0: + wave_grib2_sbs.sh[75]tstart='20210325 020000' -0: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 020000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.at_10m.inp.tmpl -0: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -0: $ WAVEWATCH-III gridded output input file -0: $ ---------------------------------------- -0: 20210325 020000 3600 1 -0: N -0: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -0: $ -0: 20210325 020000 7 11 255 0 0 -0: $ -0: $ end of input file -0: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[88]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.238224 ']' -0: ++ prep_step[4]echo gfs_ww3_grib.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[90]export err=0 -0: + wave_grib2_sbs.sh[90]err=0 -0: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[95]cat grib2_atlocn_038.out -0: -0: *** WAVEWATCH III GRIB output postp. *** -0: ============================================== -0: -0: Comment character is '$' -0: -0: Grid name : NW Atlantic 10 min wave grid -0: -0: LINEIN: -0: 20210325 020000 3600 1 -0: -0: 20210325020000 3600 1 -0: GEN_PRO -99999 -0: -0: Output time data : -0: ----------------------------------------------------- -0: First time : 2021/03/25 02:00:00 UTC -0: Interval : 01:00:00 -0: Number of requests : 1 -0: Fields : Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: Requested output fields not yet available: -0: ----------------------------------------------------- -0: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -0: -0: Successfully requested output fields : -0: ----------------------------------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: -0: Additional GRIB parameters : -0: ----------------------------------------------------- -0: Run time : 2021/03/25 02:00:00 UTC -0: GRIB center ID : 7 -0: GRIB gen. proc. ID : 11 -0: GRIB grid ID : 255 -0: GRIB GDS parameter : 0 -0: Fields in file : -0: -------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: CHOSEN GRID TYPE: : LLRECTILINEAR -0: -0: -0: -0: Generating file -0: ----------------------------------------------------- -0: Data for 2021/03/25 02:00:00 UTC 0H forecast. -0: -0: End of program -0: ========================================= -0: WAVEWATCH III GRIB output -0: -0: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -0: + wave_grib2_sbs.sh[102][[ 38 -gt 0 ]] -0: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '38 hour fcst' -grib gfs.wave.t12z.atlocn.0p16.f038.grib2 -0: 1:0:d=2021032312:SPC:surface:38 hour fcst: -0: 2:20022:d=2021032312:DIRC:surface:38 hour fcst: -0: 3:59332:d=2021032312:UOGRD:surface:38 hour fcst: -0: 4:79427:d=2021032312:VOGRD:surface:38 hour fcst: -0: 5:99152:d=2021032312:WIND:surface:38 hour fcst: -0: 6:129844:d=2021032312:WDIR:surface:38 hour fcst: -0: 7:172370:d=2021032312:UGRD:surface:38 hour fcst: -0: 8:202697:d=2021032312:VGRD:surface:38 hour fcst: -0: 9:233359:d=2021032312:ICEC:surface:38 hour fcst: -0: 10:245994:d=2021032312:HTSGW:surface:38 hour fcst: -0: 11:270324:d=2021032312:IMWF:surface:38 hour fcst: -0: 12:296441:d=2021032312:MWSPER:surface:38 hour fcst: -0: 13:322759:d=2021032312:PERPW:surface:38 hour fcst: -0: 14:350575:d=2021032312:WWSDIR:surface:38 hour fcst: -0: 15:392028:d=2021032312:DIRPW:surface:38 hour fcst: -0: 16:434423:d=2021032312:WVHGT:surface:38 hour fcst: -0: 17:458355:d=2021032312:SWELL:1 in sequence:38 hour fcst: -0: 18:477012:d=2021032312:SWELL:2 in sequence:38 hour fcst: -0: 19:492046:d=2021032312:SWELL:3 in sequence:38 hour fcst: -0: 20:505420:d=2021032312:WVPER:surface:38 hour fcst: -0: 21:532111:d=2021032312:SWPER:1 in sequence:38 hour fcst: -0: 22:552802:d=2021032312:SWPER:2 in sequence:38 hour fcst: -0: 23:569560:d=2021032312:SWPER:3 in sequence:38 hour fcst: -0: 24:583805:d=2021032312:WVDIR:surface:38 hour fcst: -0: 25:622135:d=2021032312:SWDIR:1 in sequence:38 hour fcst: -0: 26:650996:d=2021032312:SWDIR:2 in sequence:38 hour fcst: -0: 27:672551:d=2021032312:SWDIR:3 in sequence:38 hour fcst: -0: + wave_grib2_sbs.sh[104]err=0 -0: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.atlocn.0p16.f038.grib2 -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f038.grib2 ]] -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f038.grib2.idx ]] -0: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.atlocn.0p16.f038.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f038.grib2 -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f038.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f038.grib2 = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f038.grib2 ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f038.grib2 -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f038.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f038.grib2.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f038.grib2.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f038.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f038.grib2 -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.atlocn.0p16.f038.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f038.grib2.idx -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f038.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f038.grib2.idx = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f038.grib2.idx ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f038.grib2.idx -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f038.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f038.grib2.idx.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f038.grib2.idx.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f038.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f038.grib2.idx -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.atlocn.0p16.f038.grib2 and gfs.wave.t12z.atlocn.0p16.f038.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_at_10m to COM' -0: INFO: Copied gfs.wave.t12z.atlocn.0p16.f038.grib2 and gfs.wave.t12z.atlocn.0p16.f038.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_at_10m to COM -0: + wave_grib2_sbs.sh[130][[ at_10m == '' ]] -0: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -0: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.atlocn.0p16.f038.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -0: INFO: gfs.wave.t12z.atlocn.0p16.f038.grib2 is global.0p50 or SENDDBN is NO, no alert sent -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + cmdfile.4[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032502 3600. 9999 -3: + cmdfile.4[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grid_interp_sbs.sh[25]grdID=glo_30m -3: + wave_grid_interp_sbs.sh[26]valid_time=2021032502 -3: + wave_grid_interp_sbs.sh[27]dt=3600. -3: + wave_grid_interp_sbs.sh[28]nst=9999 -3: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551 -3: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.uglo_100km ./out_grd.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mod_def.uglo_100km ./mod_def.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mod_def.glo_30m ./mod_def.glo_30m -3: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ]] -3: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m'\''' -3: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m' -3: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[51]weights_found=1 -3: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 020000' -3: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 020000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/glo_30m/g ww3_gint.inp.tmpl -3: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -3: $ Input file for interpolation of uglo_100km to glo_30m -3: $------------------------------------------------ -3: $ Start Time 3600. NSteps -3: 20210325 020000 3600. 9999 -3: $ Total number of grids -3: 2 -3: $ Grid extensions -3: 'uglo_100km' -3: 'glo_30m' -3: $ -3: 0 -3: $ -3: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[70]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.238224 ']' -3: ++ prep_step[4]echo gfs_ww3_gint.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -3: INFO: Executing 'gfs_ww3_gint.x' -3: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[73]cat grid_interp.glo_30m.out -3: -3: *** WAVEWATCH III Grid interpolation *** -3: =============================================== -3: -3: Comment character is '$' -3: -3: Time Information : -3: --------------------------------------------- -3: Starting Time : 2021/03/25 02:00:00 UTC -3: Interval (in sec) : 3600.00 -3: Number of requests : 9999 -3: --------------------------------------------- -3: Number of grids (including output grid) = 2 -3: -3: -3: Extension for grid 1 is --> uglo_100km -3: -3: Grid Particulars are : -3: Dimensions = 45166 1 -3: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Extension for grid 2 is --> glo_30m -3: -3: Grid Particulars are : -3: Dimensions = 720 336 -3: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -3: -3: -3: Preparing interpolation weights for output grid -3: Total number of wet points for interpolation 167619 -3: -3: -3: Variable: Grid Interpolation Map Units: 0.100E+01 -3: -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 336 | | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 291 | 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 | -3: 246 | 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 201 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 156 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 111 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 66 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 21 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 | -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: -3: -3: Interpolating fields .... -3: -3: Output group 1 -3: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -3: Output group 2 -3: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -3: Output group 3 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 4 -3: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -3: Output group 5 -3: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -3: Output group 6 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 7 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 8 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 9 -3: Output variables skipped -3: Output group 10 -3: Output variables skipped -3: ------------------------------------------------ -3: 1Current vel. -3: 1Wind speed -3: 1Ice concentration -3: 2Wave height -3: 2Mean wave period(+2) -3: 2Mean wave period(+1) -3: 2Peak frequency -3: 2Mean wave dir. a1b1 -3: 2Peak direction -3: 4Part. wave height -3: 4Part. peak period -3: 4Part. mean direction -3: 5Charnock parameter -3: ------------------------------------------------ -3: OUTPUT TIME : 2021/03/25 02:00:00 UTC -3: -3: End of file reached -3: -3: -3: *** End of Grid interpolation Routine *** -3: =============================================== -3: -3: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -3: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -3: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_glo_30m/out_grd.glo_30m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.glo_30m -3: + cmdfile.4[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032502 38 global 0p50 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + cmdfile.4[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib2_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grib2_sbs.sh[30]grdID=glo_30m -3: + wave_grib2_sbs.sh[31]GRIDNR=255 -3: + wave_grib2_sbs.sh[32]MODNR=11 -3: + wave_grib2_sbs.sh[33]valid_time=2021032502 -3: + wave_grib2_sbs.sh[34]fhr=38 -3: + wave_grib2_sbs.sh[35]grid_region=global -3: + wave_grib2_sbs.sh[36]grid_res=0p50 -3: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551 -3: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_glo_30m -3: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_glo_30m -3: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_glo_30m -3: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_glo_30m -3: ++ wave_grib2_sbs.sh[47]printf %03i 38 -3: + wave_grib2_sbs.sh[47]FH3=038 -3: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p50 -3: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p50.f038.grib2 -3: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f038.grib2 ]] -3: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mod_def.glo_30m ./mod_def.ww3 -3: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.glo_30m ./out_grd.ww3 -3: + wave_grib2_sbs.sh[73]ngrib=1 -3: + wave_grib2_sbs.sh[74]dtgrib=3600 -3: + wave_grib2_sbs.sh[75]tstart='20210325 020000' -3: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 020000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.glo_30m.inp.tmpl -3: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -3: $ WAVEWATCH-III gridded output input file -3: $ ---------------------------------------- -3: 20210325 020000 3600 1 -3: N -3: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -3: $ -3: 20210325 020000 7 11 255 0 0 -3: $ -3: $ end of input file -3: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[88]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.238224 ']' -3: ++ prep_step[4]echo gfs_ww3_grib.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[90]export err=0 -3: + wave_grib2_sbs.sh[90]err=0 -3: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[95]cat grib2_global_038.out -3: -3: *** WAVEWATCH III GRIB output postp. *** -3: ============================================== -3: -3: Comment character is '$' -3: -3: Grid name : Global 30 min wave grid -3: -3: LINEIN: -3: 20210325 020000 3600 1 -3: -3: 20210325020000 3600 1 -3: GEN_PRO -99999 -3: -3: Output time data : -3: ----------------------------------------------------- -3: First time : 2021/03/25 02:00:00 UTC -3: Interval : 01:00:00 -3: Number of requests : 1 -3: Fields : Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: Requested output fields not yet available: -3: ----------------------------------------------------- -3: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -3: -3: Successfully requested output fields : -3: ----------------------------------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: -3: Additional GRIB parameters : -3: ----------------------------------------------------- -3: Run time : 2021/03/25 02:00:00 UTC -3: GRIB center ID : 7 -3: GRIB gen. proc. ID : 11 -3: GRIB grid ID : 255 -3: GRIB GDS parameter : 0 -3: Fields in file : -3: -------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: CHOSEN GRID TYPE: : LLRECTILINEAR -3: -3: -3: -3: Generating file -3: ----------------------------------------------------- -3: Data for 2021/03/25 02:00:00 UTC 0H forecast. -3: -3: End of program -3: ========================================= -3: WAVEWATCH III GRIB output -3: -3: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -3: + wave_grib2_sbs.sh[102][[ 38 -gt 0 ]] -3: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '38 hour fcst' -grib gfs.wave.t12z.global.0p50.f038.grib2 -3: 1:0:d=2021032312:SPC:surface:38 hour fcst: -3: 2:77944:d=2021032312:DIRC:surface:38 hour fcst: -3: 3:319685:d=2021032312:UOGRD:surface:38 hour fcst: -3: 4:394718:d=2021032312:VOGRD:surface:38 hour fcst: -3: 5:474572:d=2021032312:WIND:surface:38 hour fcst: -3: 6:631846:d=2021032312:WDIR:surface:38 hour fcst: -3: 7:868196:d=2021032312:UGRD:surface:38 hour fcst: -3: 8:1022411:d=2021032312:VGRD:surface:38 hour fcst: -3: 9:1179435:d=2021032312:ICEC:surface:38 hour fcst: -3: 10:1219820:d=2021032312:HTSGW:surface:38 hour fcst: -3: 11:1318481:d=2021032312:IMWF:surface:38 hour fcst: -3: 12:1426847:d=2021032312:MWSPER:surface:38 hour fcst: -3: 13:1536557:d=2021032312:PERPW:surface:38 hour fcst: -3: 14:1658044:d=2021032312:WWSDIR:surface:38 hour fcst: -3: 15:1866536:d=2021032312:DIRPW:surface:38 hour fcst: -3: 16:2087140:d=2021032312:WVHGT:surface:38 hour fcst: -3: 17:2188401:d=2021032312:SWELL:1 in sequence:38 hour fcst: -3: 18:2271203:d=2021032312:SWELL:2 in sequence:38 hour fcst: -3: 19:2320938:d=2021032312:SWELL:3 in sequence:38 hour fcst: -3: 20:2354504:d=2021032312:WVPER:surface:38 hour fcst: -3: 21:2471981:d=2021032312:SWPER:1 in sequence:38 hour fcst: -3: 22:2566632:d=2021032312:SWPER:2 in sequence:38 hour fcst: -3: 23:2624899:d=2021032312:SWPER:3 in sequence:38 hour fcst: -3: 24:2660682:d=2021032312:WVDIR:surface:38 hour fcst: -3: 25:2848724:d=2021032312:SWDIR:1 in sequence:38 hour fcst: -3: 26:3006344:d=2021032312:SWDIR:2 in sequence:38 hour fcst: -3: 27:3091567:d=2021032312:SWDIR:3 in sequence:38 hour fcst: -3: + wave_grib2_sbs.sh[104]err=0 -3: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p50.f038.grib2 -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f038.grib2 ]] -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f038.grib2.idx ]] -3: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p50.f038.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f038.grib2 -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f038.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f038.grib2 = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f038.grib2 ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f038.grib2 -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f038.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f038.grib2.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f038.grib2.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f038.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f038.grib2 -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p50.f038.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f038.grib2.idx -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f038.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f038.grib2.idx = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f038.grib2.idx ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f038.grib2.idx -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f038.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f038.grib2.idx.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f038.grib2.idx.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f038.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f038.grib2.idx -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p50.f038.grib2 and gfs.wave.t12z.global.0p50.f038.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_glo_30m to COM' -3: INFO: Copied gfs.wave.t12z.global.0p50.f038.grib2 and gfs.wave.t12z.global.0p50.f038.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_glo_30m to COM -3: + wave_grib2_sbs.sh[130][[ glo_30m == '' ]] -3: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -3: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p50.f038.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -3: INFO: gfs.wave.t12z.global.0p50.f038.grib2 is global.0p50 or SENDDBN is NO, no alert sent -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + cmdfile.3[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032502 3600. 9999 -2: + cmdfile.3[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grid_interp_sbs.sh[25]grdID=wc_10m -2: + wave_grid_interp_sbs.sh[26]valid_time=2021032502 -2: + wave_grid_interp_sbs.sh[27]dt=3600. -2: + wave_grid_interp_sbs.sh[28]nst=9999 -2: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551 -2: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.uglo_100km ./out_grd.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mod_def.uglo_100km ./mod_def.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mod_def.wc_10m ./mod_def.wc_10m -2: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ]] -2: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m'\''' -2: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m' -2: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[51]weights_found=1 -2: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 020000' -2: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 020000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/wc_10m/g ww3_gint.inp.tmpl -2: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -2: $ Input file for interpolation of uglo_100km to wc_10m -2: $------------------------------------------------ -2: $ Start Time 3600. NSteps -2: 20210325 020000 3600. 9999 -2: $ Total number of grids -2: 2 -2: $ Grid extensions -2: 'uglo_100km' -2: 'wc_10m' -2: $ -2: 0 -2: $ -2: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[70]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.238224 ']' -2: ++ prep_step[4]echo gfs_ww3_gint.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -2: INFO: Executing 'gfs_ww3_gint.x' -2: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[73]cat grid_interp.wc_10m.out -2: -2: *** WAVEWATCH III Grid interpolation *** -2: =============================================== -2: -2: Comment character is '$' -2: -2: Time Information : -2: --------------------------------------------- -2: Starting Time : 2021/03/25 02:00:00 UTC -2: Interval (in sec) : 3600.00 -2: Number of requests : 9999 -2: --------------------------------------------- -2: Number of grids (including output grid) = 2 -2: -2: -2: Extension for grid 1 is --> uglo_100km -2: -2: Grid Particulars are : -2: Dimensions = 45166 1 -2: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Extension for grid 2 is --> wc_10m -2: -2: Grid Particulars are : -2: Dimensions = 241 151 -2: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -2: -2: -2: Preparing interpolation weights for output grid -2: Total number of wet points for interpolation 11044 -2: -2: -2: Variable: Grid Interpolation Map Units: 0.100E+01 -2: -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: +---------------------------------------------------------------------------------------------------------------+ -2: 151 | 0 0 0 0 | -2: | 0 0 0 0 0 | -2: | 0 0 0 0 0 | -2: 130 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 109 | 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 88 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 | -2: 67 | 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 | -2: 46 | 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: 25 | 0 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 0 | -2: | | -2: 4 | | -2: +---------------------------------------------------------------------------------------------------------------+ -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: -2: -2: Interpolating fields .... -2: -2: Output group 1 -2: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -2: Output group 2 -2: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -2: Output group 3 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 4 -2: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -2: Output group 5 -2: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -2: Output group 6 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 7 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 8 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 9 -2: Output variables skipped -2: Output group 10 -2: Output variables skipped -2: ------------------------------------------------ -2: 1Current vel. -2: 1Wind speed -2: 1Ice concentration -2: 2Wave height -2: 2Mean wave period(+2) -2: 2Mean wave period(+1) -2: 2Peak frequency -2: 2Mean wave dir. a1b1 -2: 2Peak direction -2: 4Part. wave height -2: 4Part. peak period -2: 4Part. mean direction -2: 5Charnock parameter -2: ------------------------------------------------ -2: OUTPUT TIME : 2021/03/25 02:00:00 UTC -2: -2: End of file reached -2: -2: -2: *** End of Grid interpolation Routine *** -2: =============================================== -2: -2: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -2: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -2: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_wc_10m/out_grd.wc_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.wc_10m -2: + cmdfile.3[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032502 38 wcoast 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + cmdfile.3[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib2_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grib2_sbs.sh[30]grdID=wc_10m -2: + wave_grib2_sbs.sh[31]GRIDNR=255 -2: + wave_grib2_sbs.sh[32]MODNR=11 -2: + wave_grib2_sbs.sh[33]valid_time=2021032502 -2: + wave_grib2_sbs.sh[34]fhr=38 -2: + wave_grib2_sbs.sh[35]grid_region=wcoast -2: + wave_grib2_sbs.sh[36]grid_res=0p16 -2: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551 -2: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_wc_10m -2: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_wc_10m -2: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_wc_10m -2: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_wc_10m -2: ++ wave_grib2_sbs.sh[47]printf %03i 38 -2: + wave_grib2_sbs.sh[47]FH3=038 -2: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_wcoast_0p16 -2: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.wcoast.0p16.f038.grib2 -2: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f038.grib2 ]] -2: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mod_def.wc_10m ./mod_def.ww3 -2: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.wc_10m ./out_grd.ww3 -2: + wave_grib2_sbs.sh[73]ngrib=1 -2: + wave_grib2_sbs.sh[74]dtgrib=3600 -2: + wave_grib2_sbs.sh[75]tstart='20210325 020000' -2: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 020000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.wc_10m.inp.tmpl -2: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -2: $ WAVEWATCH-III gridded output input file -2: $ ---------------------------------------- -2: 20210325 020000 3600 1 -2: N -2: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -2: $ -2: 20210325 020000 7 11 255 0 0 -2: $ -2: $ end of input file -2: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[88]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.238224 ']' -2: ++ prep_step[4]echo gfs_ww3_grib.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[90]export err=0 -2: + wave_grib2_sbs.sh[90]err=0 -2: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[95]cat grib2_wcoast_038.out -2: -2: *** WAVEWATCH III GRIB output postp. *** -2: ============================================== -2: -2: Comment character is '$' -2: -2: Grid name : West Coast 10 min wave grid -2: -2: LINEIN: -2: 20210325 020000 3600 1 -2: -2: 20210325020000 3600 1 -2: GEN_PRO -99999 -2: -2: Output time data : -2: ----------------------------------------------------- -2: First time : 2021/03/25 02:00:00 UTC -2: Interval : 01:00:00 -2: Number of requests : 1 -2: Fields : Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: Requested output fields not yet available: -2: ----------------------------------------------------- -2: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -2: -2: Successfully requested output fields : -2: ----------------------------------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: -2: Additional GRIB parameters : -2: ----------------------------------------------------- -2: Run time : 2021/03/25 02:00:00 UTC -2: GRIB center ID : 7 -2: GRIB gen. proc. ID : 11 -2: GRIB grid ID : 255 -2: GRIB GDS parameter : 0 -2: Fields in file : -2: -------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: CHOSEN GRID TYPE: : LLRECTILINEAR -2: -2: -2: -2: Generating file -2: ----------------------------------------------------- -2: Data for 2021/03/25 02:00:00 UTC 0H forecast. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: -2: End of program -2: ========================================= -2: WAVEWATCH III GRIB output -2: -2: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -2: + wave_grib2_sbs.sh[102][[ 38 -gt 0 ]] -2: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '38 hour fcst' -grib gfs.wave.t12z.wcoast.0p16.f038.grib2 -2: 1:0:d=2021032312:SPC:surface:38 hour fcst: -2: 2:6963:d=2021032312:DIRC:surface:38 hour fcst: -2: 3:23296:d=2021032312:UOGRD:surface:38 hour fcst: -2: 4:30461:d=2021032312:VOGRD:surface:38 hour fcst: -2: 5:37818:d=2021032312:WIND:surface:38 hour fcst: -2: 6:49254:d=2021032312:WDIR:surface:38 hour fcst: -2: 7:64874:d=2021032312:UGRD:surface:38 hour fcst: -2: 8:76226:d=2021032312:VGRD:surface:38 hour fcst: -2: 9:87762:d=2021032312:ICEC:surface:38 hour fcst: -2: 10:92492:d=2021032312:HTSGW:surface:38 hour fcst: -2: 11:101968:d=2021032312:IMWF:surface:38 hour fcst: -2: 12:111563:d=2021032312:MWSPER:surface:38 hour fcst: -2: 13:121226:d=2021032312:PERPW:surface:38 hour fcst: -2: 14:131032:d=2021032312:WWSDIR:surface:38 hour fcst: -2: 15:146175:d=2021032312:DIRPW:surface:38 hour fcst: -2: 16:161782:d=2021032312:WVHGT:surface:38 hour fcst: -2: 17:171141:d=2021032312:SWELL:1 in sequence:38 hour fcst: -2: 18:177051:d=2021032312:SWELL:2 in sequence:38 hour fcst: -2: 19:181781:d=2021032312:SWELL:3 in sequence:38 hour fcst: -2: 20:186511:d=2021032312:WVPER:surface:38 hour fcst: -2: 21:196114:d=2021032312:SWPER:1 in sequence:38 hour fcst: -2: 22:202125:d=2021032312:SWPER:2 in sequence:38 hour fcst: -2: 23:206855:d=2021032312:SWPER:3 in sequence:38 hour fcst: -2: 24:211585:d=2021032312:WVDIR:surface:38 hour fcst: -2: 25:226443:d=2021032312:SWDIR:1 in sequence:38 hour fcst: -2: 26:233840:d=2021032312:SWDIR:2 in sequence:38 hour fcst: -2: 27:238570:d=2021032312:SWDIR:3 in sequence:38 hour fcst: -2: + wave_grib2_sbs.sh[104]err=0 -2: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.wcoast.0p16.f038.grib2 -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f038.grib2 ]] -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f038.grib2.idx ]] -2: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.wcoast.0p16.f038.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f038.grib2 -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f038.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f038.grib2 = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f038.grib2 ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f038.grib2 -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f038.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f038.grib2.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f038.grib2.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f038.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f038.grib2 -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.wcoast.0p16.f038.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f038.grib2.idx -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f038.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f038.grib2.idx = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f038.grib2.idx ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f038.grib2.idx -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f038.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f038.grib2.idx.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f038.grib2.idx.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f038.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f038.grib2.idx -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.wcoast.0p16.f038.grib2 and gfs.wave.t12z.wcoast.0p16.f038.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_wc_10m to COM' -2: INFO: Copied gfs.wave.t12z.wcoast.0p16.f038.grib2 and gfs.wave.t12z.wcoast.0p16.f038.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_wc_10m to COM -2: + wave_grib2_sbs.sh[130][[ wc_10m == '' ]] -2: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -2: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.wcoast.0p16.f038.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -2: INFO: gfs.wave.t12z.wcoast.0p16.f038.grib2 is global.0p50 or SENDDBN is NO, no alert sent -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + cmdfile.5[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032502 3600. 9999 -4: + cmdfile.5[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grid_interp_sbs.sh[25]grdID=aoc_9km -4: + wave_grid_interp_sbs.sh[26]valid_time=2021032502 -4: + wave_grid_interp_sbs.sh[27]dt=3600. -4: + wave_grid_interp_sbs.sh[28]nst=9999 -4: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551 -4: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.uglo_100km ./out_grd.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mod_def.uglo_100km ./mod_def.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mod_def.aoc_9km ./mod_def.aoc_9km -4: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km'\''' -4: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km' -4: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[51]weights_found=1 -4: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 020000' -4: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 020000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/aoc_9km/g ww3_gint.inp.tmpl -4: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -4: $ Input file for interpolation of uglo_100km to aoc_9km -4: $------------------------------------------------ -4: $ Start Time 3600. NSteps -4: 20210325 020000 3600. 9999 -4: $ Total number of grids -4: 2 -4: $ Grid extensions -4: 'uglo_100km' -4: 'aoc_9km' -4: $ -4: 0 -4: $ -4: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[70]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.238224 ']' -4: ++ prep_step[4]echo gfs_ww3_gint.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -4: INFO: Executing 'gfs_ww3_gint.x' -4: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[73]cat grid_interp.aoc_9km.out -4: -4: *** WAVEWATCH III Grid interpolation *** -4: =============================================== -4: -4: Comment character is '$' -4: -4: Time Information : -4: --------------------------------------------- -4: Starting Time : 2021/03/25 02:00:00 UTC -4: Interval (in sec) : 3600.00 -4: Number of requests : 9999 -4: --------------------------------------------- -4: Number of grids (including output grid) = 2 -4: -4: -4: Extension for grid 1 is --> uglo_100km -4: -4: Grid Particulars are : -4: Dimensions = 45166 1 -4: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Extension for grid 2 is --> aoc_9km -4: -4: Grid Particulars are : -4: Dimensions = 1006 1006 -4: Grid Type = 2 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -4: -4: -4: Preparing interpolation weights for output grid -4: Total number of wet points for interpolation 360052 -4: -4: -4: Variable: Grid Interpolation Map Units: 0.100E+01 -4: -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: *** | | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: 880 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 | -4: 754 | 0 0 0 | -4: | 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 628 | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 | -4: 502 | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 376 | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 250 | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 124 | 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 | -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: -4: -4: Interpolating fields .... -4: -4: Output group 1 -4: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -4: Output group 2 -4: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -4: Output group 3 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 4 -4: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -4: Output group 5 -4: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -4: Output group 6 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 7 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 8 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 9 -4: Output variables skipped -4: Output group 10 -4: Output variables skipped -4: ------------------------------------------------ -4: 1Current vel. -4: 1Wind speed -4: 1Ice concentration -4: 2Wave height -4: 2Mean wave period(+2) -4: 2Mean wave period(+1) -4: 2Peak frequency -4: 2Mean wave dir. a1b1 -4: 2Peak direction -4: 4Part. wave height -4: 4Part. peak period -4: 4Part. mean direction -4: 5Charnock parameter -4: ------------------------------------------------ -4: OUTPUT TIME : 2021/03/25 02:00:00 UTC -4: -4: End of file reached -4: -4: -4: *** End of Grid interpolation Routine *** -4: =============================================== -4: -4: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -4: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -4: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_aoc_9km/out_grd.aoc_9km /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.aoc_9km -4: + cmdfile.5[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032502 38 arctic 9km 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + cmdfile.5[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib2_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grib2_sbs.sh[30]grdID=aoc_9km -4: + wave_grib2_sbs.sh[31]GRIDNR=255 -4: + wave_grib2_sbs.sh[32]MODNR=11 -4: + wave_grib2_sbs.sh[33]valid_time=2021032502 -4: + wave_grib2_sbs.sh[34]fhr=38 -4: + wave_grib2_sbs.sh[35]grid_region=arctic -4: + wave_grib2_sbs.sh[36]grid_res=9km -4: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551 -4: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_aoc_9km -4: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_aoc_9km -4: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_aoc_9km -4: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_aoc_9km -4: ++ wave_grib2_sbs.sh[47]printf %03i 38 -4: + wave_grib2_sbs.sh[47]FH3=038 -4: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_arctic_9km -4: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.arctic.9km.f038.grib2 -4: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f038.grib2 ]] -4: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mod_def.aoc_9km ./mod_def.ww3 -4: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.aoc_9km ./out_grd.ww3 -4: + wave_grib2_sbs.sh[73]ngrib=1 -4: + wave_grib2_sbs.sh[74]dtgrib=3600 -4: + wave_grib2_sbs.sh[75]tstart='20210325 020000' -4: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 020000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.aoc_9km.inp.tmpl -4: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -4: $ WAVEWATCH-III gridded output input file -4: $ ---------------------------------------- -4: 20210325 020000 3600 1 -4: N -4: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -4: $ -4: 20210325 020000 7 11 255 0 20 -4: $ -4: 70 0 9.0 9.0 64 -4: $ 60 0 8.64919046313 8.64919046313 64 -4: $ end of input file -4: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[88]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.238224 ']' -4: ++ prep_step[4]echo gfs_ww3_grib.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[90]export err=0 -4: + wave_grib2_sbs.sh[90]err=0 -4: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[95]cat grib2_arctic_038.out -4: -4: *** WAVEWATCH III GRIB output postp. *** -4: ============================================== -4: -4: Comment character is '$' -4: -4: Grid name : Arctic Ocean PolarStereo 9km -4: -4: LINEIN: -4: 20210325 020000 3600 1 -4: -4: 20210325020000 3600 1 -4: GEN_PRO -99999 -4: -4: Output time data : -4: ----------------------------------------------------- -4: First time : 2021/03/25 02:00:00 UTC -4: Interval : 01:00:00 -4: Number of requests : 1 -4: Fields : Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: Requested output fields not yet available: -4: ----------------------------------------------------- -4: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -4: -4: Successfully requested output fields : -4: ----------------------------------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: -4: Additional GRIB parameters : -4: ----------------------------------------------------- -4: Run time : 2021/03/25 02:00:00 UTC -4: GRIB center ID : 7 -4: GRIB gen. proc. ID : 11 -4: GRIB grid ID : 255 -4: GRIB GDS parameter : 0 -4: Fields in file : -4: -------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: CHOSEN GRID TYPE: : POLARSTEREO -4: -4: -4: -4: Generating file -4: ----------------------------------------------------- -4: Data for 2021/03/25 02:00:00 UTC 0H forecast. -4: -4: End of program -4: ========================================= -4: WAVEWATCH III GRIB output -4: -4: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -4: + wave_grib2_sbs.sh[102][[ 38 -gt 0 ]] -4: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '38 hour fcst' -grib gfs.wave.t12z.arctic.9km.f038.grib2 -4: 1:0:d=2021032312:SPC:surface:38 hour fcst: -4: 2:191735:d=2021032312:DIRC:surface:38 hour fcst: -4: 3:661166:d=2021032312:UOGRD:surface:38 hour fcst: -4: 4:849990:d=2021032312:VOGRD:surface:38 hour fcst: -4: 5:1044250:d=2021032312:WIND:surface:38 hour fcst: -4: 6:1342054:d=2021032312:WDIR:surface:38 hour fcst: -4: 7:1764120:d=2021032312:UGRD:surface:38 hour fcst: -4: 8:2059457:d=2021032312:VGRD:surface:38 hour fcst: -4: 9:2356284:d=2021032312:ICEC:surface:38 hour fcst: -4: 10:2527724:d=2021032312:HTSGW:surface:38 hour fcst: -4: 11:2737228:d=2021032312:IMWF:surface:38 hour fcst: -4: 12:2950477:d=2021032312:MWSPER:surface:38 hour fcst: -4: 13:3165359:d=2021032312:PERPW:surface:38 hour fcst: -4: 14:3390252:d=2021032312:WWSDIR:surface:38 hour fcst: -4: 15:3688285:d=2021032312:DIRPW:surface:38 hour fcst: -4: 16:3997691:d=2021032312:WVHGT:surface:38 hour fcst: -4: 17:4204765:d=2021032312:SWELL:1 in sequence:38 hour fcst: -4: 18:4394768:d=2021032312:SWELL:2 in sequence:38 hour fcst: -4: 19:4539411:d=2021032312:SWELL:3 in sequence:38 hour fcst: -4: 20:4669648:d=2021032312:WVPER:surface:38 hour fcst: -4: 21:4888973:d=2021032312:SWPER:1 in sequence:38 hour fcst: -4: 22:5087788:d=2021032312:SWPER:2 in sequence:38 hour fcst: -4: 23:5238406:d=2021032312:SWPER:3 in sequence:38 hour fcst: -4: 24:5370707:d=2021032312:WVDIR:surface:38 hour fcst: -4: 25:5654031:d=2021032312:SWDIR:1 in sequence:38 hour fcst: -4: 26:5919118:d=2021032312:SWDIR:2 in sequence:38 hour fcst: -4: 27:6092211:d=2021032312:SWDIR:3 in sequence:38 hour fcst: -4: + wave_grib2_sbs.sh[104]err=0 -4: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.arctic.9km.f038.grib2 -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f038.grib2 ]] -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f038.grib2.idx ]] -4: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.arctic.9km.f038.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f038.grib2 -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f038.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f038.grib2 = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f038.grib2 ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f038.grib2 -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f038.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f038.grib2.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f038.grib2.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f038.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f038.grib2 -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.arctic.9km.f038.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f038.grib2.idx -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f038.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f038.grib2.idx = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f038.grib2.idx ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f038.grib2.idx -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f038.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f038.grib2.idx.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f038.grib2.idx.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f038.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f038.grib2.idx -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.arctic.9km.f038.grib2 and gfs.wave.t12z.arctic.9km.f038.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_aoc_9km to COM' -4: INFO: Copied gfs.wave.t12z.arctic.9km.f038.grib2 and gfs.wave.t12z.arctic.9km.f038.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_aoc_9km to COM -4: + wave_grib2_sbs.sh[130][[ aoc_9km == '' ]] -4: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -4: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.arctic.9km.f038.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -4: INFO: gfs.wave.t12z.arctic.9km.f038.grib2 is global.0p50 or SENDDBN is NO, no alert sent -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + cmdfile.7[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032502 3600. 9999 -6: + cmdfile.7[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grid_interp_sbs.sh[25]grdID=gsh_15m -6: + wave_grid_interp_sbs.sh[26]valid_time=2021032502 -6: + wave_grid_interp_sbs.sh[27]dt=3600. -6: + wave_grid_interp_sbs.sh[28]nst=9999 -6: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551 -6: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.uglo_100km ./out_grd.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mod_def.uglo_100km ./mod_def.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mod_def.gsh_15m ./mod_def.gsh_15m -6: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m'\''' -6: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m' -6: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[51]weights_found=1 -6: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 020000' -6: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 020000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gsh_15m/g ww3_gint.inp.tmpl -6: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -6: $ Input file for interpolation of uglo_100km to gsh_15m -6: $------------------------------------------------ -6: $ Start Time 3600. NSteps -6: 20210325 020000 3600. 9999 -6: $ Total number of grids -6: 2 -6: $ Grid extensions -6: 'uglo_100km' -6: 'gsh_15m' -6: $ -6: 0 -6: $ -6: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[70]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.238224 ']' -6: ++ prep_step[4]echo gfs_ww3_gint.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -6: INFO: Executing 'gfs_ww3_gint.x' -6: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[73]cat grid_interp.gsh_15m.out -6: -6: *** WAVEWATCH III Grid interpolation *** -6: =============================================== -6: -6: Comment character is '$' -6: -6: Time Information : -6: --------------------------------------------- -6: Starting Time : 2021/03/25 02:00:00 UTC -6: Interval (in sec) : 3600.00 -6: Number of requests : 9999 -6: --------------------------------------------- -6: Number of grids (including output grid) = 2 -6: -6: -6: Extension for grid 1 is --> uglo_100km -6: -6: Grid Particulars are : -6: Dimensions = 45166 1 -6: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Extension for grid 2 is --> gsh_15m -6: -6: Grid Particulars are : -6: Dimensions = 1440 277 -6: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -6: -6: -6: Preparing interpolation weights for output grid -6: Total number of wet points for interpolation 317192 -6: -6: -6: Variable: Grid Interpolation Map Units: 0.100E+01 -6: -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 277 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 241 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 205 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 169 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 133 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 97 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 61 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 | -6: 25 | 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 | -6: | | -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: -6: -6: Interpolating fields .... -6: -6: Output group 1 -6: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -6: Output group 2 -6: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -6: Output group 3 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 4 -6: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -6: Output group 5 -6: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -6: Output group 6 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 7 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 8 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 9 -6: Output variables skipped -6: Output group 10 -6: Output variables skipped -6: ------------------------------------------------ -6: 1Current vel. -6: 1Wind speed -6: 1Ice concentration -6: 2Wave height -6: 2Mean wave period(+2) -6: 2Mean wave period(+1) -6: 2Peak frequency -6: 2Mean wave dir. a1b1 -6: 2Peak direction -6: 4Part. wave height -6: 4Part. peak period -6: 4Part. mean direction -6: 5Charnock parameter -6: ------------------------------------------------ -6: OUTPUT TIME : 2021/03/25 02:00:00 UTC -6: -6: End of file reached -6: -6: -6: *** End of Grid interpolation Routine *** -6: =============================================== -6: -6: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -6: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -6: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_gsh_15m/out_grd.gsh_15m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.gsh_15m -6: + cmdfile.7[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032502 38 gsouth 0p25 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + cmdfile.7[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib2_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grib2_sbs.sh[30]grdID=gsh_15m -6: + wave_grib2_sbs.sh[31]GRIDNR=255 -6: + wave_grib2_sbs.sh[32]MODNR=11 -6: + wave_grib2_sbs.sh[33]valid_time=2021032502 -6: + wave_grib2_sbs.sh[34]fhr=38 -6: + wave_grib2_sbs.sh[35]grid_region=gsouth -6: + wave_grib2_sbs.sh[36]grid_res=0p25 -6: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551 -6: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_gsh_15m -6: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_gsh_15m -6: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_gsh_15m -6: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_gsh_15m -6: ++ wave_grib2_sbs.sh[47]printf %03i 38 -6: + wave_grib2_sbs.sh[47]FH3=038 -6: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -6: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.gsouth.0p25.f038.grib2 -6: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f038.grib2 ]] -6: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mod_def.gsh_15m ./mod_def.ww3 -6: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.gsh_15m ./out_grd.ww3 -6: + wave_grib2_sbs.sh[73]ngrib=1 -6: + wave_grib2_sbs.sh[74]dtgrib=3600 -6: + wave_grib2_sbs.sh[75]tstart='20210325 020000' -6: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 020000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gsh_15m.inp.tmpl -6: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -6: $ WAVEWATCH-III gridded output input file -6: $ ---------------------------------------- -6: 20210325 020000 3600 1 -6: N -6: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -6: $ -6: 20210325 020000 7 11 255 0 0 -6: $ -6: $ end of input file -6: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[88]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.238224 ']' -6: ++ prep_step[4]echo gfs_ww3_grib.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[90]export err=0 -6: + wave_grib2_sbs.sh[90]err=0 -6: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[95]cat grib2_gsouth_038.out -6: -6: *** WAVEWATCH III GRIB output postp. *** -6: ============================================== -6: -6: Comment character is '$' -6: -6: Grid name : GFSv16-wave S Hemisphere 1/4 d -6: -6: LINEIN: -6: 20210325 020000 3600 1 -6: -6: 20210325020000 3600 1 -6: GEN_PRO -99999 -6: -6: Output time data : -6: ----------------------------------------------------- -6: First time : 2021/03/25 02:00:00 UTC -6: Interval : 01:00:00 -6: Number of requests : 1 -6: Fields : Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: Requested output fields not yet available: -6: ----------------------------------------------------- -6: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -6: -6: Successfully requested output fields : -6: ----------------------------------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: -6: Additional GRIB parameters : -6: ----------------------------------------------------- -6: Run time : 2021/03/25 02:00:00 UTC -6: GRIB center ID : 7 -6: GRIB gen. proc. ID : 11 -6: GRIB grid ID : 255 -6: GRIB GDS parameter : 0 -6: Fields in file : -6: -------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: CHOSEN GRID TYPE: : LLRECTILINEAR -6: -6: -6: -6: Generating file -6: ----------------------------------------------------- -6: Data for 2021/03/25 02:00:00 UTC 0H forecast. -6: -6: End of program -6: ========================================= -6: WAVEWATCH III GRIB output -6: -6: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -6: + wave_grib2_sbs.sh[102][[ 38 -gt 0 ]] -6: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '38 hour fcst' -grib gfs.wave.t12z.gsouth.0p25.f038.grib2 -6: 1:0:d=2021032312:SPC:surface:38 hour fcst: -6: 2:118308:d=2021032312:DIRC:surface:38 hour fcst: -6: 3:503282:d=2021032312:UOGRD:surface:38 hour fcst: -6: 4:616728:d=2021032312:VOGRD:surface:38 hour fcst: -6: 5:739651:d=2021032312:WIND:surface:38 hour fcst: -6: 6:984092:d=2021032312:WDIR:surface:38 hour fcst: -6: 7:1364402:d=2021032312:UGRD:surface:38 hour fcst: -6: 8:1601533:d=2021032312:VGRD:surface:38 hour fcst: -6: 9:1844363:d=2021032312:ICEC:surface:38 hour fcst: -6: 10:1905631:d=2021032312:HTSGW:surface:38 hour fcst: -6: 11:2063117:d=2021032312:IMWF:surface:38 hour fcst: -6: 12:2232882:d=2021032312:MWSPER:surface:38 hour fcst: -6: 13:2404427:d=2021032312:PERPW:surface:38 hour fcst: -6: 14:2597120:d=2021032312:WWSDIR:surface:38 hour fcst: -6: 15:2948313:d=2021032312:DIRPW:surface:38 hour fcst: -6: 16:3322314:d=2021032312:WVHGT:surface:38 hour fcst: -6: 17:3491333:d=2021032312:SWELL:1 in sequence:38 hour fcst: -6: 18:3634760:d=2021032312:SWELL:2 in sequence:38 hour fcst: -6: 19:3721240:d=2021032312:SWELL:3 in sequence:38 hour fcst: -6: 20:3776948:d=2021032312:WVPER:surface:38 hour fcst: -6: 21:3971431:d=2021032312:SWPER:1 in sequence:38 hour fcst: -6: 22:4134520:d=2021032312:SWPER:2 in sequence:38 hour fcst: -6: 23:4237640:d=2021032312:SWPER:3 in sequence:38 hour fcst: -6: 24:4297801:d=2021032312:WVDIR:surface:38 hour fcst: -6: 25:4621094:d=2021032312:SWDIR:1 in sequence:38 hour fcst: -6: 26:4912613:d=2021032312:SWDIR:2 in sequence:38 hour fcst: -6: 27:5075356:d=2021032312:SWDIR:3 in sequence:38 hour fcst: -6: + wave_grib2_sbs.sh[104]err=0 -6: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.gsouth.0p25.f038.grib2 -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f038.grib2 ]] -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f038.grib2.idx ]] -6: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.gsouth.0p25.f038.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f038.grib2 -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f038.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f038.grib2 = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f038.grib2 ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f038.grib2 -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f038.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f038.grib2.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f038.grib2.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f038.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f038.grib2 -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.gsouth.0p25.f038.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f038.grib2.idx -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f038.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f038.grib2.idx = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f038.grib2.idx ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f038.grib2.idx -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f038.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f038.grib2.idx.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f038.grib2.idx.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f038.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f038.grib2.idx -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.gsouth.0p25.f038.grib2 and gfs.wave.t12z.gsouth.0p25.f038.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_gsh_15m to COM' -6: INFO: Copied gfs.wave.t12z.gsouth.0p25.f038.grib2 and gfs.wave.t12z.gsouth.0p25.f038.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_gsh_15m to COM -6: + wave_grib2_sbs.sh[130][[ gsh_15m == '' ]] -6: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -6: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.gsouth.0p25.f038.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -6: INFO: gfs.wave.t12z.gsouth.0p25.f038.grib2 is global.0p50 or SENDDBN is NO, no alert sent -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + cmdfile.6[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032502 3600. 9999 -5: + cmdfile.6[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grid_interp_sbs.sh[25]grdID=gnh_10m -5: + wave_grid_interp_sbs.sh[26]valid_time=2021032502 -5: + wave_grid_interp_sbs.sh[27]dt=3600. -5: + wave_grid_interp_sbs.sh[28]nst=9999 -5: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551 -5: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.uglo_100km ./out_grd.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mod_def.uglo_100km ./mod_def.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mod_def.gnh_10m ./mod_def.gnh_10m -5: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m'\''' -5: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m' -5: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[51]weights_found=1 -5: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 020000' -5: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 020000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gnh_10m/g ww3_gint.inp.tmpl -5: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -5: $ Input file for interpolation of uglo_100km to gnh_10m -5: $------------------------------------------------ -5: $ Start Time 3600. NSteps -5: 20210325 020000 3600. 9999 -5: $ Total number of grids -5: 2 -5: $ Grid extensions -5: 'uglo_100km' -5: 'gnh_10m' -5: $ -5: 0 -5: $ -5: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[70]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.238224 ']' -5: ++ prep_step[4]echo gfs_ww3_gint.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -5: INFO: Executing 'gfs_ww3_gint.x' -5: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[73]cat grid_interp.gnh_10m.out -5: -5: *** WAVEWATCH III Grid interpolation *** -5: =============================================== -5: -5: Comment character is '$' -5: -5: Time Information : -5: --------------------------------------------- -5: Starting Time : 2021/03/25 02:00:00 UTC -5: Interval (in sec) : 3600.00 -5: Number of requests : 9999 -5: --------------------------------------------- -5: Number of grids (including output grid) = 2 -5: -5: -5: Extension for grid 1 is --> uglo_100km -5: -5: Grid Particulars are : -5: Dimensions = 45166 1 -5: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Extension for grid 2 is --> gnh_10m -5: -5: Grid Particulars are : -5: Dimensions = 2160 406 -5: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -5: -5: -5: Preparing interpolation weights for output grid -5: Total number of wet points for interpolation 571209 -5: -5: -5: Variable: Grid Interpolation Map Units: 0.100E+01 -5: -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 406 | | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: 355 | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 304 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 253 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 202 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 151 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 100 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 49 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: -5: -5: Interpolating fields .... -5: -5: Output group 1 -5: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -5: Output group 2 -5: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -5: Output group 3 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 4 -5: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -5: Output group 5 -5: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -5: Output group 6 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 7 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 8 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 9 -5: Output variables skipped -5: Output group 10 -5: Output variables skipped -5: ------------------------------------------------ -5: 1Current vel. -5: 1Wind speed -5: 1Ice concentration -5: 2Wave height -5: 2Mean wave period(+2) -5: 2Mean wave period(+1) -5: 2Peak frequency -5: 2Mean wave dir. a1b1 -5: 2Peak direction -5: 4Part. wave height -5: 4Part. peak period -5: 4Part. mean direction -5: 5Charnock parameter -5: ------------------------------------------------ -5: OUTPUT TIME : 2021/03/25 02:00:00 UTC -5: -5: End of file reached -5: -5: -5: *** End of Grid interpolation Routine *** -5: =============================================== -5: -5: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -5: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -5: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grid_interp_gnh_10m/out_grd.gnh_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.gnh_10m -5: + cmdfile.6[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032502 38 global 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + cmdfile.6[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib2_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grib2_sbs.sh[30]grdID=gnh_10m -5: + wave_grib2_sbs.sh[31]GRIDNR=255 -5: + wave_grib2_sbs.sh[32]MODNR=11 -5: + wave_grib2_sbs.sh[33]valid_time=2021032502 -5: + wave_grib2_sbs.sh[34]fhr=38 -5: + wave_grib2_sbs.sh[35]grid_region=global -5: + wave_grib2_sbs.sh[36]grid_res=0p16 -5: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551 -5: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_gnh_10m -5: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_gnh_10m -5: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_gnh_10m -5: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_gnh_10m -5: ++ wave_grib2_sbs.sh[47]printf %03i 38 -5: + wave_grib2_sbs.sh[47]FH3=038 -5: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p16 -5: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p16.f038.grib2 -5: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f038.grib2 ]] -5: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/mod_def.gnh_10m ./mod_def.ww3 -5: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/out_grd.gnh_10m ./out_grd.ww3 -5: + wave_grib2_sbs.sh[73]ngrib=1 -5: + wave_grib2_sbs.sh[74]dtgrib=3600 -5: + wave_grib2_sbs.sh[75]tstart='20210325 020000' -5: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 020000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gnh_10m.inp.tmpl -5: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -5: $ WAVEWATCH-III gridded output input file -5: $ ---------------------------------------- -5: 20210325 020000 3600 1 -5: N -5: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -5: $ -5: 20210325 020000 7 11 255 0 0 -5: $ -5: $ end of input file -5: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[88]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.238224 ']' -5: ++ prep_step[4]echo gfs_ww3_grib.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[90]export err=0 -5: + wave_grib2_sbs.sh[90]err=0 -5: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[95]cat grib2_global_038.out -5: -5: *** WAVEWATCH III GRIB output postp. *** -5: ============================================== -5: -5: Comment character is '$' -5: -5: Grid name : GFSv16-wave N Hemisphere 1/6 d -5: -5: LINEIN: -5: 20210325 020000 3600 1 -5: -5: 20210325020000 3600 1 -5: GEN_PRO -99999 -5: -5: Output time data : -5: ----------------------------------------------------- -5: First time : 2021/03/25 02:00:00 UTC -5: Interval : 01:00:00 -5: Number of requests : 1 -5: Fields : Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: Requested output fields not yet available: -5: ----------------------------------------------------- -5: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -5: -5: Successfully requested output fields : -5: ----------------------------------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: -5: Additional GRIB parameters : -5: ----------------------------------------------------- -5: Run time : 2021/03/25 02:00:00 UTC -5: GRIB center ID : 7 -5: GRIB gen. proc. ID : 11 -5: GRIB grid ID : 255 -5: GRIB GDS parameter : 0 -5: Fields in file : -5: -------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: CHOSEN GRID TYPE: : LLRECTILINEAR -5: -5: -5: -5: Generating file -5: ----------------------------------------------------- -5: Data for 2021/03/25 02:00:00 UTC 0H forecast. -5: -5: End of program -5: ========================================= -5: WAVEWATCH III GRIB output -5: -5: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -5: + wave_grib2_sbs.sh[102][[ 38 -gt 0 ]] -5: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '38 hour fcst' -grib gfs.wave.t12z.global.0p16.f038.grib2 -5: 1:0:d=2021032312:SPC:surface:38 hour fcst: -5: 2:235360:d=2021032312:DIRC:surface:38 hour fcst: -5: 3:860067:d=2021032312:UOGRD:surface:38 hour fcst: -5: 4:1090347:d=2021032312:VOGRD:surface:38 hour fcst: -5: 5:1327926:d=2021032312:WIND:surface:38 hour fcst: -5: 6:1728310:d=2021032312:WDIR:surface:38 hour fcst: -5: 7:2375930:d=2021032312:UGRD:surface:38 hour fcst: -5: 8:2772786:d=2021032312:VGRD:surface:38 hour fcst: -5: 9:3173230:d=2021032312:ICEC:surface:38 hour fcst: -5: 10:3287969:d=2021032312:HTSGW:surface:38 hour fcst: -5: 11:3569497:d=2021032312:IMWF:surface:38 hour fcst: -5: 12:3877948:d=2021032312:MWSPER:surface:38 hour fcst: -5: 13:4188892:d=2021032312:PERPW:surface:38 hour fcst: -5: 14:4524803:d=2021032312:WWSDIR:surface:38 hour fcst: -5: 15:5134934:d=2021032312:DIRPW:surface:38 hour fcst: -5: 16:5786035:d=2021032312:WVHGT:surface:38 hour fcst: -5: 17:6069264:d=2021032312:SWELL:1 in sequence:38 hour fcst: -5: 18:6293095:d=2021032312:SWELL:2 in sequence:38 hour fcst: -5: 19:6442963:d=2021032312:SWELL:3 in sequence:38 hour fcst: -5: 20:6558767:d=2021032312:WVPER:surface:38 hour fcst: -5: 21:6885926:d=2021032312:SWPER:1 in sequence:38 hour fcst: -5: 22:7142833:d=2021032312:SWPER:2 in sequence:38 hour fcst: -5: 23:7313746:d=2021032312:SWPER:3 in sequence:38 hour fcst: -5: 24:7434315:d=2021032312:WVDIR:surface:38 hour fcst: -5: 25:7983957:d=2021032312:SWDIR:1 in sequence:38 hour fcst: -5: 26:8421302:d=2021032312:SWDIR:2 in sequence:38 hour fcst: -5: 27:8668824:d=2021032312:SWDIR:3 in sequence:38 hour fcst: -5: + wave_grib2_sbs.sh[104]err=0 -5: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p16.f038.grib2 -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f038.grib2 ]] -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f038.grib2.idx ]] -5: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p16.f038.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f038.grib2 -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f038.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f038.grib2 = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f038.grib2 ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f038.grib2 -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f038.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f038.grib2.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f038.grib2.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f038.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f038.grib2 -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p16.f038.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f038.grib2.idx -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f038.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f038.grib2.idx = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f038.grib2.idx ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f038.grib2.idx -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f038.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f038.grib2.idx.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f038.grib2.idx.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f038.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f038.grib2.idx -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p16.f038.grib2 and gfs.wave.t12z.global.0p16.f038.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_gnh_10m to COM' -5: INFO: Copied gfs.wave.t12z.global.0p16.f038.grib2 and gfs.wave.t12z.global.0p16.f038.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551/grib_gnh_10m to COM -5: + wave_grib2_sbs.sh[130][[ gnh_10m == '' ]] -5: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -5: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p16.f038.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -5: INFO: gfs.wave.t12z.global.0p16.f038.grib2 is global.0p50 or SENDDBN is NO, no alert sent -+ run_mpmd.sh[113]exit 0 -+ run_mpmd.sh[1]postamble run_mpmd.sh 1753758322 0 -+ preamble.sh[62]set +x -End run_mpmd.sh at 03:05:32 with error code 0 (time elapsed: 00:00:10) -+ exgfs_wave_post_gridded_sbs.sh[122]true -+ exgfs_wave_post_gridded_sbs.sh[123]export err=0 -+ exgfs_wave_post_gridded_sbs.sh[123]err=0 -+ exgfs_wave_post_gridded_sbs.sh[124][[ 0 -ne 0 ]] -+ exgfs_wave_post_gridded_sbs.sh[130]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ exgfs_wave_post_gridded_sbs.sh[131]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ exgfs_wave_post_gridded_sbs.sh[132]gribchk=gfs.wave.t12z.gsouth.0p25.f038.grib2 -+ exgfs_wave_post_gridded_sbs.sh[133][[ ! -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f038.grib2 ]] -+ exgfs_wave_post_gridded_sbs.sh[138]exit 0 -+ JGLOBAL_WAVE_POST_SBS[28]true -+ JGLOBAL_WAVE_POST_SBS[29]export err=0 -+ JGLOBAL_WAVE_POST_SBS[29]err=0 -+ JGLOBAL_WAVE_POST_SBS[30][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[37]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312 -+ JGLOBAL_WAVE_POST_SBS[38][[ NO != \Y\E\S ]] -+ JGLOBAL_WAVE_POST_SBS[39]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f038.231551 -+ JGLOBAL_WAVE_POST_SBS[42]exit 0 -+ JGLOBAL_WAVE_POST_SBS[1]postamble /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS 1753758311 0 -+ preamble.sh[62]set +x -End /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at 03:05:33 with error code 0 (time elapsed: 00:00:22) diff --git a/ci/error_logs/HECULES_PR_LOGS/C48_S2SW_logs_2021032312_gfs_wavepostsbs_f057-f063.log b/ci/error_logs/HECULES_PR_LOGS/C48_S2SW_logs_2021032312_gfs_wavepostsbs_f057-f063.log deleted file mode 100644 index 529b346e..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/C48_S2SW_logs_2021032312_gfs_wavepostsbs_f057-f063.log +++ /dev/null @@ -1,12866 +0,0 @@ -+ source /work2/noaa/global/mterry/global-workflow_forked/ush/load_ufswm_modules.sh -++ [[ NO == \N\O ]] -++ echo 'Loading modules quietly...' -Loading modules quietly... -++ set +x -Running "module reset". Resetting modules to system default. The following $MODULEPATH directories have been removed: None - -Currently Loaded Modules: - 1) contrib/0.1 25) git-lfs/3.1.2 - 2) intel-oneapi-compilers/2023.1.0 26) crtm/2.4.0.1 - 3) stack-intel/2021.9.0 27) g2/3.5.1 - 4) intel-oneapi-mpi/2021.9.0 28) g2tmpl/1.13.0 - 5) stack-intel-oneapi-mpi/2021.9.0 29) ip/4.3.0 - 6) nghttp2/1.57.0 30) sp/2.5.0 - 7) curl/8.4.0 31) w3emc/2.10.0 - 8) cmake/3.23.1 32) gftl/1.10.0 - 9) libjpeg/2.1.0 33) gftl-shared/1.6.1 - 10) jasper/2.0.32 34) fargparse/1.5.0 - 11) zlib/1.2.13 35) pigz/2.7 - 12) libpng/1.6.37 36) tar/1.34 - 13) hdf5/1.14.0 37) gettext/0.21.1 - 14) snappy/1.1.10 38) libxcrypt/4.4.35 - 15) zstd/1.5.2 39) sqlite/3.43.2 - 16) c-blosc/1.21.5 40) util-linux-uuid/2.38.1 - 17) netcdf-c/4.9.2 41) python/3.10.13 - 18) netcdf-fortran/4.6.1 42) mapl/2.40.3-esmf-8.6.0 - 19) parallel-netcdf/1.12.2 43) scotch/7.0.4 - 20) parallelio/2.5.10 44) ufs_common - 21) esmf/8.6.0 45) nccmp/1.9.0.1 - 22) fms/2024.01.02 46) ufs_hercules.intel - 23) bacio/2.4.1 47) prod_util/2.1.1 - 24) crtm-fix/2.4.0.1_emc 48) wgrib2/3.1.1 - - - -Begin /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at Tue Jul 29 03:04:26 UTC 2025 -++ jjob_header.sh[46]OPTIND=1 -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[50]env_job=wavepostsbs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[49]read -ra configs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[61]shift 4 -++ jjob_header.sh[63][[ -z wavepostsbs ]] -++ jjob_header.sh[71]export DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463 -++ jjob_header.sh[71]DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463 -++ jjob_header.sh[72][[ YES == \Y\E\S ]] -++ jjob_header.sh[73]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463 -++ jjob_header.sh[75]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463 -++ jjob_header.sh[76]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463 -++ jjob_header.sh[85]export pid=454158 -++ jjob_header.sh[85]pid=454158 -++ jjob_header.sh[86]export pgmout=OUTPUT.454158 -++ jjob_header.sh[86]pgmout=OUTPUT.454158 -++ jjob_header.sh[87]export pgmerr=errfile -++ jjob_header.sh[87]pgmerr=errfile -++ jjob_header.sh[90]export pgm= -++ jjob_header.sh[90]pgm= -++ jjob_header.sh[96]export cycle=t12z -++ jjob_header.sh[96]cycle=t12z -++ jjob_header.sh[97]setpdy.sh -+ setpdy.sh[20]'[' /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463 == /home/mterry ']' -+ setpdy.sh[25][[ ! t12z =~ t??z ]] -+ setpdy.sh[30]case $# in -+ setpdy.sh[31]dates_before_PDY=7 -+ setpdy.sh[32]dates_after_PDY=7 -+ setpdy.sh[50]COMDATEROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+ setpdy.sh[53]'[' -z 20210323 ']' -+ setpdy.sh[57]sed 's/[0-9]\{8\}/20210323/' /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z -sed: can't read /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z: No such file or directory -++ jjob_header.sh[97]true -++ jjob_header.sh[98]source ./PDY -/work2/noaa/global/mterry/global-workflow_forked/ush/jjob_header.sh: line 98: ./PDY: No such file or directory -++ jjob_header.sh[98]true -++ jjob_header.sh[104]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[104]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.base -+++ config.base[6]echo 'BEGIN: config.base' -BEGIN: config.base -+++ config.base[9]export machine=HERCULES -+++ config.base[9]machine=HERCULES -+++ config.base[12]export RUN_ENVIR=emc -+++ config.base[12]RUN_ENVIR=emc -+++ config.base[15]export ACCOUNT=fv3-cpu -+++ config.base[15]ACCOUNT=fv3-cpu -+++ config.base[16]export QUEUE=batch -+++ config.base[16]QUEUE=batch -+++ config.base[17]export QUEUE_SERVICE=batch -+++ config.base[17]QUEUE_SERVICE=batch -+++ config.base[18]export QUEUE_DTN=batch -+++ config.base[18]QUEUE_DTN=batch -+++ config.base[19]export PARTITION_BATCH=hercules -+++ config.base[19]PARTITION_BATCH=hercules -+++ config.base[20]export PARTITION_SERVICE=service -+++ config.base[20]PARTITION_SERVICE=service -+++ config.base[21]export PARTITION_DTN= -+++ config.base[21]PARTITION_DTN= -+++ config.base[22]export RESERVATION= -+++ config.base[22]RESERVATION= -+++ config.base[23]export CLUSTERS= -+++ config.base[23]CLUSTERS= -+++ config.base[24]export CLUSTERS_SERVICE= -+++ config.base[24]CLUSTERS_SERVICE= -+++ config.base[25]export CLUSTERS_DTN= -+++ config.base[25]CLUSTERS_DTN= -+++ config.base[28]export HPSS_PROJECT=emc-global -+++ config.base[28]HPSS_PROJECT=emc-global -+++ config.base[31]export HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[31]HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[32]export EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[32]EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[33]export FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[33]FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[34]export PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[34]PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[35]export SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[35]SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[36]export USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[36]USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[38]export FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[38]FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[39]export FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[39]FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[40]export FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[40]FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[41]export FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[41]FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[42]export FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[42]FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[43]export FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[43]FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[44]export FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[44]FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[45]export FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[45]FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[50]export PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[50]PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[51]export COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[51]COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[52]export COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[52]COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[53]export DMPDIR=/work/noaa/rstprod/dump -+++ config.base[53]DMPDIR=/work/noaa/rstprod/dump -+++ config.base[57]export COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[57]COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[58]export COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[58]COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[59]export COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[59]COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[62]export HOMEDIR=/work2/noaa/global/mterry -+++ config.base[62]HOMEDIR=/work2/noaa/global/mterry -+++ config.base[63]export STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[63]STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]export PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[65]export NOSCRUB=/work2/noaa/global/mterry -+++ config.base[65]NOSCRUB=/work2/noaa/global/mterry -+++ config.base[68]export BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[68]BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[71]export BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[71]BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[74]export DO_PREP_SFC=NO -+++ config.base[74]DO_PREP_SFC=NO -+++ config.base[77]export DO_GOES=NO -+++ config.base[77]DO_GOES=NO -+++ config.base[78]export DO_BUFRSND=NO -+++ config.base[78]DO_BUFRSND=NO -+++ config.base[79]export DO_GEMPAK=NO -+++ config.base[79]DO_GEMPAK=NO -+++ config.base[80]export DO_AWIPS=NO -+++ config.base[80]DO_AWIPS=NO -+++ config.base[81]export DO_NPOESS=NO -+++ config.base[81]DO_NPOESS=NO -+++ config.base[82]export DO_TRACKER=YES -+++ config.base[82]DO_TRACKER=YES -+++ config.base[83]export DO_GENESIS=YES -+++ config.base[83]DO_GENESIS=YES -+++ config.base[84]export DO_GENESIS_FSU=NO -+++ config.base[84]DO_GENESIS_FSU=NO -+++ config.base[85]export DO_VERFOZN=YES -+++ config.base[85]DO_VERFOZN=YES -+++ config.base[86]export DO_VERFRAD=YES -+++ config.base[86]DO_VERFRAD=YES -+++ config.base[87]export DO_VMINMON=YES -+++ config.base[87]DO_VMINMON=YES -+++ config.base[88]export DO_ANLSTAT=NO -+++ config.base[88]DO_ANLSTAT=NO -+++ config.base[91]export MODE=forecast-only -+++ config.base[91]MODE=forecast-only -+++ config.base[92]export DO_TEST_MODE=YES -+++ config.base[92]DO_TEST_MODE=YES -+++ config.base[101]export FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[101]FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[102]export HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[102]HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[103]export HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[103]HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[106]export NMV=/bin/mv -+++ config.base[106]NMV=/bin/mv -+++ config.base[107]export 'NLN=/bin/ln -sf' -+++ config.base[107]NLN='/bin/ln -sf' -+++ config.base[108]export VERBOSE=YES -+++ config.base[108]VERBOSE=YES -+++ config.base[109]export KEEPDATA=NO -+++ config.base[109]KEEPDATA=NO -+++ config.base[110]export DEBUG_POSTSCRIPT=NO -+++ config.base[110]DEBUG_POSTSCRIPT=NO -+++ config.base[111]export CHGRP_RSTPROD=YES -+++ config.base[111]CHGRP_RSTPROD=YES -+++ config.base[112]export 'CHGRP_CMD=chgrp rstprod' -+++ config.base[112]CHGRP_CMD='chgrp rstprod' -+++ config.base[113]export NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[113]NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[114]export NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[114]NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[117]export BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[117]BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[120]export SDATE=2021032312 -+++ config.base[120]SDATE=2021032312 -+++ config.base[121]export EDATE=2021032312 -+++ config.base[121]EDATE=2021032312 -+++ config.base[122]export EXP_WARM_START=.false. -+++ config.base[122]EXP_WARM_START=.false. -+++ config.base[123]export assim_freq=6 -+++ config.base[123]assim_freq=6 -+++ config.base[124]export PSLOT=C48_S2SW -+++ config.base[124]PSLOT=C48_S2SW -+++ config.base[125]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[125]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[126]export ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[126]ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[127]export DUMP_SUFFIX= -+++ config.base[127]DUMP_SUFFIX= -+++ config.base[128][[ 2021032312 -ge 2019092100 ]] -+++ config.base[128][[ 2021032312 -le 2019110700 ]] -+++ config.base[131]export ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[131]ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[132]export ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[132]ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[133]export FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[133]FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[136]export envir=prod -+++ config.base[136]envir=prod -+++ config.base[137]export NET=gfs -+++ config.base[137]NET=gfs -+++ config.base[138]export RUN=gfs -+++ config.base[138]RUN=gfs -+++ config.base[141]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.com -++++ config.com[4]echo 'BEGIN: config.com' -BEGIN: config.com -++++ config.com[38][[ emc == \n\c\o ]] -++++ config.com[43]COM_OBSPROC_TMPL='${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos' -++++ config.com[44]COM_RTOFS_TMPL='${DMPDIR}' -++++ config.com[45]COM_TCVITAL_TMPL='${DMPDIR}/${RUN}.${YMD}/${HH}/atmos' -++++ config.com[47]declare -rx 'COM_OBS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/obs' -++++ config.com[48]declare -rx COM_OBSPROC_TMPL COM_RTOFS_TMPL -++++ config.com[50]COM_BASE='${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}' -++++ config.com[52]declare -rx 'COM_TOP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}' -++++ config.com[54]declare -rx 'COM_CONF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf' -++++ config.com[55]declare -rx 'COM_OBS_JEDI=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi' -++++ config.com[57]declare -rx 'COM_ATMOS_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input' -++++ config.com[58]declare -rx 'COM_ATMOS_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart' -++++ config.com[59]declare -rx 'COM_ATMOS_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos' -++++ config.com[60]declare -rx 'COM_SNOW_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow' -++++ config.com[61]declare -rx 'COM_SNOW_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon' -++++ config.com[62]declare -rx 'COM_ATMOS_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history' -++++ config.com[63]declare -rx 'COM_ATMOS_MASTER_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master' -++++ config.com[64]declare -rx 'COM_ATMOS_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2' -++++ config.com[65]declare -rx 'COM_ATMOS_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}' -++++ config.com[66]declare -rx 'COM_ATMOS_BUFR_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr' -++++ config.com[67]declare -rx 'COM_ATMOS_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}' -++++ config.com[68]declare -rx 'COM_ATMOS_GENESIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital' -++++ config.com[69]declare -rx 'COM_ATMOS_TRACK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks' -++++ config.com[70]declare -rx 'COM_ATMOS_GOES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim' -++++ config.com[71]declare -rx 'COM_ATMOS_IMAGERY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery' -++++ config.com[72]declare -rx 'COM_ATMOS_OZNMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon' -++++ config.com[73]declare -rx 'COM_ATMOS_RADMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon' -++++ config.com[74]declare -rx 'COM_ATMOS_MINMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon' -++++ config.com[75]declare -rx 'COM_ATMOS_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon' -++++ config.com[76]declare -rx 'COM_ATMOS_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo' -++++ config.com[78]declare -rx 'COM_WAVE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart' -++++ config.com[79]declare -rx 'COM_WAVE_PREP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep' -++++ config.com[80]declare -rx 'COM_WAVE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history' -++++ config.com[81]declare -rx 'COM_WAVE_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded' -++++ config.com[82]declare -rx 'COM_WAVE_GRID_RES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}' -++++ config.com[83]declare -rx 'COM_WAVE_STATION_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station' -++++ config.com[84]declare -rx 'COM_WAVE_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak' -++++ config.com[85]declare -rx 'COM_WAVE_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo' -++++ config.com[87]declare -rx 'COM_OCEAN_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history' -++++ config.com[88]declare -rx 'COM_OCEAN_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart' -++++ config.com[89]declare -rx 'COM_OCEAN_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input' -++++ config.com[90]declare -rx 'COM_OCEAN_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean' -++++ config.com[91]declare -rx 'COM_OCEAN_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon' -++++ config.com[92]declare -rx 'COM_OCEAN_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf' -++++ config.com[93]declare -rx 'COM_OCEAN_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean' -++++ config.com[94]declare -rx 'COM_OCEAN_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf' -++++ config.com[95]declare -rx 'COM_OCEAN_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2' -++++ config.com[96]declare -rx 'COM_OCEAN_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}' -++++ config.com[98]declare -rx 'COM_ICE_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice' -++++ config.com[99]declare -rx 'COM_ICE_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf' -++++ config.com[100]declare -rx 'COM_ICE_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon' -++++ config.com[101]declare -rx 'COM_ICE_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice' -++++ config.com[102]declare -rx 'COM_ICE_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input' -++++ config.com[103]declare -rx 'COM_ICE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history' -++++ config.com[104]declare -rx 'COM_ICE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart' -++++ config.com[105]declare -rx 'COM_ICE_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf' -++++ config.com[106]declare -rx 'COM_ICE_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2' -++++ config.com[107]declare -rx 'COM_ICE_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}' -++++ config.com[109]declare -rx 'COM_CHEM_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history' -++++ config.com[110]declare -rx 'COM_CHEM_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem' -++++ config.com[111]declare -rx 'COM_CHEM_BMAT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix' -++++ config.com[112]declare -rx 'COM_CHEM_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon' -++++ config.com[114]declare -rx 'COM_MED_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart' -+++ config.base[143]export LOGSCRIPT= -+++ config.base[143]LOGSCRIPT= -+++ config.base[145]export 'REDOUT=1>' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wave -+++ config.wave[6]echo 'BEGIN: config.wave' -BEGIN: config.wave -+++ config.wave[13]export RUNRSTwave=gdas -+++ config.wave[13]RUNRSTwave=gdas -+++ config.wave[16]export MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[16]MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[19]case "${waveGRD}" in -+++ config.wave[64]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[64]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[65]export wavepostGRD= -+++ config.wave[65]wavepostGRD= -+++ config.wave[66]export waveuoutpGRD=uglo_100km -+++ config.wave[66]waveuoutpGRD=uglo_100km -+++ config.wave[75]export WAVEWND_DID= -+++ config.wave[75]WAVEWND_DID= -+++ config.wave[76]export WAVEWND_FID= -+++ config.wave[76]WAVEWND_FID= -+++ config.wave[79][[ gfs == \g\f\s ]] -+++ config.wave[80]export FHMAX_WAV=120 -+++ config.wave[80]FHMAX_WAV=120 -+++ config.wave[82]export WAVHINDH=0 -+++ config.wave[82]WAVHINDH=0 -+++ config.wave[83]export FHMAX_WAV_IBP=180 -+++ config.wave[83]FHMAX_WAV_IBP=180 -+++ config.wave[84](( FHMAX_WAV < FHMAX_WAV_IBP )) -+++ config.wave[84]export FHMAX_WAV_IBP=120 -+++ config.wave[84]FHMAX_WAV_IBP=120 -+++ config.wave[87]export DTFLD_WAV=3600 -+++ config.wave[87]DTFLD_WAV=3600 -+++ config.wave[88]export DTPNT_WAV=3600 -+++ config.wave[88]DTPNT_WAV=3600 -+++ config.wave[89]export FHINCP_WAV=1 -+++ config.wave[89]FHINCP_WAV=1 -+++ config.wave[92]export 'OUTPARS_WAV=WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[92]OUTPARS_WAV='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[95][[ gfs == \g\d\a\s ]] -+++ config.wave[99](( INTERVAL_GFS > 0 )) -+++ config.wave[100]export WAVHCYC=6 -+++ config.wave[100]WAVHCYC=6 -+++ config.wave[101]export FHMAX_WAV_CUR=192 -+++ config.wave[101]FHMAX_WAV_CUR=192 -+++ config.wave[109]export RSTTYPE_WAV=T -+++ config.wave[109]RSTTYPE_WAV=T -+++ config.wave[110][[ gfs != gfs ]] -+++ config.wave[115]rst_dt_gfs=43200 -+++ config.wave[116][[ 43200 -gt 0 ]] -+++ config.wave[117]export DT_1_RST_WAV=0 -+++ config.wave[117]DT_1_RST_WAV=0 -+++ config.wave[120]export DT_2_RST_WAV=43200 -+++ config.wave[120]DT_2_RST_WAV=43200 -+++ config.wave[126]export RSTIOFF_WAV=0 -+++ config.wave[126]RSTIOFF_WAV=0 -+++ config.wave[131]export RUNMEM=-1 -+++ config.wave[131]RUNMEM=-1 -+++ config.wave[134](( RUNMEM == -1 )) -+++ config.wave[136]export waveMEMB= -+++ config.wave[136]waveMEMB= -+++ config.wave[143]export WW3ATMINP=CPL -+++ config.wave[143]WW3ATMINP=CPL -+++ config.wave[144][[ YES == \Y\E\S ]] -+++ config.wave[145]export WW3ICEINP=CPL -+++ config.wave[145]WW3ICEINP=CPL -+++ config.wave[146]export WAVEICE_FID= -+++ config.wave[146]WAVEICE_FID= -+++ config.wave[152][[ YES == \Y\E\S ]] -+++ config.wave[153]export WW3CURINP=CPL -+++ config.wave[153]WW3CURINP=CPL -+++ config.wave[154]export WAVECUR_FID= -+++ config.wave[154]WAVECUR_FID= -+++ config.wave[161]export WW3ATMIENS=F -+++ config.wave[161]WW3ATMIENS=F -+++ config.wave[162]export WW3ICEIENS=F -+++ config.wave[162]WW3ICEIENS=F -+++ config.wave[163]export WW3CURIENS=F -+++ config.wave[163]WW3CURIENS=F -+++ config.wave[165]export GOFILETYPE=1 -+++ config.wave[165]GOFILETYPE=1 -+++ config.wave[166]export POFILETYPE=1 -+++ config.wave[166]POFILETYPE=1 -+++ config.wave[170]export FUNIPNT=T -+++ config.wave[170]FUNIPNT=T -+++ config.wave[172]export IOSRV=1 -+++ config.wave[172]IOSRV=1 -+++ config.wave[174]export FPNTPROC=T -+++ config.wave[174]FPNTPROC=T -+++ config.wave[176]export FGRDPROC=F -+++ config.wave[176]FGRDPROC=F -+++ config.wave[178]export FLAGMASKCOMP=F -+++ config.wave[178]FLAGMASKCOMP=F -+++ config.wave[180]export FLAGMASKOUT=F -+++ config.wave[180]FLAGMASKOUT=F -+++ config.wave[182]echo 'END: config.wave' -END: config.wave -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wavepostsbs -+++ config.wavepostsbs[6]echo 'BEGIN: config.wavepostsbs' -BEGIN: config.wavepostsbs -+++ config.wavepostsbs[9]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources wavepostsbs -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=wavepostsbs -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[156]ntasks=8 -++++ config.resources[157]threads_per_task=1 -++++ config.resources[158]tasks_per_node=80 -++++ config.resources[159]NTASKS=8 -++++ config.resources[160]memory=20GB -++++ config.resources[161]walltime=00:15:00 -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export NTASKS -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export memory -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.wavepostsbs[12]export MAX_TASKS=25 -+++ config.wavepostsbs[12]MAX_TASKS=25 -+++ config.wavepostsbs[15]export WAV_SUBGRBSRC= -+++ config.wavepostsbs[15]WAV_SUBGRBSRC= -+++ config.wavepostsbs[16]export WAV_SUBGRB= -+++ config.wavepostsbs[16]WAV_SUBGRB= -+++ config.wavepostsbs[19]export DOFLD_WAV=YES -+++ config.wavepostsbs[19]DOFLD_WAV=YES -+++ config.wavepostsbs[20]export DOPNT_WAV=YES -+++ config.wavepostsbs[20]DOPNT_WAV=YES -+++ config.wavepostsbs[21]export DOGRB_WAV=YES -+++ config.wavepostsbs[21]DOGRB_WAV=YES -+++ config.wavepostsbs[22][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+++ config.wavepostsbs[23]export DOGRI_WAV=YES -+++ config.wavepostsbs[23]DOGRI_WAV=YES -+++ config.wavepostsbs[27]export DOSPC_WAV=YES -+++ config.wavepostsbs[27]DOSPC_WAV=YES -+++ config.wavepostsbs[28]export DOBLL_WAV=YES -+++ config.wavepostsbs[28]DOBLL_WAV=YES -+++ config.wavepostsbs[30]echo 'END: config.wavepostsbs' -END: config.wavepostsbs -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env wavepostsbs -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=wavepostsbs -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 20971520 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 8 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[34]max_threads_per_task=1 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 1 ]] -+++ HERCULES.env[40][[ 1 -gt 1 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 8' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[63]export USE_CFP=YES -+++ HERCULES.env[63]USE_CFP=YES -+++ HERCULES.env[64][[ wavepostsbs == \w\a\v\e\p\r\e\p ]] -+++ HERCULES.env[65]export 'wavempexec=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[65]wavempexec='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[66]export 'wave_mpmd=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[66]wave_mpmd='--multi-prog --output=mpmd.%j.%t.out' -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[4]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ JGLOBAL_WAVE_POST_SBS[7]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[7]HH=12 -+ JGLOBAL_WAVE_POST_SBS[7]declare_from_tmpl -rx COMIN_WAVE_PREP:COM_WAVE_PREP_TMPL COMIN_WAVE_HISTORY:COM_WAVE_HISTORY_TMPL COMOUT_WAVE_GRID:COM_WAVE_GRID_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_WAVE_PREP=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep -declare_from_tmpl :: COMIN_WAVE_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history -declare_from_tmpl :: COMOUT_WAVE_GRID=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded -+ JGLOBAL_WAVE_POST_SBS[12][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n '' ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_atlocn_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_atlocn_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_atlocn_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_epacif_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_epacif_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_epacif_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_wcoast_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_wcoast_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_wcoast_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p50 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p50:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p50=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p50 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=arctic.9km -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_arctic_9km:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_arctic_9km=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_arctic_9km -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_gsouth_0p25:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_gsouth_0p25=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 ]] -+ JGLOBAL_WAVE_POST_SBS[28]/work2/noaa/global/mterry/global-workflow_forked/scripts/exgfs_wave_post_gridded_sbs.sh -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ exgfs_wave_post_gridded_sbs.sh[24]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ exgfs_wave_post_gridded_sbs.sh[26]DOGRI_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[27]DOGRB_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[29]export waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[29]waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[30]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[30]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[31]export wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[31]wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[33]cat - INFO: Grid information: - INFO: Native wave grids: uglo_100km - INFO: Interpolated grids: at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m - INFO: Post-process grids: -++ exgfs_wave_post_gridded_sbs.sh[40]printf %03i 57 -+ exgfs_wave_post_gridded_sbs.sh[40]fhr3=057 -++ exgfs_wave_post_gridded_sbs.sh[41]date -u -d '20210323 12 + 57 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[41]valid_time=2021032521 -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[49]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f057.bin ./out_grd.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f057.bin ./out_grd.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[52][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[62]cat - INFO: Summary: - INFO: Grid interp: DOGRI_WAV="YES" - INFO: Grib files: DOGRB_WAV="YES" - INFO: Fields to be included in grib files: - INFO: OUTPARS_WAV="WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA" -+ exgfs_wave_post_gridded_sbs.sh[70][[ YES == \N\O ]] -+ exgfs_wave_post_gridded_sbs.sh[76]rm -f 'cmdfile.*' cmdfile -+ exgfs_wave_post_gridded_sbs.sh[77]count=0 -+ exgfs_wave_post_gridded_sbs.sh[80][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[81]dt_int=3600. -+ exgfs_wave_post_gridded_sbs.sh[82]n_int=9999 -++ exgfs_wave_post_gridded_sbs.sh[83]date -u -d '20210325 21 - 0 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[83]ymdh_int=2021032521 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=1 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032521 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032521 57 atlocn 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib2_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib2_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=2 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032521 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032521 57 epacif 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib2_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib2_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=3 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032521 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032521 57 wcoast 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib2_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib2_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=4 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032521 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032521 57 global 0p50 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib2_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib2_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=5 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032521 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032521 57 arctic 9km '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib2_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib2_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=6 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032521 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032521 57 global 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib2_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib2_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=7 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032521 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032521 57 gsouth 0p25 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib2_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib2_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[100][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[113][[ 8 -lt 7 ]] -+ exgfs_wave_post_gridded_sbs.sh[121]echo 'INFO: Running MPMD job with 7 commands' -INFO: Running MPMD job with 7 commands -+ exgfs_wave_post_gridded_sbs.sh[122]/work2/noaa/global/mterry/global-workflow_forked/ush/run_mpmd.sh /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/cmdfile -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ run_mpmd.sh[31]source /work2/noaa/global/mterry/global-workflow_forked/ush/preamble.sh -++ preamble.sh[20]set +x -Begin run_mpmd.sh at Tue Jul 29 03:04:38 UTC 2025 -+ run_mpmd.sh[33]cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/cmdfile -+ run_mpmd.sh[36][[ YES != \Y\E\S ]] -+ run_mpmd.sh[46]export OMP_NUM_THREADS=1 -+ run_mpmd.sh[46]OMP_NUM_THREADS=1 -++ run_mpmd.sh[49]wc -l -+ run_mpmd.sh[49]nprocs=7 -+ run_mpmd.sh[52]mpmd_cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mpmd_cmdfile -+ run_mpmd.sh[53][[ -s /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mpmd_cmdfile ]] -+ run_mpmd.sh[55]cat - INFO: Executing MPMD job, STDOUT redirected for each process separately - INFO: On failure, logs for each job will be available in /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mpmd.proc_num.out - INFO: The proc_num corresponds to the line in '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mpmd_cmdfile' -+ run_mpmd.sh[61][[ srun -l --export=ALL --hint=nomultithread =~ ^srun.* ]] -+ run_mpmd.sh[65]nm=0 -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '0 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/cmdfile.1' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '1 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/cmdfile.2' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '2 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/cmdfile.3' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '3 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/cmdfile.4' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '4 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/cmdfile.5' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '5 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/cmdfile.6' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '6 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/cmdfile.7' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[72]set +e -+ run_mpmd.sh[74]srun -l --export=ALL --hint=nomultithread --multi-prog --output=mpmd.%j.%t.out -n 7 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mpmd_cmdfile -+ run_mpmd.sh[75]err=0 -+ run_mpmd.sh[76]set_strict -+ preamble.sh[35][[ YES == \Y\E\S ]] -+ preamble.sh[37]set -eu -+ run_mpmd.sh[103][[ 0 -eq 0 ]] -+ run_mpmd.sh[104]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mpmd_cmdfile -++ run_mpmd.sh[105]find . -name 'mpmd.*.out' -+ run_mpmd.sh[105]out_files='./mpmd.5951794.1.out -./mpmd.5951794.5.out -./mpmd.5951794.3.out -./mpmd.5951794.2.out -./mpmd.5951794.0.out -./mpmd.5951794.4.out -./mpmd.5951794.6.out' -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951794.1.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951794.1.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951794.5.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951794.5.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951794.3.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951794.3.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951794.2.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951794.2.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951794.0.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951794.0.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951794.4.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951794.4.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951794.6.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951794.6.out -+ run_mpmd.sh[110]cat mpmd.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + cmdfile.2[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032521 3600. 9999 -1: + cmdfile.2[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grid_interp_sbs.sh[25]grdID=ep_10m -1: + wave_grid_interp_sbs.sh[26]valid_time=2021032521 -1: + wave_grid_interp_sbs.sh[27]dt=3600. -1: + wave_grid_interp_sbs.sh[28]nst=9999 -1: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463 -1: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.uglo_100km ./out_grd.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mod_def.uglo_100km ./mod_def.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mod_def.ep_10m ./mod_def.ep_10m -1: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ]] -1: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m'\''' -1: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m' -1: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[51]weights_found=1 -1: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 210000' -1: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 210000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/ep_10m/g ww3_gint.inp.tmpl -1: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -1: $ Input file for interpolation of uglo_100km to ep_10m -1: $------------------------------------------------ -1: $ Start Time 3600. NSteps -1: 20210325 210000 3600. 9999 -1: $ Total number of grids -1: 2 -1: $ Grid extensions -1: 'uglo_100km' -1: 'ep_10m' -1: $ -1: 0 -1: $ -1: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[70]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.454158 ']' -1: ++ prep_step[4]echo gfs_ww3_gint.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -1: INFO: Executing 'gfs_ww3_gint.x' -1: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[73]cat grid_interp.ep_10m.out -1: -1: *** WAVEWATCH III Grid interpolation *** -1: =============================================== -1: -1: Comment character is '$' -1: -1: Time Information : -1: --------------------------------------------- -1: Starting Time : 2021/03/25 21:00:00 UTC -1: Interval (in sec) : 3600.00 -1: Number of requests : 9999 -1: --------------------------------------------- -1: Number of grids (including output grid) = 2 -1: -1: -1: Extension for grid 1 is --> uglo_100km -1: -1: Grid Particulars are : -1: Dimensions = 45166 1 -1: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Extension for grid 2 is --> ep_10m -1: -1: Grid Particulars are : -1: Dimensions = 511 301 -1: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -1: -1: -1: Preparing interpolation weights for output grid -1: Total number of wet points for interpolation 7439 -1: -1: -1: Variable: Grid Interpolation Map Units: 0.100E+01 -1: -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 301 | | -1: | | -1: | 0 0 | -1: 262 | 0 0 0 0 | -1: | 0 0 0 0 | -1: | 0 0 0 0 | -1: 223 | 0 0 0 | -1: | 0 | -1: | 0 | -1: 184 | | -1: | 0 0 0 | -1: | 0 0 0 0 | -1: 145 | 0 | -1: | | -1: | | -1: 106 | | -1: | | -1: | | -1: 67 | | -1: | 0 | -1: | | -1: 28 | 0 | -1: | | -1: | | -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: -1: -1: Interpolating fields .... -1: -1: Output group 1 -1: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -1: Output group 2 -1: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -1: Output group 3 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 4 -1: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -1: Output group 5 -1: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -1: Output group 6 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 7 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 8 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 9 -1: Output variables skipped -1: Output group 10 -1: Output variables skipped -1: ------------------------------------------------ -1: 1Current vel. -1: 1Wind speed -1: 1Ice concentration -1: 2Wave height -1: 2Mean wave period(+2) -1: 2Mean wave period(+1) -1: 2Peak frequency -1: 2Mean wave dir. a1b1 -1: 2Peak direction -1: 4Part. wave height -1: 4Part. peak period -1: 4Part. mean direction -1: 5Charnock parameter -1: ------------------------------------------------ -1: OUTPUT TIME : 2021/03/25 21:00:00 UTC -1: -1: End of file reached -1: -1: -1: *** End of Grid interpolation Routine *** -1: =============================================== -1: -1: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -1: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -1: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_ep_10m/out_grd.ep_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.ep_10m -1: + cmdfile.2[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032521 57 epacif 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + cmdfile.2[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib2_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grib2_sbs.sh[30]grdID=ep_10m -1: + wave_grib2_sbs.sh[31]GRIDNR=255 -1: + wave_grib2_sbs.sh[32]MODNR=11 -1: + wave_grib2_sbs.sh[33]valid_time=2021032521 -1: + wave_grib2_sbs.sh[34]fhr=57 -1: + wave_grib2_sbs.sh[35]grid_region=epacif -1: + wave_grib2_sbs.sh[36]grid_res=0p16 -1: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463 -1: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_ep_10m -1: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_ep_10m -1: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_ep_10m -1: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_ep_10m -1: ++ wave_grib2_sbs.sh[47]printf %03i 57 -1: + wave_grib2_sbs.sh[47]FH3=057 -1: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_epacif_0p16 -1: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.epacif.0p16.f057.grib2 -1: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f057.grib2 ]] -1: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mod_def.ep_10m ./mod_def.ww3 -1: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.ep_10m ./out_grd.ww3 -1: + wave_grib2_sbs.sh[73]ngrib=1 -1: + wave_grib2_sbs.sh[74]dtgrib=3600 -1: + wave_grib2_sbs.sh[75]tstart='20210325 210000' -1: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 210000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.ep_10m.inp.tmpl -1: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -1: $ WAVEWATCH-III gridded output input file -1: $ ---------------------------------------- -1: 20210325 210000 3600 1 -1: N -1: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -1: $ -1: 20210325 210000 7 11 255 0 0 -1: $ -1: $ end of input file -1: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[88]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.454158 ']' -1: ++ prep_step[4]echo gfs_ww3_grib.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[90]export err=0 -1: + wave_grib2_sbs.sh[90]err=0 -1: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[95]cat grib2_epacif_057.out -1: -1: *** WAVEWATCH III GRIB output postp. *** -1: ============================================== -1: -1: Comment character is '$' -1: -1: Grid name : East Pacific 10 min wave grid -1: -1: LINEIN: -1: 20210325 210000 3600 1 -1: -1: 20210325210000 3600 1 -1: GEN_PRO -99999 -1: -1: Output time data : -1: ----------------------------------------------------- -1: First time : 2021/03/25 21:00:00 UTC -1: Interval : 01:00:00 -1: Number of requests : 1 -1: Fields : Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: Requested output fields not yet available: -1: ----------------------------------------------------- -1: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -1: -1: Successfully requested output fields : -1: ----------------------------------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: -1: Additional GRIB parameters : -1: ----------------------------------------------------- -1: Run time : 2021/03/25 21:00:00 UTC -1: GRIB center ID : 7 -1: GRIB gen. proc. ID : 11 -1: GRIB grid ID : 255 -1: GRIB GDS parameter : 0 -1: Fields in file : -1: -------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: CHOSEN GRID TYPE: : LLRECTILINEAR -1: -1: -1: -1: Generating file -1: ----------------------------------------------------- -1: Data for 2021/03/25 21:00:00 UTC 0H forecast. -1: -1: End of program -1: ========================================= -1: WAVEWATCH III GRIB output -1: -1: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -1: + wave_grib2_sbs.sh[102][[ 57 -gt 0 ]] -1: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '57 hour fcst' -grib gfs.wave.t12z.epacif.0p16.f057.grib2 -1: 1:0:d=2021032312:SPC:surface:57 hour fcst: -1: 2:21672:d=2021032312:DIRC:surface:57 hour fcst: -1: 3:49712:d=2021032312:UOGRD:surface:57 hour fcst: -1: 4:71543:d=2021032312:VOGRD:surface:57 hour fcst: -1: 5:93564:d=2021032312:WIND:surface:57 hour fcst: -1: 6:118091:d=2021032312:WDIR:surface:57 hour fcst: -1: 7:145644:d=2021032312:UGRD:surface:57 hour fcst: -1: 8:170210:d=2021032312:VGRD:surface:57 hour fcst: -1: 9:194704:d=2021032312:ICEC:surface:57 hour fcst: -1: 10:214112:d=2021032312:HTSGW:surface:57 hour fcst: -1: 11:236992:d=2021032312:IMWF:surface:57 hour fcst: -1: 12:259858:d=2021032312:MWSPER:surface:57 hour fcst: -1: 13:282731:d=2021032312:PERPW:surface:57 hour fcst: -1: 14:305845:d=2021032312:WWSDIR:surface:57 hour fcst: -1: 15:333061:d=2021032312:DIRPW:surface:57 hour fcst: -1: 16:360414:d=2021032312:WVHGT:surface:57 hour fcst: -1: 17:382943:d=2021032312:SWELL:1 in sequence:57 hour fcst: -1: 18:405180:d=2021032312:SWELL:2 in sequence:57 hour fcst: -1: 19:425770:d=2021032312:SWELL:3 in sequence:57 hour fcst: -1: 20:445387:d=2021032312:WVPER:surface:57 hour fcst: -1: 21:467986:d=2021032312:SWPER:1 in sequence:57 hour fcst: -1: 22:490380:d=2021032312:SWPER:2 in sequence:57 hour fcst: -1: 23:511233:d=2021032312:SWPER:3 in sequence:57 hour fcst: -1: 24:530873:d=2021032312:WVDIR:surface:57 hour fcst: -1: 25:555958:d=2021032312:SWDIR:1 in sequence:57 hour fcst: -1: 26:581842:d=2021032312:SWDIR:2 in sequence:57 hour fcst: -1: 27:604392:d=2021032312:SWDIR:3 in sequence:57 hour fcst: -1: + wave_grib2_sbs.sh[104]err=0 -1: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.epacif.0p16.f057.grib2 -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f057.grib2 ]] -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f057.grib2.idx ]] -1: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.epacif.0p16.f057.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f057.grib2 -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f057.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f057.grib2 = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f057.grib2 ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f057.grib2 -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f057.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f057.grib2.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f057.grib2.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f057.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f057.grib2 -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.epacif.0p16.f057.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f057.grib2.idx -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f057.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f057.grib2.idx = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f057.grib2.idx ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f057.grib2.idx -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f057.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f057.grib2.idx.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f057.grib2.idx.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f057.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f057.grib2.idx -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.epacif.0p16.f057.grib2 and gfs.wave.t12z.epacif.0p16.f057.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_ep_10m to COM' -1: INFO: Copied gfs.wave.t12z.epacif.0p16.f057.grib2 and gfs.wave.t12z.epacif.0p16.f057.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_ep_10m to COM -1: + wave_grib2_sbs.sh[130][[ ep_10m == '' ]] -1: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -1: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.epacif.0p16.f057.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -1: INFO: gfs.wave.t12z.epacif.0p16.f057.grib2 is global.0p50 or SENDDBN is NO, no alert sent -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + cmdfile.6[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032521 3600. 9999 -5: + cmdfile.6[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grid_interp_sbs.sh[25]grdID=gnh_10m -5: + wave_grid_interp_sbs.sh[26]valid_time=2021032521 -5: + wave_grid_interp_sbs.sh[27]dt=3600. -5: + wave_grid_interp_sbs.sh[28]nst=9999 -5: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463 -5: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.uglo_100km ./out_grd.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mod_def.uglo_100km ./mod_def.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mod_def.gnh_10m ./mod_def.gnh_10m -5: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m'\''' -5: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m' -5: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[51]weights_found=1 -5: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 210000' -5: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 210000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gnh_10m/g ww3_gint.inp.tmpl -5: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -5: $ Input file for interpolation of uglo_100km to gnh_10m -5: $------------------------------------------------ -5: $ Start Time 3600. NSteps -5: 20210325 210000 3600. 9999 -5: $ Total number of grids -5: 2 -5: $ Grid extensions -5: 'uglo_100km' -5: 'gnh_10m' -5: $ -5: 0 -5: $ -5: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[70]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.454158 ']' -5: ++ prep_step[4]echo gfs_ww3_gint.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -5: INFO: Executing 'gfs_ww3_gint.x' -5: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[73]cat grid_interp.gnh_10m.out -5: -5: *** WAVEWATCH III Grid interpolation *** -5: =============================================== -5: -5: Comment character is '$' -5: -5: Time Information : -5: --------------------------------------------- -5: Starting Time : 2021/03/25 21:00:00 UTC -5: Interval (in sec) : 3600.00 -5: Number of requests : 9999 -5: --------------------------------------------- -5: Number of grids (including output grid) = 2 -5: -5: -5: Extension for grid 1 is --> uglo_100km -5: -5: Grid Particulars are : -5: Dimensions = 45166 1 -5: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Extension for grid 2 is --> gnh_10m -5: -5: Grid Particulars are : -5: Dimensions = 2160 406 -5: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -5: -5: -5: Preparing interpolation weights for output grid -5: Total number of wet points for interpolation 571209 -5: -5: -5: Variable: Grid Interpolation Map Units: 0.100E+01 -5: -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 406 | | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: 355 | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 304 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 253 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 202 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 151 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 100 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 49 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: -5: -5: Interpolating fields .... -5: -5: Output group 1 -5: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -5: Output group 2 -5: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -5: Output group 3 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 4 -5: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -5: Output group 5 -5: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -5: Output group 6 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 7 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 8 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 9 -5: Output variables skipped -5: Output group 10 -5: Output variables skipped -5: ------------------------------------------------ -5: 1Current vel. -5: 1Wind speed -5: 1Ice concentration -5: 2Wave height -5: 2Mean wave period(+2) -5: 2Mean wave period(+1) -5: 2Peak frequency -5: 2Mean wave dir. a1b1 -5: 2Peak direction -5: 4Part. wave height -5: 4Part. peak period -5: 4Part. mean direction -5: 5Charnock parameter -5: ------------------------------------------------ -5: OUTPUT TIME : 2021/03/25 21:00:00 UTC -5: -5: End of file reached -5: -5: -5: *** End of Grid interpolation Routine *** -5: =============================================== -5: -5: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -5: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -5: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_gnh_10m/out_grd.gnh_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.gnh_10m -5: + cmdfile.6[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032521 57 global 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + cmdfile.6[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib2_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grib2_sbs.sh[30]grdID=gnh_10m -5: + wave_grib2_sbs.sh[31]GRIDNR=255 -5: + wave_grib2_sbs.sh[32]MODNR=11 -5: + wave_grib2_sbs.sh[33]valid_time=2021032521 -5: + wave_grib2_sbs.sh[34]fhr=57 -5: + wave_grib2_sbs.sh[35]grid_region=global -5: + wave_grib2_sbs.sh[36]grid_res=0p16 -5: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463 -5: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_gnh_10m -5: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_gnh_10m -5: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_gnh_10m -5: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_gnh_10m -5: ++ wave_grib2_sbs.sh[47]printf %03i 57 -5: + wave_grib2_sbs.sh[47]FH3=057 -5: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p16 -5: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p16.f057.grib2 -5: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f057.grib2 ]] -5: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mod_def.gnh_10m ./mod_def.ww3 -5: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.gnh_10m ./out_grd.ww3 -5: + wave_grib2_sbs.sh[73]ngrib=1 -5: + wave_grib2_sbs.sh[74]dtgrib=3600 -5: + wave_grib2_sbs.sh[75]tstart='20210325 210000' -5: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 210000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gnh_10m.inp.tmpl -5: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -5: $ WAVEWATCH-III gridded output input file -5: $ ---------------------------------------- -5: 20210325 210000 3600 1 -5: N -5: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -5: $ -5: 20210325 210000 7 11 255 0 0 -5: $ -5: $ end of input file -5: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[88]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.454158 ']' -5: ++ prep_step[4]echo gfs_ww3_grib.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[90]export err=0 -5: + wave_grib2_sbs.sh[90]err=0 -5: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[95]cat grib2_global_057.out -5: -5: *** WAVEWATCH III GRIB output postp. *** -5: ============================================== -5: -5: Comment character is '$' -5: -5: Grid name : GFSv16-wave N Hemisphere 1/6 d -5: -5: LINEIN: -5: 20210325 210000 3600 1 -5: -5: 20210325210000 3600 1 -5: GEN_PRO -99999 -5: -5: Output time data : -5: ----------------------------------------------------- -5: First time : 2021/03/25 21:00:00 UTC -5: Interval : 01:00:00 -5: Number of requests : 1 -5: Fields : Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: Requested output fields not yet available: -5: ----------------------------------------------------- -5: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -5: -5: Successfully requested output fields : -5: ----------------------------------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: -5: Additional GRIB parameters : -5: ----------------------------------------------------- -5: Run time : 2021/03/25 21:00:00 UTC -5: GRIB center ID : 7 -5: GRIB gen. proc. ID : 11 -5: GRIB grid ID : 255 -5: GRIB GDS parameter : 0 -5: Fields in file : -5: -------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: CHOSEN GRID TYPE: : LLRECTILINEAR -5: -5: -5: -5: Generating file -5: ----------------------------------------------------- -5: Data for 2021/03/25 21:00:00 UTC 0H forecast. -5: -5: End of program -5: ========================================= -5: WAVEWATCH III GRIB output -5: -5: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -5: + wave_grib2_sbs.sh[102][[ 57 -gt 0 ]] -5: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '57 hour fcst' -grib gfs.wave.t12z.global.0p16.f057.grib2 -5: 1:0:d=2021032312:SPC:surface:57 hour fcst: -5: 2:235963:d=2021032312:DIRC:surface:57 hour fcst: -5: 3:864486:d=2021032312:UOGRD:surface:57 hour fcst: -5: 4:1095340:d=2021032312:VOGRD:surface:57 hour fcst: -5: 5:1332644:d=2021032312:WIND:surface:57 hour fcst: -5: 6:1734894:d=2021032312:WDIR:surface:57 hour fcst: -5: 7:2379127:d=2021032312:UGRD:surface:57 hour fcst: -5: 8:2778171:d=2021032312:VGRD:surface:57 hour fcst: -5: 9:3176241:d=2021032312:ICEC:surface:57 hour fcst: -5: 10:3290939:d=2021032312:HTSGW:surface:57 hour fcst: -5: 11:3574284:d=2021032312:IMWF:surface:57 hour fcst: -5: 12:3883642:d=2021032312:MWSPER:surface:57 hour fcst: -5: 13:4194961:d=2021032312:PERPW:surface:57 hour fcst: -5: 14:4526458:d=2021032312:WWSDIR:surface:57 hour fcst: -5: 15:5121464:d=2021032312:DIRPW:surface:57 hour fcst: -5: 16:5744314:d=2021032312:WVHGT:surface:57 hour fcst: -5: 17:6025383:d=2021032312:SWELL:1 in sequence:57 hour fcst: -5: 18:6275204:d=2021032312:SWELL:2 in sequence:57 hour fcst: -5: 19:6441118:d=2021032312:SWELL:3 in sequence:57 hour fcst: -5: 20:6559150:d=2021032312:WVPER:surface:57 hour fcst: -5: 21:6878549:d=2021032312:SWPER:1 in sequence:57 hour fcst: -5: 22:7164164:d=2021032312:SWPER:2 in sequence:57 hour fcst: -5: 23:7359048:d=2021032312:SWPER:3 in sequence:57 hour fcst: -5: 24:7484022:d=2021032312:WVDIR:surface:57 hour fcst: -5: 25:8009190:d=2021032312:SWDIR:1 in sequence:57 hour fcst: -5: 26:8508467:d=2021032312:SWDIR:2 in sequence:57 hour fcst: -5: 27:8798930:d=2021032312:SWDIR:3 in sequence:57 hour fcst: -5: + wave_grib2_sbs.sh[104]err=0 -5: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p16.f057.grib2 -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f057.grib2 ]] -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f057.grib2.idx ]] -5: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p16.f057.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f057.grib2 -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f057.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f057.grib2 = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f057.grib2 ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f057.grib2 -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f057.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f057.grib2.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f057.grib2.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f057.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f057.grib2 -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p16.f057.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f057.grib2.idx -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f057.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f057.grib2.idx = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f057.grib2.idx ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f057.grib2.idx -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f057.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f057.grib2.idx.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f057.grib2.idx.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f057.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f057.grib2.idx -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p16.f057.grib2 and gfs.wave.t12z.global.0p16.f057.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_gnh_10m to COM' -5: INFO: Copied gfs.wave.t12z.global.0p16.f057.grib2 and gfs.wave.t12z.global.0p16.f057.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_gnh_10m to COM -5: + wave_grib2_sbs.sh[130][[ gnh_10m == '' ]] -5: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -5: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p16.f057.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -5: INFO: gfs.wave.t12z.global.0p16.f057.grib2 is global.0p50 or SENDDBN is NO, no alert sent -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + cmdfile.4[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032521 3600. 9999 -3: + cmdfile.4[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grid_interp_sbs.sh[25]grdID=glo_30m -3: + wave_grid_interp_sbs.sh[26]valid_time=2021032521 -3: + wave_grid_interp_sbs.sh[27]dt=3600. -3: + wave_grid_interp_sbs.sh[28]nst=9999 -3: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463 -3: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.uglo_100km ./out_grd.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mod_def.uglo_100km ./mod_def.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mod_def.glo_30m ./mod_def.glo_30m -3: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ]] -3: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m'\''' -3: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m' -3: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[51]weights_found=1 -3: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 210000' -3: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 210000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/glo_30m/g ww3_gint.inp.tmpl -3: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -3: $ Input file for interpolation of uglo_100km to glo_30m -3: $------------------------------------------------ -3: $ Start Time 3600. NSteps -3: 20210325 210000 3600. 9999 -3: $ Total number of grids -3: 2 -3: $ Grid extensions -3: 'uglo_100km' -3: 'glo_30m' -3: $ -3: 0 -3: $ -3: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[70]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.454158 ']' -3: ++ prep_step[4]echo gfs_ww3_gint.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -3: INFO: Executing 'gfs_ww3_gint.x' -3: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[73]cat grid_interp.glo_30m.out -3: -3: *** WAVEWATCH III Grid interpolation *** -3: =============================================== -3: -3: Comment character is '$' -3: -3: Time Information : -3: --------------------------------------------- -3: Starting Time : 2021/03/25 21:00:00 UTC -3: Interval (in sec) : 3600.00 -3: Number of requests : 9999 -3: --------------------------------------------- -3: Number of grids (including output grid) = 2 -3: -3: -3: Extension for grid 1 is --> uglo_100km -3: -3: Grid Particulars are : -3: Dimensions = 45166 1 -3: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Extension for grid 2 is --> glo_30m -3: -3: Grid Particulars are : -3: Dimensions = 720 336 -3: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -3: -3: -3: Preparing interpolation weights for output grid -3: Total number of wet points for interpolation 167619 -3: -3: -3: Variable: Grid Interpolation Map Units: 0.100E+01 -3: -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 336 | | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 291 | 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 | -3: 246 | 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 201 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 156 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 111 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 66 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 21 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 | -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: -3: -3: Interpolating fields .... -3: -3: Output group 1 -3: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -3: Output group 2 -3: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -3: Output group 3 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 4 -3: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -3: Output group 5 -3: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -3: Output group 6 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 7 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 8 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 9 -3: Output variables skipped -3: Output group 10 -3: Output variables skipped -3: ------------------------------------------------ -3: 1Current vel. -3: 1Wind speed -3: 1Ice concentration -3: 2Wave height -3: 2Mean wave period(+2) -3: 2Mean wave period(+1) -3: 2Peak frequency -3: 2Mean wave dir. a1b1 -3: 2Peak direction -3: 4Part. wave height -3: 4Part. peak period -3: 4Part. mean direction -3: 5Charnock parameter -3: ------------------------------------------------ -3: OUTPUT TIME : 2021/03/25 21:00:00 UTC -3: -3: End of file reached -3: -3: -3: *** End of Grid interpolation Routine *** -3: =============================================== -3: -3: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -3: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -3: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_glo_30m/out_grd.glo_30m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.glo_30m -3: + cmdfile.4[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032521 57 global 0p50 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + cmdfile.4[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib2_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grib2_sbs.sh[30]grdID=glo_30m -3: + wave_grib2_sbs.sh[31]GRIDNR=255 -3: + wave_grib2_sbs.sh[32]MODNR=11 -3: + wave_grib2_sbs.sh[33]valid_time=2021032521 -3: + wave_grib2_sbs.sh[34]fhr=57 -3: + wave_grib2_sbs.sh[35]grid_region=global -3: + wave_grib2_sbs.sh[36]grid_res=0p50 -3: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463 -3: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_glo_30m -3: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_glo_30m -3: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_glo_30m -3: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_glo_30m -3: ++ wave_grib2_sbs.sh[47]printf %03i 57 -3: + wave_grib2_sbs.sh[47]FH3=057 -3: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p50 -3: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p50.f057.grib2 -3: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f057.grib2 ]] -3: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mod_def.glo_30m ./mod_def.ww3 -3: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.glo_30m ./out_grd.ww3 -3: + wave_grib2_sbs.sh[73]ngrib=1 -3: + wave_grib2_sbs.sh[74]dtgrib=3600 -3: + wave_grib2_sbs.sh[75]tstart='20210325 210000' -3: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 210000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.glo_30m.inp.tmpl -3: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -3: $ WAVEWATCH-III gridded output input file -3: $ ---------------------------------------- -3: 20210325 210000 3600 1 -3: N -3: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -3: $ -3: 20210325 210000 7 11 255 0 0 -3: $ -3: $ end of input file -3: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[88]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.454158 ']' -3: ++ prep_step[4]echo gfs_ww3_grib.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[90]export err=0 -3: + wave_grib2_sbs.sh[90]err=0 -3: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[95]cat grib2_global_057.out -3: -3: *** WAVEWATCH III GRIB output postp. *** -3: ============================================== -3: -3: Comment character is '$' -3: -3: Grid name : Global 30 min wave grid -3: -3: LINEIN: -3: 20210325 210000 3600 1 -3: -3: 20210325210000 3600 1 -3: GEN_PRO -99999 -3: -3: Output time data : -3: ----------------------------------------------------- -3: First time : 2021/03/25 21:00:00 UTC -3: Interval : 01:00:00 -3: Number of requests : 1 -3: Fields : Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: Requested output fields not yet available: -3: ----------------------------------------------------- -3: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -3: -3: Successfully requested output fields : -3: ----------------------------------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: -3: Additional GRIB parameters : -3: ----------------------------------------------------- -3: Run time : 2021/03/25 21:00:00 UTC -3: GRIB center ID : 7 -3: GRIB gen. proc. ID : 11 -3: GRIB grid ID : 255 -3: GRIB GDS parameter : 0 -3: Fields in file : -3: -------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: CHOSEN GRID TYPE: : LLRECTILINEAR -3: -3: -3: -3: Generating file -3: ----------------------------------------------------- -3: Data for 2021/03/25 21:00:00 UTC 0H forecast. -3: -3: End of program -3: ========================================= -3: WAVEWATCH III GRIB output -3: -3: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -3: + wave_grib2_sbs.sh[102][[ 57 -gt 0 ]] -3: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '57 hour fcst' -grib gfs.wave.t12z.global.0p50.f057.grib2 -3: 1:0:d=2021032312:SPC:surface:57 hour fcst: -3: 2:77620:d=2021032312:DIRC:surface:57 hour fcst: -3: 3:319373:d=2021032312:UOGRD:surface:57 hour fcst: -3: 4:394838:d=2021032312:VOGRD:surface:57 hour fcst: -3: 5:474139:d=2021032312:WIND:surface:57 hour fcst: -3: 6:632192:d=2021032312:WDIR:surface:57 hour fcst: -3: 7:867658:d=2021032312:UGRD:surface:57 hour fcst: -3: 8:1023196:d=2021032312:VGRD:surface:57 hour fcst: -3: 9:1180084:d=2021032312:ICEC:surface:57 hour fcst: -3: 10:1220437:d=2021032312:HTSGW:surface:57 hour fcst: -3: 11:1319296:d=2021032312:IMWF:surface:57 hour fcst: -3: 12:1427829:d=2021032312:MWSPER:surface:57 hour fcst: -3: 13:1537764:d=2021032312:PERPW:surface:57 hour fcst: -3: 14:1657197:d=2021032312:WWSDIR:surface:57 hour fcst: -3: 15:1862094:d=2021032312:DIRPW:surface:57 hour fcst: -3: 16:2075808:d=2021032312:WVHGT:surface:57 hour fcst: -3: 17:2175875:d=2021032312:SWELL:1 in sequence:57 hour fcst: -3: 18:2268819:d=2021032312:SWELL:2 in sequence:57 hour fcst: -3: 19:2327334:d=2021032312:SWELL:3 in sequence:57 hour fcst: -3: 20:2363254:d=2021032312:WVPER:surface:57 hour fcst: -3: 21:2476960:d=2021032312:SWPER:1 in sequence:57 hour fcst: -3: 22:2582925:d=2021032312:SWPER:2 in sequence:57 hour fcst: -3: 23:2654100:d=2021032312:SWPER:3 in sequence:57 hour fcst: -3: 24:2693540:d=2021032312:WVDIR:surface:57 hour fcst: -3: 25:2871702:d=2021032312:SWDIR:1 in sequence:57 hour fcst: -3: 26:3051172:d=2021032312:SWDIR:2 in sequence:57 hour fcst: -3: 27:3160025:d=2021032312:SWDIR:3 in sequence:57 hour fcst: -3: + wave_grib2_sbs.sh[104]err=0 -3: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p50.f057.grib2 -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f057.grib2 ]] -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f057.grib2.idx ]] -3: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p50.f057.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f057.grib2 -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f057.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f057.grib2 = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f057.grib2 ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f057.grib2 -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f057.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f057.grib2.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f057.grib2.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f057.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f057.grib2 -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p50.f057.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f057.grib2.idx -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f057.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f057.grib2.idx = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f057.grib2.idx ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f057.grib2.idx -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f057.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f057.grib2.idx.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f057.grib2.idx.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f057.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f057.grib2.idx -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p50.f057.grib2 and gfs.wave.t12z.global.0p50.f057.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_glo_30m to COM' -3: INFO: Copied gfs.wave.t12z.global.0p50.f057.grib2 and gfs.wave.t12z.global.0p50.f057.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_glo_30m to COM -3: + wave_grib2_sbs.sh[130][[ glo_30m == '' ]] -3: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -3: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p50.f057.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -3: INFO: gfs.wave.t12z.global.0p50.f057.grib2 is global.0p50 or SENDDBN is NO, no alert sent -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + cmdfile.3[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032521 3600. 9999 -2: + cmdfile.3[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grid_interp_sbs.sh[25]grdID=wc_10m -2: + wave_grid_interp_sbs.sh[26]valid_time=2021032521 -2: + wave_grid_interp_sbs.sh[27]dt=3600. -2: + wave_grid_interp_sbs.sh[28]nst=9999 -2: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463 -2: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.uglo_100km ./out_grd.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mod_def.uglo_100km ./mod_def.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mod_def.wc_10m ./mod_def.wc_10m -2: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ]] -2: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m'\''' -2: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m' -2: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[51]weights_found=1 -2: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 210000' -2: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 210000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/wc_10m/g ww3_gint.inp.tmpl -2: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -2: $ Input file for interpolation of uglo_100km to wc_10m -2: $------------------------------------------------ -2: $ Start Time 3600. NSteps -2: 20210325 210000 3600. 9999 -2: $ Total number of grids -2: 2 -2: $ Grid extensions -2: 'uglo_100km' -2: 'wc_10m' -2: $ -2: 0 -2: $ -2: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[70]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.454158 ']' -2: ++ prep_step[4]echo gfs_ww3_gint.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -2: INFO: Executing 'gfs_ww3_gint.x' -2: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[73]cat grid_interp.wc_10m.out -2: -2: *** WAVEWATCH III Grid interpolation *** -2: =============================================== -2: -2: Comment character is '$' -2: -2: Time Information : -2: --------------------------------------------- -2: Starting Time : 2021/03/25 21:00:00 UTC -2: Interval (in sec) : 3600.00 -2: Number of requests : 9999 -2: --------------------------------------------- -2: Number of grids (including output grid) = 2 -2: -2: -2: Extension for grid 1 is --> uglo_100km -2: -2: Grid Particulars are : -2: Dimensions = 45166 1 -2: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Extension for grid 2 is --> wc_10m -2: -2: Grid Particulars are : -2: Dimensions = 241 151 -2: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -2: -2: -2: Preparing interpolation weights for output grid -2: Total number of wet points for interpolation 11044 -2: -2: -2: Variable: Grid Interpolation Map Units: 0.100E+01 -2: -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: +---------------------------------------------------------------------------------------------------------------+ -2: 151 | 0 0 0 0 | -2: | 0 0 0 0 0 | -2: | 0 0 0 0 0 | -2: 130 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 109 | 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 88 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 | -2: 67 | 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 | -2: 46 | 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: 25 | 0 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 0 | -2: | | -2: 4 | | -2: +---------------------------------------------------------------------------------------------------------------+ -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: -2: -2: Interpolating fields .... -2: -2: Output group 1 -2: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -2: Output group 2 -2: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -2: Output group 3 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 4 -2: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -2: Output group 5 -2: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -2: Output group 6 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 7 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 8 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 9 -2: Output variables skipped -2: Output group 10 -2: Output variables skipped -2: ------------------------------------------------ -2: 1Current vel. -2: 1Wind speed -2: 1Ice concentration -2: 2Wave height -2: 2Mean wave period(+2) -2: 2Mean wave period(+1) -2: 2Peak frequency -2: 2Mean wave dir. a1b1 -2: 2Peak direction -2: 4Part. wave height -2: 4Part. peak period -2: 4Part. mean direction -2: 5Charnock parameter -2: ------------------------------------------------ -2: OUTPUT TIME : 2021/03/25 21:00:00 UTC -2: -2: End of file reached -2: -2: -2: *** End of Grid interpolation Routine *** -2: =============================================== -2: -2: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -2: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -2: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_wc_10m/out_grd.wc_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.wc_10m -2: + cmdfile.3[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032521 57 wcoast 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + cmdfile.3[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib2_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grib2_sbs.sh[30]grdID=wc_10m -2: + wave_grib2_sbs.sh[31]GRIDNR=255 -2: + wave_grib2_sbs.sh[32]MODNR=11 -2: + wave_grib2_sbs.sh[33]valid_time=2021032521 -2: + wave_grib2_sbs.sh[34]fhr=57 -2: + wave_grib2_sbs.sh[35]grid_region=wcoast -2: + wave_grib2_sbs.sh[36]grid_res=0p16 -2: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463 -2: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_wc_10m -2: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_wc_10m -2: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_wc_10m -2: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_wc_10m -2: ++ wave_grib2_sbs.sh[47]printf %03i 57 -2: + wave_grib2_sbs.sh[47]FH3=057 -2: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_wcoast_0p16 -2: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.wcoast.0p16.f057.grib2 -2: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f057.grib2 ]] -2: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mod_def.wc_10m ./mod_def.ww3 -2: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.wc_10m ./out_grd.ww3 -2: + wave_grib2_sbs.sh[73]ngrib=1 -2: + wave_grib2_sbs.sh[74]dtgrib=3600 -2: + wave_grib2_sbs.sh[75]tstart='20210325 210000' -2: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 210000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.wc_10m.inp.tmpl -2: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -2: $ WAVEWATCH-III gridded output input file -2: $ ---------------------------------------- -2: 20210325 210000 3600 1 -2: N -2: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -2: $ -2: 20210325 210000 7 11 255 0 0 -2: $ -2: $ end of input file -2: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[88]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.454158 ']' -2: ++ prep_step[4]echo gfs_ww3_grib.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[90]export err=0 -2: + wave_grib2_sbs.sh[90]err=0 -2: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[95]cat grib2_wcoast_057.out -2: -2: *** WAVEWATCH III GRIB output postp. *** -2: ============================================== -2: -2: Comment character is '$' -2: -2: Grid name : West Coast 10 min wave grid -2: -2: LINEIN: -2: 20210325 210000 3600 1 -2: -2: 20210325210000 3600 1 -2: GEN_PRO -99999 -2: -2: Output time data : -2: ----------------------------------------------------- -2: First time : 2021/03/25 21:00:00 UTC -2: Interval : 01:00:00 -2: Number of requests : 1 -2: Fields : Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: Requested output fields not yet available: -2: ----------------------------------------------------- -2: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -2: -2: Successfully requested output fields : -2: ----------------------------------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: -2: Additional GRIB parameters : -2: ----------------------------------------------------- -2: Run time : 2021/03/25 21:00:00 UTC -2: GRIB center ID : 7 -2: GRIB gen. proc. ID : 11 -2: GRIB grid ID : 255 -2: GRIB GDS parameter : 0 -2: Fields in file : -2: -------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: CHOSEN GRID TYPE: : LLRECTILINEAR -2: -2: -2: -2: Generating file -2: ----------------------------------------------------- -2: Data for 2021/03/25 21:00:00 UTC 0H forecast. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: -2: End of program -2: ========================================= -2: WAVEWATCH III GRIB output -2: -2: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -2: + wave_grib2_sbs.sh[102][[ 57 -gt 0 ]] -2: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '57 hour fcst' -grib gfs.wave.t12z.wcoast.0p16.f057.grib2 -2: 1:0:d=2021032312:SPC:surface:57 hour fcst: -2: 2:7352:d=2021032312:DIRC:surface:57 hour fcst: -2: 3:23878:d=2021032312:UOGRD:surface:57 hour fcst: -2: 4:30898:d=2021032312:VOGRD:surface:57 hour fcst: -2: 5:38501:d=2021032312:WIND:surface:57 hour fcst: -2: 6:49733:d=2021032312:WDIR:surface:57 hour fcst: -2: 7:65642:d=2021032312:UGRD:surface:57 hour fcst: -2: 8:76934:d=2021032312:VGRD:surface:57 hour fcst: -2: 9:88334:d=2021032312:ICEC:surface:57 hour fcst: -2: 10:93064:d=2021032312:HTSGW:surface:57 hour fcst: -2: 11:102628:d=2021032312:IMWF:surface:57 hour fcst: -2: 12:112260:d=2021032312:MWSPER:surface:57 hour fcst: -2: 13:121982:d=2021032312:PERPW:surface:57 hour fcst: -2: 14:131898:d=2021032312:WWSDIR:surface:57 hour fcst: -2: 15:146758:d=2021032312:DIRPW:surface:57 hour fcst: -2: 16:162047:d=2021032312:WVHGT:surface:57 hour fcst: -2: 17:171617:d=2021032312:SWELL:1 in sequence:57 hour fcst: -2: 18:177201:d=2021032312:SWELL:2 in sequence:57 hour fcst: -2: 19:181931:d=2021032312:SWELL:3 in sequence:57 hour fcst: -2: 20:186661:d=2021032312:WVPER:surface:57 hour fcst: -2: 21:196547:d=2021032312:SWPER:1 in sequence:57 hour fcst: -2: 22:202004:d=2021032312:SWPER:2 in sequence:57 hour fcst: -2: 23:206734:d=2021032312:SWPER:3 in sequence:57 hour fcst: -2: 24:211464:d=2021032312:WVDIR:surface:57 hour fcst: -2: 25:226119:d=2021032312:SWDIR:1 in sequence:57 hour fcst: -2: 26:232385:d=2021032312:SWDIR:2 in sequence:57 hour fcst: -2: 27:237115:d=2021032312:SWDIR:3 in sequence:57 hour fcst: -2: + wave_grib2_sbs.sh[104]err=0 -2: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.wcoast.0p16.f057.grib2 -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f057.grib2 ]] -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f057.grib2.idx ]] -2: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.wcoast.0p16.f057.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f057.grib2 -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f057.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f057.grib2 = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f057.grib2 ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f057.grib2 -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f057.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f057.grib2.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f057.grib2.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f057.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f057.grib2 -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.wcoast.0p16.f057.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f057.grib2.idx -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f057.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f057.grib2.idx = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f057.grib2.idx ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f057.grib2.idx -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f057.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f057.grib2.idx.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f057.grib2.idx.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f057.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f057.grib2.idx -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.wcoast.0p16.f057.grib2 and gfs.wave.t12z.wcoast.0p16.f057.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_wc_10m to COM' -2: INFO: Copied gfs.wave.t12z.wcoast.0p16.f057.grib2 and gfs.wave.t12z.wcoast.0p16.f057.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_wc_10m to COM -2: + wave_grib2_sbs.sh[130][[ wc_10m == '' ]] -2: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -2: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.wcoast.0p16.f057.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -2: INFO: gfs.wave.t12z.wcoast.0p16.f057.grib2 is global.0p50 or SENDDBN is NO, no alert sent -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + cmdfile.1[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032521 3600. 9999 -0: + cmdfile.1[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grid_interp_sbs.sh[25]grdID=at_10m -0: + wave_grid_interp_sbs.sh[26]valid_time=2021032521 -0: + wave_grid_interp_sbs.sh[27]dt=3600. -0: + wave_grid_interp_sbs.sh[28]nst=9999 -0: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463 -0: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.uglo_100km ./out_grd.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mod_def.uglo_100km ./mod_def.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mod_def.at_10m ./mod_def.at_10m -0: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ]] -0: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m'\''' -0: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m' -0: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[51]weights_found=1 -0: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 210000' -0: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 210000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/at_10m/g ww3_gint.inp.tmpl -0: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -0: $ Input file for interpolation of uglo_100km to at_10m -0: $------------------------------------------------ -0: $ Start Time 3600. NSteps -0: 20210325 210000 3600. 9999 -0: $ Total number of grids -0: 2 -0: $ Grid extensions -0: 'uglo_100km' -0: 'at_10m' -0: $ -0: 0 -0: $ -0: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[70]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.454158 ']' -0: ++ prep_step[4]echo gfs_ww3_gint.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -0: INFO: Executing 'gfs_ww3_gint.x' -0: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[73]cat grid_interp.at_10m.out -0: -0: *** WAVEWATCH III Grid interpolation *** -0: =============================================== -0: -0: Comment character is '$' -0: -0: Time Information : -0: --------------------------------------------- -0: Starting Time : 2021/03/25 21:00:00 UTC -0: Interval (in sec) : 3600.00 -0: Number of requests : 9999 -0: --------------------------------------------- -0: Number of grids (including output grid) = 2 -0: -0: -0: Extension for grid 1 is --> uglo_100km -0: -0: Grid Particulars are : -0: Dimensions = 45166 1 -0: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Extension for grid 2 is --> at_10m -0: -0: Grid Particulars are : -0: Dimensions = 301 331 -0: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -0: -0: -0: Preparing interpolation weights for output grid -0: Total number of wet points for interpolation 29591 -0: -0: -0: Variable: Grid Interpolation Map Units: 0.100E+01 -0: -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 331 | | -0: | | -0: | | -0: 289 | | -0: | | -0: | 0 0 0 | -0: 247 | 0 0 0 0 | -0: | 0 0 0 0 0 | -0: | 0 0 0 0 | -0: 205 | 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 163 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 121 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 79 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 | -0: | 0 0 | -0: 37 | 0 | -0: | | -0: | | -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: -0: -0: Interpolating fields .... -0: -0: Output group 1 -0: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -0: Output group 2 -0: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -0: Output group 3 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 4 -0: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -0: Output group 5 -0: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -0: Output group 6 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 7 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 8 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 9 -0: Output variables skipped -0: Output group 10 -0: Output variables skipped -0: ------------------------------------------------ -0: 1Current vel. -0: 1Wind speed -0: 1Ice concentration -0: 2Wave height -0: 2Mean wave period(+2) -0: 2Mean wave period(+1) -0: 2Peak frequency -0: 2Mean wave dir. a1b1 -0: 2Peak direction -0: 4Part. wave height -0: 4Part. peak period -0: 4Part. mean direction -0: 5Charnock parameter -0: ------------------------------------------------ -0: OUTPUT TIME : 2021/03/25 21:00:00 UTC -0: -0: End of file reached -0: -0: -0: *** End of Grid interpolation Routine *** -0: =============================================== -0: -0: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -0: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -0: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_at_10m/out_grd.at_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.at_10m -0: + cmdfile.1[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032521 57 atlocn 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + cmdfile.1[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib2_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grib2_sbs.sh[30]grdID=at_10m -0: + wave_grib2_sbs.sh[31]GRIDNR=255 -0: + wave_grib2_sbs.sh[32]MODNR=11 -0: + wave_grib2_sbs.sh[33]valid_time=2021032521 -0: + wave_grib2_sbs.sh[34]fhr=57 -0: + wave_grib2_sbs.sh[35]grid_region=atlocn -0: + wave_grib2_sbs.sh[36]grid_res=0p16 -0: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463 -0: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_at_10m -0: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_at_10m -0: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_at_10m -0: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_at_10m -0: ++ wave_grib2_sbs.sh[47]printf %03i 57 -0: + wave_grib2_sbs.sh[47]FH3=057 -0: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_atlocn_0p16 -0: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.atlocn.0p16.f057.grib2 -0: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f057.grib2 ]] -0: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mod_def.at_10m ./mod_def.ww3 -0: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.at_10m ./out_grd.ww3 -0: + wave_grib2_sbs.sh[73]ngrib=1 -0: + wave_grib2_sbs.sh[74]dtgrib=3600 -0: + wave_grib2_sbs.sh[75]tstart='20210325 210000' -0: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 210000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.at_10m.inp.tmpl -0: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -0: $ WAVEWATCH-III gridded output input file -0: $ ---------------------------------------- -0: 20210325 210000 3600 1 -0: N -0: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -0: $ -0: 20210325 210000 7 11 255 0 0 -0: $ -0: $ end of input file -0: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[88]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.454158 ']' -0: ++ prep_step[4]echo gfs_ww3_grib.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[90]export err=0 -0: + wave_grib2_sbs.sh[90]err=0 -0: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[95]cat grib2_atlocn_057.out -0: -0: *** WAVEWATCH III GRIB output postp. *** -0: ============================================== -0: -0: Comment character is '$' -0: -0: Grid name : NW Atlantic 10 min wave grid -0: -0: LINEIN: -0: 20210325 210000 3600 1 -0: -0: 20210325210000 3600 1 -0: GEN_PRO -99999 -0: -0: Output time data : -0: ----------------------------------------------------- -0: First time : 2021/03/25 21:00:00 UTC -0: Interval : 01:00:00 -0: Number of requests : 1 -0: Fields : Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: Requested output fields not yet available: -0: ----------------------------------------------------- -0: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -0: -0: Successfully requested output fields : -0: ----------------------------------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: -0: Additional GRIB parameters : -0: ----------------------------------------------------- -0: Run time : 2021/03/25 21:00:00 UTC -0: GRIB center ID : 7 -0: GRIB gen. proc. ID : 11 -0: GRIB grid ID : 255 -0: GRIB GDS parameter : 0 -0: Fields in file : -0: -------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: CHOSEN GRID TYPE: : LLRECTILINEAR -0: -0: -0: -0: Generating file -0: ----------------------------------------------------- -0: Data for 2021/03/25 21:00:00 UTC 0H forecast. -0: -0: End of program -0: ========================================= -0: WAVEWATCH III GRIB output -0: -0: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -0: + wave_grib2_sbs.sh[102][[ 57 -gt 0 ]] -0: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '57 hour fcst' -grib gfs.wave.t12z.atlocn.0p16.f057.grib2 -0: 1:0:d=2021032312:SPC:surface:57 hour fcst: -0: 2:20210:d=2021032312:DIRC:surface:57 hour fcst: -0: 3:60928:d=2021032312:UOGRD:surface:57 hour fcst: -0: 4:80853:d=2021032312:VOGRD:surface:57 hour fcst: -0: 5:100851:d=2021032312:WIND:surface:57 hour fcst: -0: 6:131501:d=2021032312:WDIR:surface:57 hour fcst: -0: 7:172927:d=2021032312:UGRD:surface:57 hour fcst: -0: 8:203149:d=2021032312:VGRD:surface:57 hour fcst: -0: 9:233176:d=2021032312:ICEC:surface:57 hour fcst: -0: 10:245811:d=2021032312:HTSGW:surface:57 hour fcst: -0: 11:270653:d=2021032312:IMWF:surface:57 hour fcst: -0: 12:297090:d=2021032312:MWSPER:surface:57 hour fcst: -0: 13:323767:d=2021032312:PERPW:surface:57 hour fcst: -0: 14:351587:d=2021032312:WWSDIR:surface:57 hour fcst: -0: 15:391926:d=2021032312:DIRPW:surface:57 hour fcst: -0: 16:433288:d=2021032312:WVHGT:surface:57 hour fcst: -0: 17:457350:d=2021032312:SWELL:1 in sequence:57 hour fcst: -0: 18:477633:d=2021032312:SWELL:2 in sequence:57 hour fcst: -0: 19:493916:d=2021032312:SWELL:3 in sequence:57 hour fcst: -0: 20:507534:d=2021032312:WVPER:surface:57 hour fcst: -0: 21:533907:d=2021032312:SWPER:1 in sequence:57 hour fcst: -0: 22:555894:d=2021032312:SWPER:2 in sequence:57 hour fcst: -0: 23:573832:d=2021032312:SWPER:3 in sequence:57 hour fcst: -0: 24:587956:d=2021032312:WVDIR:surface:57 hour fcst: -0: 25:625164:d=2021032312:SWDIR:1 in sequence:57 hour fcst: -0: 26:656491:d=2021032312:SWDIR:2 in sequence:57 hour fcst: -0: 27:679898:d=2021032312:SWDIR:3 in sequence:57 hour fcst: -0: + wave_grib2_sbs.sh[104]err=0 -0: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.atlocn.0p16.f057.grib2 -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f057.grib2 ]] -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f057.grib2.idx ]] -0: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.atlocn.0p16.f057.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f057.grib2 -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f057.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f057.grib2 = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f057.grib2 ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f057.grib2 -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f057.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f057.grib2.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f057.grib2.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f057.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f057.grib2 -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.atlocn.0p16.f057.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f057.grib2.idx -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f057.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f057.grib2.idx = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f057.grib2.idx ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f057.grib2.idx -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f057.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f057.grib2.idx.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f057.grib2.idx.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f057.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f057.grib2.idx -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.atlocn.0p16.f057.grib2 and gfs.wave.t12z.atlocn.0p16.f057.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_at_10m to COM' -0: INFO: Copied gfs.wave.t12z.atlocn.0p16.f057.grib2 and gfs.wave.t12z.atlocn.0p16.f057.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_at_10m to COM -0: + wave_grib2_sbs.sh[130][[ at_10m == '' ]] -0: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -0: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.atlocn.0p16.f057.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -0: INFO: gfs.wave.t12z.atlocn.0p16.f057.grib2 is global.0p50 or SENDDBN is NO, no alert sent -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + cmdfile.5[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032521 3600. 9999 -4: + cmdfile.5[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grid_interp_sbs.sh[25]grdID=aoc_9km -4: + wave_grid_interp_sbs.sh[26]valid_time=2021032521 -4: + wave_grid_interp_sbs.sh[27]dt=3600. -4: + wave_grid_interp_sbs.sh[28]nst=9999 -4: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463 -4: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.uglo_100km ./out_grd.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mod_def.uglo_100km ./mod_def.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mod_def.aoc_9km ./mod_def.aoc_9km -4: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km'\''' -4: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km' -4: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[51]weights_found=1 -4: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 210000' -4: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 210000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/aoc_9km/g ww3_gint.inp.tmpl -4: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -4: $ Input file for interpolation of uglo_100km to aoc_9km -4: $------------------------------------------------ -4: $ Start Time 3600. NSteps -4: 20210325 210000 3600. 9999 -4: $ Total number of grids -4: 2 -4: $ Grid extensions -4: 'uglo_100km' -4: 'aoc_9km' -4: $ -4: 0 -4: $ -4: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[70]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.454158 ']' -4: ++ prep_step[4]echo gfs_ww3_gint.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -4: INFO: Executing 'gfs_ww3_gint.x' -4: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[73]cat grid_interp.aoc_9km.out -4: -4: *** WAVEWATCH III Grid interpolation *** -4: =============================================== -4: -4: Comment character is '$' -4: -4: Time Information : -4: --------------------------------------------- -4: Starting Time : 2021/03/25 21:00:00 UTC -4: Interval (in sec) : 3600.00 -4: Number of requests : 9999 -4: --------------------------------------------- -4: Number of grids (including output grid) = 2 -4: -4: -4: Extension for grid 1 is --> uglo_100km -4: -4: Grid Particulars are : -4: Dimensions = 45166 1 -4: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Extension for grid 2 is --> aoc_9km -4: -4: Grid Particulars are : -4: Dimensions = 1006 1006 -4: Grid Type = 2 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -4: -4: -4: Preparing interpolation weights for output grid -4: Total number of wet points for interpolation 360052 -4: -4: -4: Variable: Grid Interpolation Map Units: 0.100E+01 -4: -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: *** | | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: 880 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 | -4: 754 | 0 0 0 | -4: | 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 628 | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 | -4: 502 | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 376 | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 250 | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 124 | 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 | -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: -4: -4: Interpolating fields .... -4: -4: Output group 1 -4: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -4: Output group 2 -4: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -4: Output group 3 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 4 -4: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -4: Output group 5 -4: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -4: Output group 6 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 7 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 8 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 9 -4: Output variables skipped -4: Output group 10 -4: Output variables skipped -4: ------------------------------------------------ -4: 1Current vel. -4: 1Wind speed -4: 1Ice concentration -4: 2Wave height -4: 2Mean wave period(+2) -4: 2Mean wave period(+1) -4: 2Peak frequency -4: 2Mean wave dir. a1b1 -4: 2Peak direction -4: 4Part. wave height -4: 4Part. peak period -4: 4Part. mean direction -4: 5Charnock parameter -4: ------------------------------------------------ -4: OUTPUT TIME : 2021/03/25 21:00:00 UTC -4: -4: End of file reached -4: -4: -4: *** End of Grid interpolation Routine *** -4: =============================================== -4: -4: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -4: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -4: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_aoc_9km/out_grd.aoc_9km /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.aoc_9km -4: + cmdfile.5[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032521 57 arctic 9km 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + cmdfile.5[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib2_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grib2_sbs.sh[30]grdID=aoc_9km -4: + wave_grib2_sbs.sh[31]GRIDNR=255 -4: + wave_grib2_sbs.sh[32]MODNR=11 -4: + wave_grib2_sbs.sh[33]valid_time=2021032521 -4: + wave_grib2_sbs.sh[34]fhr=57 -4: + wave_grib2_sbs.sh[35]grid_region=arctic -4: + wave_grib2_sbs.sh[36]grid_res=9km -4: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463 -4: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_aoc_9km -4: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_aoc_9km -4: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_aoc_9km -4: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_aoc_9km -4: ++ wave_grib2_sbs.sh[47]printf %03i 57 -4: + wave_grib2_sbs.sh[47]FH3=057 -4: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_arctic_9km -4: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.arctic.9km.f057.grib2 -4: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f057.grib2 ]] -4: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mod_def.aoc_9km ./mod_def.ww3 -4: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.aoc_9km ./out_grd.ww3 -4: + wave_grib2_sbs.sh[73]ngrib=1 -4: + wave_grib2_sbs.sh[74]dtgrib=3600 -4: + wave_grib2_sbs.sh[75]tstart='20210325 210000' -4: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 210000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.aoc_9km.inp.tmpl -4: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -4: $ WAVEWATCH-III gridded output input file -4: $ ---------------------------------------- -4: 20210325 210000 3600 1 -4: N -4: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -4: $ -4: 20210325 210000 7 11 255 0 20 -4: $ -4: 70 0 9.0 9.0 64 -4: $ 60 0 8.64919046313 8.64919046313 64 -4: $ end of input file -4: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[88]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.454158 ']' -4: ++ prep_step[4]echo gfs_ww3_grib.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[90]export err=0 -4: + wave_grib2_sbs.sh[90]err=0 -4: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[95]cat grib2_arctic_057.out -4: -4: *** WAVEWATCH III GRIB output postp. *** -4: ============================================== -4: -4: Comment character is '$' -4: -4: Grid name : Arctic Ocean PolarStereo 9km -4: -4: LINEIN: -4: 20210325 210000 3600 1 -4: -4: 20210325210000 3600 1 -4: GEN_PRO -99999 -4: -4: Output time data : -4: ----------------------------------------------------- -4: First time : 2021/03/25 21:00:00 UTC -4: Interval : 01:00:00 -4: Number of requests : 1 -4: Fields : Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: Requested output fields not yet available: -4: ----------------------------------------------------- -4: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -4: -4: Successfully requested output fields : -4: ----------------------------------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: -4: Additional GRIB parameters : -4: ----------------------------------------------------- -4: Run time : 2021/03/25 21:00:00 UTC -4: GRIB center ID : 7 -4: GRIB gen. proc. ID : 11 -4: GRIB grid ID : 255 -4: GRIB GDS parameter : 0 -4: Fields in file : -4: -------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: CHOSEN GRID TYPE: : POLARSTEREO -4: -4: -4: -4: Generating file -4: ----------------------------------------------------- -4: Data for 2021/03/25 21:00:00 UTC 0H forecast. -4: -4: End of program -4: ========================================= -4: WAVEWATCH III GRIB output -4: -4: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -4: + wave_grib2_sbs.sh[102][[ 57 -gt 0 ]] -4: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '57 hour fcst' -grib gfs.wave.t12z.arctic.9km.f057.grib2 -4: 1:0:d=2021032312:SPC:surface:57 hour fcst: -4: 2:192030:d=2021032312:DIRC:surface:57 hour fcst: -4: 3:662066:d=2021032312:UOGRD:surface:57 hour fcst: -4: 4:851575:d=2021032312:VOGRD:surface:57 hour fcst: -4: 5:1045057:d=2021032312:WIND:surface:57 hour fcst: -4: 6:1344531:d=2021032312:WDIR:surface:57 hour fcst: -4: 7:1772455:d=2021032312:UGRD:surface:57 hour fcst: -4: 8:2069933:d=2021032312:VGRD:surface:57 hour fcst: -4: 9:2367357:d=2021032312:ICEC:surface:57 hour fcst: -4: 10:2537859:d=2021032312:HTSGW:surface:57 hour fcst: -4: 11:2748037:d=2021032312:IMWF:surface:57 hour fcst: -4: 12:2962202:d=2021032312:MWSPER:surface:57 hour fcst: -4: 13:3178523:d=2021032312:PERPW:surface:57 hour fcst: -4: 14:3404145:d=2021032312:WWSDIR:surface:57 hour fcst: -4: 15:3704585:d=2021032312:DIRPW:surface:57 hour fcst: -4: 16:4013068:d=2021032312:WVHGT:surface:57 hour fcst: -4: 17:4220418:d=2021032312:SWELL:1 in sequence:57 hour fcst: -4: 18:4412072:d=2021032312:SWELL:2 in sequence:57 hour fcst: -4: 19:4563381:d=2021032312:SWELL:3 in sequence:57 hour fcst: -4: 20:4691970:d=2021032312:WVPER:surface:57 hour fcst: -4: 21:4909168:d=2021032312:SWPER:1 in sequence:57 hour fcst: -4: 22:5110758:d=2021032312:SWPER:2 in sequence:57 hour fcst: -4: 23:5269225:d=2021032312:SWPER:3 in sequence:57 hour fcst: -4: 24:5398625:d=2021032312:WVDIR:surface:57 hour fcst: -4: 25:5679104:d=2021032312:SWDIR:1 in sequence:57 hour fcst: -4: 26:5952822:d=2021032312:SWDIR:2 in sequence:57 hour fcst: -4: 27:6139557:d=2021032312:SWDIR:3 in sequence:57 hour fcst: -4: + wave_grib2_sbs.sh[104]err=0 -4: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.arctic.9km.f057.grib2 -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f057.grib2 ]] -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f057.grib2.idx ]] -4: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.arctic.9km.f057.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f057.grib2 -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f057.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f057.grib2 = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f057.grib2 ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f057.grib2 -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f057.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f057.grib2.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f057.grib2.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f057.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f057.grib2 -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.arctic.9km.f057.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f057.grib2.idx -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f057.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f057.grib2.idx = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f057.grib2.idx ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f057.grib2.idx -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f057.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f057.grib2.idx.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f057.grib2.idx.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f057.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f057.grib2.idx -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.arctic.9km.f057.grib2 and gfs.wave.t12z.arctic.9km.f057.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_aoc_9km to COM' -4: INFO: Copied gfs.wave.t12z.arctic.9km.f057.grib2 and gfs.wave.t12z.arctic.9km.f057.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_aoc_9km to COM -4: + wave_grib2_sbs.sh[130][[ aoc_9km == '' ]] -4: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -4: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.arctic.9km.f057.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -4: INFO: gfs.wave.t12z.arctic.9km.f057.grib2 is global.0p50 or SENDDBN is NO, no alert sent -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + cmdfile.7[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032521 3600. 9999 -6: + cmdfile.7[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grid_interp_sbs.sh[25]grdID=gsh_15m -6: + wave_grid_interp_sbs.sh[26]valid_time=2021032521 -6: + wave_grid_interp_sbs.sh[27]dt=3600. -6: + wave_grid_interp_sbs.sh[28]nst=9999 -6: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463 -6: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.uglo_100km ./out_grd.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mod_def.uglo_100km ./mod_def.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mod_def.gsh_15m ./mod_def.gsh_15m -6: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m'\''' -6: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m' -6: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[51]weights_found=1 -6: + wave_grid_interp_sbs.sh[59]ymdhms='20210325 210000' -6: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210325 210000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gsh_15m/g ww3_gint.inp.tmpl -6: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -6: $ Input file for interpolation of uglo_100km to gsh_15m -6: $------------------------------------------------ -6: $ Start Time 3600. NSteps -6: 20210325 210000 3600. 9999 -6: $ Total number of grids -6: 2 -6: $ Grid extensions -6: 'uglo_100km' -6: 'gsh_15m' -6: $ -6: 0 -6: $ -6: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[70]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.454158 ']' -6: ++ prep_step[4]echo gfs_ww3_gint.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -6: INFO: Executing 'gfs_ww3_gint.x' -6: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[73]cat grid_interp.gsh_15m.out -6: -6: *** WAVEWATCH III Grid interpolation *** -6: =============================================== -6: -6: Comment character is '$' -6: -6: Time Information : -6: --------------------------------------------- -6: Starting Time : 2021/03/25 21:00:00 UTC -6: Interval (in sec) : 3600.00 -6: Number of requests : 9999 -6: --------------------------------------------- -6: Number of grids (including output grid) = 2 -6: -6: -6: Extension for grid 1 is --> uglo_100km -6: -6: Grid Particulars are : -6: Dimensions = 45166 1 -6: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Extension for grid 2 is --> gsh_15m -6: -6: Grid Particulars are : -6: Dimensions = 1440 277 -6: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -6: -6: -6: Preparing interpolation weights for output grid -6: Total number of wet points for interpolation 317192 -6: -6: -6: Variable: Grid Interpolation Map Units: 0.100E+01 -6: -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 277 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 241 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 205 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 169 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 133 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 97 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 61 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 | -6: 25 | 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 | -6: | | -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: -6: -6: Interpolating fields .... -6: -6: Output group 1 -6: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -6: Output group 2 -6: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -6: Output group 3 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 4 -6: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -6: Output group 5 -6: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -6: Output group 6 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 7 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 8 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 9 -6: Output variables skipped -6: Output group 10 -6: Output variables skipped -6: ------------------------------------------------ -6: 1Current vel. -6: 1Wind speed -6: 1Ice concentration -6: 2Wave height -6: 2Mean wave period(+2) -6: 2Mean wave period(+1) -6: 2Peak frequency -6: 2Mean wave dir. a1b1 -6: 2Peak direction -6: 4Part. wave height -6: 4Part. peak period -6: 4Part. mean direction -6: 5Charnock parameter -6: ------------------------------------------------ -6: OUTPUT TIME : 2021/03/25 21:00:00 UTC -6: -6: End of file reached -6: -6: -6: *** End of Grid interpolation Routine *** -6: =============================================== -6: -6: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -6: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -6: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grid_interp_gsh_15m/out_grd.gsh_15m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.gsh_15m -6: + cmdfile.7[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032521 57 gsouth 0p25 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + cmdfile.7[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib2_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grib2_sbs.sh[30]grdID=gsh_15m -6: + wave_grib2_sbs.sh[31]GRIDNR=255 -6: + wave_grib2_sbs.sh[32]MODNR=11 -6: + wave_grib2_sbs.sh[33]valid_time=2021032521 -6: + wave_grib2_sbs.sh[34]fhr=57 -6: + wave_grib2_sbs.sh[35]grid_region=gsouth -6: + wave_grib2_sbs.sh[36]grid_res=0p25 -6: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463 -6: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_gsh_15m -6: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_gsh_15m -6: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_gsh_15m -6: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_gsh_15m -6: ++ wave_grib2_sbs.sh[47]printf %03i 57 -6: + wave_grib2_sbs.sh[47]FH3=057 -6: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -6: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.gsouth.0p25.f057.grib2 -6: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f057.grib2 ]] -6: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/mod_def.gsh_15m ./mod_def.ww3 -6: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/out_grd.gsh_15m ./out_grd.ww3 -6: + wave_grib2_sbs.sh[73]ngrib=1 -6: + wave_grib2_sbs.sh[74]dtgrib=3600 -6: + wave_grib2_sbs.sh[75]tstart='20210325 210000' -6: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210325 210000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gsh_15m.inp.tmpl -6: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -6: $ WAVEWATCH-III gridded output input file -6: $ ---------------------------------------- -6: 20210325 210000 3600 1 -6: N -6: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -6: $ -6: 20210325 210000 7 11 255 0 0 -6: $ -6: $ end of input file -6: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[88]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.454158 ']' -6: ++ prep_step[4]echo gfs_ww3_grib.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[90]export err=0 -6: + wave_grib2_sbs.sh[90]err=0 -6: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[95]cat grib2_gsouth_057.out -6: -6: *** WAVEWATCH III GRIB output postp. *** -6: ============================================== -6: -6: Comment character is '$' -6: -6: Grid name : GFSv16-wave S Hemisphere 1/4 d -6: -6: LINEIN: -6: 20210325 210000 3600 1 -6: -6: 20210325210000 3600 1 -6: GEN_PRO -99999 -6: -6: Output time data : -6: ----------------------------------------------------- -6: First time : 2021/03/25 21:00:00 UTC -6: Interval : 01:00:00 -6: Number of requests : 1 -6: Fields : Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: Requested output fields not yet available: -6: ----------------------------------------------------- -6: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -6: -6: Successfully requested output fields : -6: ----------------------------------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: -6: Additional GRIB parameters : -6: ----------------------------------------------------- -6: Run time : 2021/03/25 21:00:00 UTC -6: GRIB center ID : 7 -6: GRIB gen. proc. ID : 11 -6: GRIB grid ID : 255 -6: GRIB GDS parameter : 0 -6: Fields in file : -6: -------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: CHOSEN GRID TYPE: : LLRECTILINEAR -6: -6: -6: -6: Generating file -6: ----------------------------------------------------- -6: Data for 2021/03/25 21:00:00 UTC 0H forecast. -6: -6: End of program -6: ========================================= -6: WAVEWATCH III GRIB output -6: -6: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -6: + wave_grib2_sbs.sh[102][[ 57 -gt 0 ]] -6: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '57 hour fcst' -grib gfs.wave.t12z.gsouth.0p25.f057.grib2 -6: 1:0:d=2021032312:SPC:surface:57 hour fcst: -6: 2:117200:d=2021032312:DIRC:surface:57 hour fcst: -6: 3:500448:d=2021032312:UOGRD:surface:57 hour fcst: -6: 4:614201:d=2021032312:VOGRD:surface:57 hour fcst: -6: 5:735692:d=2021032312:WIND:surface:57 hour fcst: -6: 6:981521:d=2021032312:WDIR:surface:57 hour fcst: -6: 7:1359877:d=2021032312:UGRD:surface:57 hour fcst: -6: 8:1599086:d=2021032312:VGRD:surface:57 hour fcst: -6: 9:1842058:d=2021032312:ICEC:surface:57 hour fcst: -6: 10:1903227:d=2021032312:HTSGW:surface:57 hour fcst: -6: 11:2060919:d=2021032312:IMWF:surface:57 hour fcst: -6: 12:2231180:d=2021032312:MWSPER:surface:57 hour fcst: -6: 13:2402260:d=2021032312:PERPW:surface:57 hour fcst: -6: 14:2592556:d=2021032312:WWSDIR:surface:57 hour fcst: -6: 15:2935854:d=2021032312:DIRPW:surface:57 hour fcst: -6: 16:3295777:d=2021032312:WVHGT:surface:57 hour fcst: -6: 17:3461745:d=2021032312:SWELL:1 in sequence:57 hour fcst: -6: 18:3623456:d=2021032312:SWELL:2 in sequence:57 hour fcst: -6: 19:3728584:d=2021032312:SWELL:3 in sequence:57 hour fcst: -6: 20:3791013:d=2021032312:WVPER:surface:57 hour fcst: -6: 21:3978942:d=2021032312:SWPER:1 in sequence:57 hour fcst: -6: 22:4162050:d=2021032312:SWPER:2 in sequence:57 hour fcst: -6: 23:4292036:d=2021032312:SWPER:3 in sequence:57 hour fcst: -6: 24:4362659:d=2021032312:WVDIR:surface:57 hour fcst: -6: 25:4666978:d=2021032312:SWDIR:1 in sequence:57 hour fcst: -6: 26:4996434:d=2021032312:SWDIR:2 in sequence:57 hour fcst: -6: 27:5212784:d=2021032312:SWDIR:3 in sequence:57 hour fcst: -6: + wave_grib2_sbs.sh[104]err=0 -6: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.gsouth.0p25.f057.grib2 -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f057.grib2 ]] -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f057.grib2.idx ]] -6: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.gsouth.0p25.f057.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f057.grib2 -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f057.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f057.grib2 = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f057.grib2 ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f057.grib2 -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f057.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f057.grib2.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f057.grib2.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f057.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f057.grib2 -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.gsouth.0p25.f057.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f057.grib2.idx -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f057.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f057.grib2.idx = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f057.grib2.idx ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f057.grib2.idx -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f057.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f057.grib2.idx.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f057.grib2.idx.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f057.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f057.grib2.idx -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.gsouth.0p25.f057.grib2 and gfs.wave.t12z.gsouth.0p25.f057.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_gsh_15m to COM' -6: INFO: Copied gfs.wave.t12z.gsouth.0p25.f057.grib2 and gfs.wave.t12z.gsouth.0p25.f057.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463/grib_gsh_15m to COM -6: + wave_grib2_sbs.sh[130][[ gsh_15m == '' ]] -6: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -6: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.gsouth.0p25.f057.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -6: INFO: gfs.wave.t12z.gsouth.0p25.f057.grib2 is global.0p50 or SENDDBN is NO, no alert sent -+ run_mpmd.sh[113]exit 0 -+ run_mpmd.sh[1]postamble run_mpmd.sh 1753758278 0 -+ preamble.sh[62]set +x -End run_mpmd.sh at 03:04:49 with error code 0 (time elapsed: 00:00:11) -+ exgfs_wave_post_gridded_sbs.sh[122]true -+ exgfs_wave_post_gridded_sbs.sh[123]export err=0 -+ exgfs_wave_post_gridded_sbs.sh[123]err=0 -+ exgfs_wave_post_gridded_sbs.sh[124][[ 0 -ne 0 ]] -+ exgfs_wave_post_gridded_sbs.sh[130]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ exgfs_wave_post_gridded_sbs.sh[131]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ exgfs_wave_post_gridded_sbs.sh[132]gribchk=gfs.wave.t12z.gsouth.0p25.f057.grib2 -+ exgfs_wave_post_gridded_sbs.sh[133][[ ! -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f057.grib2 ]] -+ exgfs_wave_post_gridded_sbs.sh[138]exit 0 -+ JGLOBAL_WAVE_POST_SBS[28]true -+ JGLOBAL_WAVE_POST_SBS[29]export err=0 -+ JGLOBAL_WAVE_POST_SBS[29]err=0 -+ JGLOBAL_WAVE_POST_SBS[30][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[37]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312 -+ JGLOBAL_WAVE_POST_SBS[38][[ NO != \Y\E\S ]] -+ JGLOBAL_WAVE_POST_SBS[39]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f057.453463 -+ JGLOBAL_WAVE_POST_SBS[42]exit 0 -+ JGLOBAL_WAVE_POST_SBS[1]postamble /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS 1753758266 0 -+ preamble.sh[62]set +x -End /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at 03:04:49 with error code 0 (time elapsed: 00:00:23) -Begin /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at Tue Jul 29 03:04:49 UTC 2025 -++ jjob_header.sh[46]OPTIND=1 -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[50]env_job=wavepostsbs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[49]read -ra configs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[61]shift 4 -++ jjob_header.sh[63][[ -z wavepostsbs ]] -++ jjob_header.sh[71]export DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463 -++ jjob_header.sh[71]DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463 -++ jjob_header.sh[72][[ YES == \Y\E\S ]] -++ jjob_header.sh[73]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463 -++ jjob_header.sh[75]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463 -++ jjob_header.sh[76]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463 -++ jjob_header.sh[85]export pid=457148 -++ jjob_header.sh[85]pid=457148 -++ jjob_header.sh[86]export pgmout=OUTPUT.457148 -++ jjob_header.sh[86]pgmout=OUTPUT.457148 -++ jjob_header.sh[87]export pgmerr=errfile -++ jjob_header.sh[87]pgmerr=errfile -++ jjob_header.sh[90]export pgm= -++ jjob_header.sh[90]pgm= -++ jjob_header.sh[96]export cycle=t12z -++ jjob_header.sh[96]cycle=t12z -++ jjob_header.sh[97]setpdy.sh -+ setpdy.sh[20]'[' /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463 == /home/mterry ']' -+ setpdy.sh[25][[ ! t12z =~ t??z ]] -+ setpdy.sh[30]case $# in -+ setpdy.sh[31]dates_before_PDY=7 -+ setpdy.sh[32]dates_after_PDY=7 -+ setpdy.sh[50]COMDATEROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+ setpdy.sh[53]'[' -z 20210323 ']' -+ setpdy.sh[57]sed 's/[0-9]\{8\}/20210323/' /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z -sed: can't read /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z: No such file or directory -++ jjob_header.sh[97]true -++ jjob_header.sh[98]source ./PDY -/work2/noaa/global/mterry/global-workflow_forked/ush/jjob_header.sh: line 98: ./PDY: No such file or directory -++ jjob_header.sh[98]true -++ jjob_header.sh[104]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[104]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.base -+++ config.base[6]echo 'BEGIN: config.base' -BEGIN: config.base -+++ config.base[9]export machine=HERCULES -+++ config.base[9]machine=HERCULES -+++ config.base[12]export RUN_ENVIR=emc -+++ config.base[12]RUN_ENVIR=emc -+++ config.base[15]export ACCOUNT=fv3-cpu -+++ config.base[15]ACCOUNT=fv3-cpu -+++ config.base[16]export QUEUE=batch -+++ config.base[16]QUEUE=batch -+++ config.base[17]export QUEUE_SERVICE=batch -+++ config.base[17]QUEUE_SERVICE=batch -+++ config.base[18]export QUEUE_DTN=batch -+++ config.base[18]QUEUE_DTN=batch -+++ config.base[19]export PARTITION_BATCH=hercules -+++ config.base[19]PARTITION_BATCH=hercules -+++ config.base[20]export PARTITION_SERVICE=service -+++ config.base[20]PARTITION_SERVICE=service -+++ config.base[21]export PARTITION_DTN= -+++ config.base[21]PARTITION_DTN= -+++ config.base[22]export RESERVATION= -+++ config.base[22]RESERVATION= -+++ config.base[23]export CLUSTERS= -+++ config.base[23]CLUSTERS= -+++ config.base[24]export CLUSTERS_SERVICE= -+++ config.base[24]CLUSTERS_SERVICE= -+++ config.base[25]export CLUSTERS_DTN= -+++ config.base[25]CLUSTERS_DTN= -+++ config.base[28]export HPSS_PROJECT=emc-global -+++ config.base[28]HPSS_PROJECT=emc-global -+++ config.base[31]export HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[31]HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[32]export EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[32]EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[33]export FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[33]FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[34]export PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[34]PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[35]export SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[35]SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[36]export USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[36]USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[38]export FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[38]FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[39]export FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[39]FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[40]export FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[40]FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[41]export FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[41]FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[42]export FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[42]FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[43]export FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[43]FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[44]export FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[44]FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[45]export FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[45]FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[50]export PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[50]PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[51]export COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[51]COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[52]export COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[52]COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[53]export DMPDIR=/work/noaa/rstprod/dump -+++ config.base[53]DMPDIR=/work/noaa/rstprod/dump -+++ config.base[57]export COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[57]COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[58]export COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[58]COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[59]export COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[59]COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[62]export HOMEDIR=/work2/noaa/global/mterry -+++ config.base[62]HOMEDIR=/work2/noaa/global/mterry -+++ config.base[63]export STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[63]STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]export PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[65]export NOSCRUB=/work2/noaa/global/mterry -+++ config.base[65]NOSCRUB=/work2/noaa/global/mterry -+++ config.base[68]export BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[68]BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[71]export BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[71]BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[74]export DO_PREP_SFC=NO -+++ config.base[74]DO_PREP_SFC=NO -+++ config.base[77]export DO_GOES=NO -+++ config.base[77]DO_GOES=NO -+++ config.base[78]export DO_BUFRSND=NO -+++ config.base[78]DO_BUFRSND=NO -+++ config.base[79]export DO_GEMPAK=NO -+++ config.base[79]DO_GEMPAK=NO -+++ config.base[80]export DO_AWIPS=NO -+++ config.base[80]DO_AWIPS=NO -+++ config.base[81]export DO_NPOESS=NO -+++ config.base[81]DO_NPOESS=NO -+++ config.base[82]export DO_TRACKER=YES -+++ config.base[82]DO_TRACKER=YES -+++ config.base[83]export DO_GENESIS=YES -+++ config.base[83]DO_GENESIS=YES -+++ config.base[84]export DO_GENESIS_FSU=NO -+++ config.base[84]DO_GENESIS_FSU=NO -+++ config.base[85]export DO_VERFOZN=YES -+++ config.base[85]DO_VERFOZN=YES -+++ config.base[86]export DO_VERFRAD=YES -+++ config.base[86]DO_VERFRAD=YES -+++ config.base[87]export DO_VMINMON=YES -+++ config.base[87]DO_VMINMON=YES -+++ config.base[88]export DO_ANLSTAT=NO -+++ config.base[88]DO_ANLSTAT=NO -+++ config.base[91]export MODE=forecast-only -+++ config.base[91]MODE=forecast-only -+++ config.base[92]export DO_TEST_MODE=YES -+++ config.base[92]DO_TEST_MODE=YES -+++ config.base[101]export FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[101]FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[102]export HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[102]HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[103]export HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[103]HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[106]export NMV=/bin/mv -+++ config.base[106]NMV=/bin/mv -+++ config.base[107]export 'NLN=/bin/ln -sf' -+++ config.base[107]NLN='/bin/ln -sf' -+++ config.base[108]export VERBOSE=YES -+++ config.base[108]VERBOSE=YES -+++ config.base[109]export KEEPDATA=NO -+++ config.base[109]KEEPDATA=NO -+++ config.base[110]export DEBUG_POSTSCRIPT=NO -+++ config.base[110]DEBUG_POSTSCRIPT=NO -+++ config.base[111]export CHGRP_RSTPROD=YES -+++ config.base[111]CHGRP_RSTPROD=YES -+++ config.base[112]export 'CHGRP_CMD=chgrp rstprod' -+++ config.base[112]CHGRP_CMD='chgrp rstprod' -+++ config.base[113]export NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[113]NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[114]export NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[114]NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[117]export BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[117]BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[120]export SDATE=2021032312 -+++ config.base[120]SDATE=2021032312 -+++ config.base[121]export EDATE=2021032312 -+++ config.base[121]EDATE=2021032312 -+++ config.base[122]export EXP_WARM_START=.false. -+++ config.base[122]EXP_WARM_START=.false. -+++ config.base[123]export assim_freq=6 -+++ config.base[123]assim_freq=6 -+++ config.base[124]export PSLOT=C48_S2SW -+++ config.base[124]PSLOT=C48_S2SW -+++ config.base[125]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[125]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[126]export ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[126]ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[127]export DUMP_SUFFIX= -+++ config.base[127]DUMP_SUFFIX= -+++ config.base[128][[ 2021032312 -ge 2019092100 ]] -+++ config.base[128][[ 2021032312 -le 2019110700 ]] -+++ config.base[131]export ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[131]ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[132]export ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[132]ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[133]export FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[133]FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[136]export envir=prod -+++ config.base[136]envir=prod -+++ config.base[137]export NET=gfs -+++ config.base[137]NET=gfs -+++ config.base[138]export RUN=gfs -+++ config.base[138]RUN=gfs -+++ config.base[141]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.com -++++ config.com[4]echo 'BEGIN: config.com' -BEGIN: config.com -++++ config.com[38][[ emc == \n\c\o ]] -++++ config.com[43]COM_OBSPROC_TMPL='${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos' -++++ config.com[44]COM_RTOFS_TMPL='${DMPDIR}' -++++ config.com[45]COM_TCVITAL_TMPL='${DMPDIR}/${RUN}.${YMD}/${HH}/atmos' -++++ config.com[47]declare -rx 'COM_OBS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/obs' -++++ config.com[48]declare -rx COM_OBSPROC_TMPL COM_RTOFS_TMPL -++++ config.com[50]COM_BASE='${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}' -++++ config.com[52]declare -rx 'COM_TOP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}' -++++ config.com[54]declare -rx 'COM_CONF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf' -++++ config.com[55]declare -rx 'COM_OBS_JEDI=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi' -++++ config.com[57]declare -rx 'COM_ATMOS_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input' -++++ config.com[58]declare -rx 'COM_ATMOS_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart' -++++ config.com[59]declare -rx 'COM_ATMOS_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos' -++++ config.com[60]declare -rx 'COM_SNOW_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow' -++++ config.com[61]declare -rx 'COM_SNOW_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon' -++++ config.com[62]declare -rx 'COM_ATMOS_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history' -++++ config.com[63]declare -rx 'COM_ATMOS_MASTER_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master' -++++ config.com[64]declare -rx 'COM_ATMOS_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2' -++++ config.com[65]declare -rx 'COM_ATMOS_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}' -++++ config.com[66]declare -rx 'COM_ATMOS_BUFR_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr' -++++ config.com[67]declare -rx 'COM_ATMOS_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}' -++++ config.com[68]declare -rx 'COM_ATMOS_GENESIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital' -++++ config.com[69]declare -rx 'COM_ATMOS_TRACK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks' -++++ config.com[70]declare -rx 'COM_ATMOS_GOES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim' -++++ config.com[71]declare -rx 'COM_ATMOS_IMAGERY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery' -++++ config.com[72]declare -rx 'COM_ATMOS_OZNMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon' -++++ config.com[73]declare -rx 'COM_ATMOS_RADMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon' -++++ config.com[74]declare -rx 'COM_ATMOS_MINMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon' -++++ config.com[75]declare -rx 'COM_ATMOS_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon' -++++ config.com[76]declare -rx 'COM_ATMOS_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo' -++++ config.com[78]declare -rx 'COM_WAVE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart' -++++ config.com[79]declare -rx 'COM_WAVE_PREP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep' -++++ config.com[80]declare -rx 'COM_WAVE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history' -++++ config.com[81]declare -rx 'COM_WAVE_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded' -++++ config.com[82]declare -rx 'COM_WAVE_GRID_RES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}' -++++ config.com[83]declare -rx 'COM_WAVE_STATION_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station' -++++ config.com[84]declare -rx 'COM_WAVE_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak' -++++ config.com[85]declare -rx 'COM_WAVE_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo' -++++ config.com[87]declare -rx 'COM_OCEAN_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history' -++++ config.com[88]declare -rx 'COM_OCEAN_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart' -++++ config.com[89]declare -rx 'COM_OCEAN_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input' -++++ config.com[90]declare -rx 'COM_OCEAN_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean' -++++ config.com[91]declare -rx 'COM_OCEAN_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon' -++++ config.com[92]declare -rx 'COM_OCEAN_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf' -++++ config.com[93]declare -rx 'COM_OCEAN_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean' -++++ config.com[94]declare -rx 'COM_OCEAN_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf' -++++ config.com[95]declare -rx 'COM_OCEAN_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2' -++++ config.com[96]declare -rx 'COM_OCEAN_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}' -++++ config.com[98]declare -rx 'COM_ICE_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice' -++++ config.com[99]declare -rx 'COM_ICE_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf' -++++ config.com[100]declare -rx 'COM_ICE_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon' -++++ config.com[101]declare -rx 'COM_ICE_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice' -++++ config.com[102]declare -rx 'COM_ICE_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input' -++++ config.com[103]declare -rx 'COM_ICE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history' -++++ config.com[104]declare -rx 'COM_ICE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart' -++++ config.com[105]declare -rx 'COM_ICE_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf' -++++ config.com[106]declare -rx 'COM_ICE_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2' -++++ config.com[107]declare -rx 'COM_ICE_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}' -++++ config.com[109]declare -rx 'COM_CHEM_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history' -++++ config.com[110]declare -rx 'COM_CHEM_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem' -++++ config.com[111]declare -rx 'COM_CHEM_BMAT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix' -++++ config.com[112]declare -rx 'COM_CHEM_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon' -++++ config.com[114]declare -rx 'COM_MED_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart' -+++ config.base[143]export LOGSCRIPT= -+++ config.base[143]LOGSCRIPT= -+++ config.base[145]export 'REDOUT=1>' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wave -+++ config.wave[6]echo 'BEGIN: config.wave' -BEGIN: config.wave -+++ config.wave[13]export RUNRSTwave=gdas -+++ config.wave[13]RUNRSTwave=gdas -+++ config.wave[16]export MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[16]MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[19]case "${waveGRD}" in -+++ config.wave[64]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[64]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[65]export wavepostGRD= -+++ config.wave[65]wavepostGRD= -+++ config.wave[66]export waveuoutpGRD=uglo_100km -+++ config.wave[66]waveuoutpGRD=uglo_100km -+++ config.wave[75]export WAVEWND_DID= -+++ config.wave[75]WAVEWND_DID= -+++ config.wave[76]export WAVEWND_FID= -+++ config.wave[76]WAVEWND_FID= -+++ config.wave[79][[ gfs == \g\f\s ]] -+++ config.wave[80]export FHMAX_WAV=120 -+++ config.wave[80]FHMAX_WAV=120 -+++ config.wave[82]export WAVHINDH=0 -+++ config.wave[82]WAVHINDH=0 -+++ config.wave[83]export FHMAX_WAV_IBP=180 -+++ config.wave[83]FHMAX_WAV_IBP=180 -+++ config.wave[84](( FHMAX_WAV < FHMAX_WAV_IBP )) -+++ config.wave[84]export FHMAX_WAV_IBP=120 -+++ config.wave[84]FHMAX_WAV_IBP=120 -+++ config.wave[87]export DTFLD_WAV=3600 -+++ config.wave[87]DTFLD_WAV=3600 -+++ config.wave[88]export DTPNT_WAV=3600 -+++ config.wave[88]DTPNT_WAV=3600 -+++ config.wave[89]export FHINCP_WAV=1 -+++ config.wave[89]FHINCP_WAV=1 -+++ config.wave[92]export 'OUTPARS_WAV=WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[92]OUTPARS_WAV='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[95][[ gfs == \g\d\a\s ]] -+++ config.wave[99](( INTERVAL_GFS > 0 )) -+++ config.wave[100]export WAVHCYC=6 -+++ config.wave[100]WAVHCYC=6 -+++ config.wave[101]export FHMAX_WAV_CUR=192 -+++ config.wave[101]FHMAX_WAV_CUR=192 -+++ config.wave[109]export RSTTYPE_WAV=T -+++ config.wave[109]RSTTYPE_WAV=T -+++ config.wave[110][[ gfs != gfs ]] -+++ config.wave[115]rst_dt_gfs=43200 -+++ config.wave[116][[ 43200 -gt 0 ]] -+++ config.wave[117]export DT_1_RST_WAV=0 -+++ config.wave[117]DT_1_RST_WAV=0 -+++ config.wave[120]export DT_2_RST_WAV=43200 -+++ config.wave[120]DT_2_RST_WAV=43200 -+++ config.wave[126]export RSTIOFF_WAV=0 -+++ config.wave[126]RSTIOFF_WAV=0 -+++ config.wave[131]export RUNMEM=-1 -+++ config.wave[131]RUNMEM=-1 -+++ config.wave[134](( RUNMEM == -1 )) -+++ config.wave[136]export waveMEMB= -+++ config.wave[136]waveMEMB= -+++ config.wave[143]export WW3ATMINP=CPL -+++ config.wave[143]WW3ATMINP=CPL -+++ config.wave[144][[ YES == \Y\E\S ]] -+++ config.wave[145]export WW3ICEINP=CPL -+++ config.wave[145]WW3ICEINP=CPL -+++ config.wave[146]export WAVEICE_FID= -+++ config.wave[146]WAVEICE_FID= -+++ config.wave[152][[ YES == \Y\E\S ]] -+++ config.wave[153]export WW3CURINP=CPL -+++ config.wave[153]WW3CURINP=CPL -+++ config.wave[154]export WAVECUR_FID= -+++ config.wave[154]WAVECUR_FID= -+++ config.wave[161]export WW3ATMIENS=F -+++ config.wave[161]WW3ATMIENS=F -+++ config.wave[162]export WW3ICEIENS=F -+++ config.wave[162]WW3ICEIENS=F -+++ config.wave[163]export WW3CURIENS=F -+++ config.wave[163]WW3CURIENS=F -+++ config.wave[165]export GOFILETYPE=1 -+++ config.wave[165]GOFILETYPE=1 -+++ config.wave[166]export POFILETYPE=1 -+++ config.wave[166]POFILETYPE=1 -+++ config.wave[170]export FUNIPNT=T -+++ config.wave[170]FUNIPNT=T -+++ config.wave[172]export IOSRV=1 -+++ config.wave[172]IOSRV=1 -+++ config.wave[174]export FPNTPROC=T -+++ config.wave[174]FPNTPROC=T -+++ config.wave[176]export FGRDPROC=F -+++ config.wave[176]FGRDPROC=F -+++ config.wave[178]export FLAGMASKCOMP=F -+++ config.wave[178]FLAGMASKCOMP=F -+++ config.wave[180]export FLAGMASKOUT=F -+++ config.wave[180]FLAGMASKOUT=F -+++ config.wave[182]echo 'END: config.wave' -END: config.wave -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wavepostsbs -+++ config.wavepostsbs[6]echo 'BEGIN: config.wavepostsbs' -BEGIN: config.wavepostsbs -+++ config.wavepostsbs[9]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources wavepostsbs -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=wavepostsbs -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[156]ntasks=8 -++++ config.resources[157]threads_per_task=1 -++++ config.resources[158]tasks_per_node=80 -++++ config.resources[159]NTASKS=8 -++++ config.resources[160]memory=20GB -++++ config.resources[161]walltime=00:15:00 -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export NTASKS -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export memory -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.wavepostsbs[12]export MAX_TASKS=25 -+++ config.wavepostsbs[12]MAX_TASKS=25 -+++ config.wavepostsbs[15]export WAV_SUBGRBSRC= -+++ config.wavepostsbs[15]WAV_SUBGRBSRC= -+++ config.wavepostsbs[16]export WAV_SUBGRB= -+++ config.wavepostsbs[16]WAV_SUBGRB= -+++ config.wavepostsbs[19]export DOFLD_WAV=YES -+++ config.wavepostsbs[19]DOFLD_WAV=YES -+++ config.wavepostsbs[20]export DOPNT_WAV=YES -+++ config.wavepostsbs[20]DOPNT_WAV=YES -+++ config.wavepostsbs[21]export DOGRB_WAV=YES -+++ config.wavepostsbs[21]DOGRB_WAV=YES -+++ config.wavepostsbs[22][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+++ config.wavepostsbs[23]export DOGRI_WAV=YES -+++ config.wavepostsbs[23]DOGRI_WAV=YES -+++ config.wavepostsbs[27]export DOSPC_WAV=YES -+++ config.wavepostsbs[27]DOSPC_WAV=YES -+++ config.wavepostsbs[28]export DOBLL_WAV=YES -+++ config.wavepostsbs[28]DOBLL_WAV=YES -+++ config.wavepostsbs[30]echo 'END: config.wavepostsbs' -END: config.wavepostsbs -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env wavepostsbs -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=wavepostsbs -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 20971520 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 8 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[34]max_threads_per_task=1 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 1 ]] -+++ HERCULES.env[40][[ 1 -gt 1 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 8' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[63]export USE_CFP=YES -+++ HERCULES.env[63]USE_CFP=YES -+++ HERCULES.env[64][[ wavepostsbs == \w\a\v\e\p\r\e\p ]] -+++ HERCULES.env[65]export 'wavempexec=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[65]wavempexec='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[66]export 'wave_mpmd=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[66]wave_mpmd='--multi-prog --output=mpmd.%j.%t.out' -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[4]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ JGLOBAL_WAVE_POST_SBS[7]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[7]HH=12 -+ JGLOBAL_WAVE_POST_SBS[7]declare_from_tmpl -rx COMIN_WAVE_PREP:COM_WAVE_PREP_TMPL COMIN_WAVE_HISTORY:COM_WAVE_HISTORY_TMPL COMOUT_WAVE_GRID:COM_WAVE_GRID_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_WAVE_PREP=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep -declare_from_tmpl :: COMIN_WAVE_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history -declare_from_tmpl :: COMOUT_WAVE_GRID=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded -+ JGLOBAL_WAVE_POST_SBS[12][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n '' ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_atlocn_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_atlocn_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_atlocn_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_epacif_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_epacif_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_epacif_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_wcoast_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_wcoast_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_wcoast_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p50 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p50:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p50=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p50 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=arctic.9km -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_arctic_9km:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_arctic_9km=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_arctic_9km -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_gsouth_0p25:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_gsouth_0p25=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 ]] -+ JGLOBAL_WAVE_POST_SBS[28]/work2/noaa/global/mterry/global-workflow_forked/scripts/exgfs_wave_post_gridded_sbs.sh -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ exgfs_wave_post_gridded_sbs.sh[24]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ exgfs_wave_post_gridded_sbs.sh[26]DOGRI_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[27]DOGRB_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[29]export waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[29]waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[30]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[30]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[31]export wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[31]wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[33]cat - INFO: Grid information: - INFO: Native wave grids: uglo_100km - INFO: Interpolated grids: at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m - INFO: Post-process grids: -++ exgfs_wave_post_gridded_sbs.sh[40]printf %03i 60 -+ exgfs_wave_post_gridded_sbs.sh[40]fhr3=060 -++ exgfs_wave_post_gridded_sbs.sh[41]date -u -d '20210323 12 + 60 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[41]valid_time=2021032600 -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[49]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f060.bin ./out_grd.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f060.bin ./out_grd.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[52][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[62]cat - INFO: Summary: - INFO: Grid interp: DOGRI_WAV="YES" - INFO: Grib files: DOGRB_WAV="YES" - INFO: Fields to be included in grib files: - INFO: OUTPARS_WAV="WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA" -+ exgfs_wave_post_gridded_sbs.sh[70][[ YES == \N\O ]] -+ exgfs_wave_post_gridded_sbs.sh[76]rm -f 'cmdfile.*' cmdfile -+ exgfs_wave_post_gridded_sbs.sh[77]count=0 -+ exgfs_wave_post_gridded_sbs.sh[80][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[81]dt_int=3600. -+ exgfs_wave_post_gridded_sbs.sh[82]n_int=9999 -++ exgfs_wave_post_gridded_sbs.sh[83]date -u -d '20210326 00 - 0 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[83]ymdh_int=2021032600 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=1 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032600 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032600 60 atlocn 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib2_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib2_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=2 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032600 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032600 60 epacif 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib2_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib2_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=3 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032600 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032600 60 wcoast 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib2_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib2_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=4 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032600 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032600 60 global 0p50 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib2_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib2_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=5 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032600 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032600 60 arctic 9km '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib2_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib2_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=6 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032600 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032600 60 global 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib2_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib2_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=7 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032600 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032600 60 gsouth 0p25 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib2_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib2_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[100][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[113][[ 8 -lt 7 ]] -+ exgfs_wave_post_gridded_sbs.sh[121]echo 'INFO: Running MPMD job with 7 commands' -INFO: Running MPMD job with 7 commands -+ exgfs_wave_post_gridded_sbs.sh[122]/work2/noaa/global/mterry/global-workflow_forked/ush/run_mpmd.sh /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/cmdfile -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ run_mpmd.sh[31]source /work2/noaa/global/mterry/global-workflow_forked/ush/preamble.sh -++ preamble.sh[20]set +x -Begin run_mpmd.sh at Tue Jul 29 03:05:01 UTC 2025 -+ run_mpmd.sh[33]cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/cmdfile -+ run_mpmd.sh[36][[ YES != \Y\E\S ]] -+ run_mpmd.sh[46]export OMP_NUM_THREADS=1 -+ run_mpmd.sh[46]OMP_NUM_THREADS=1 -++ run_mpmd.sh[49]wc -l -+ run_mpmd.sh[49]nprocs=7 -+ run_mpmd.sh[52]mpmd_cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mpmd_cmdfile -+ run_mpmd.sh[53][[ -s /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mpmd_cmdfile ]] -+ run_mpmd.sh[55]cat - INFO: Executing MPMD job, STDOUT redirected for each process separately - INFO: On failure, logs for each job will be available in /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mpmd.proc_num.out - INFO: The proc_num corresponds to the line in '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mpmd_cmdfile' -+ run_mpmd.sh[61][[ srun -l --export=ALL --hint=nomultithread =~ ^srun.* ]] -+ run_mpmd.sh[65]nm=0 -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '0 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/cmdfile.1' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '1 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/cmdfile.2' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '2 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/cmdfile.3' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '3 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/cmdfile.4' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '4 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/cmdfile.5' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '5 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/cmdfile.6' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '6 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/cmdfile.7' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[72]set +e -+ run_mpmd.sh[74]srun -l --export=ALL --hint=nomultithread --multi-prog --output=mpmd.%j.%t.out -n 7 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mpmd_cmdfile -+ run_mpmd.sh[75]err=0 -+ run_mpmd.sh[76]set_strict -+ preamble.sh[35][[ YES == \Y\E\S ]] -+ preamble.sh[37]set -eu -+ run_mpmd.sh[103][[ 0 -eq 0 ]] -+ run_mpmd.sh[104]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mpmd_cmdfile -++ run_mpmd.sh[105]find . -name 'mpmd.*.out' -+ run_mpmd.sh[105]out_files='./mpmd.5951794.1.out -./mpmd.5951794.5.out -./mpmd.5951794.3.out -./mpmd.5951794.2.out -./mpmd.5951794.0.out -./mpmd.5951794.4.out -./mpmd.5951794.6.out' -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951794.1.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951794.1.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951794.5.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951794.5.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951794.3.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951794.3.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951794.2.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951794.2.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951794.0.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951794.0.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951794.4.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951794.4.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951794.6.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951794.6.out -+ run_mpmd.sh[110]cat mpmd.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + cmdfile.2[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032600 3600. 9999 -1: + cmdfile.2[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grid_interp_sbs.sh[25]grdID=ep_10m -1: + wave_grid_interp_sbs.sh[26]valid_time=2021032600 -1: + wave_grid_interp_sbs.sh[27]dt=3600. -1: + wave_grid_interp_sbs.sh[28]nst=9999 -1: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463 -1: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.uglo_100km ./out_grd.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mod_def.uglo_100km ./mod_def.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mod_def.ep_10m ./mod_def.ep_10m -1: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ]] -1: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m'\''' -1: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m' -1: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[51]weights_found=1 -1: + wave_grid_interp_sbs.sh[59]ymdhms='20210326 000000' -1: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210326 000000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/ep_10m/g ww3_gint.inp.tmpl -1: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -1: $ Input file for interpolation of uglo_100km to ep_10m -1: $------------------------------------------------ -1: $ Start Time 3600. NSteps -1: 20210326 000000 3600. 9999 -1: $ Total number of grids -1: 2 -1: $ Grid extensions -1: 'uglo_100km' -1: 'ep_10m' -1: $ -1: 0 -1: $ -1: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[70]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.457148 ']' -1: ++ prep_step[4]echo gfs_ww3_gint.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -1: INFO: Executing 'gfs_ww3_gint.x' -1: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[73]cat grid_interp.ep_10m.out -1: -1: *** WAVEWATCH III Grid interpolation *** -1: =============================================== -1: -1: Comment character is '$' -1: -1: Time Information : -1: --------------------------------------------- -1: Starting Time : 2021/03/26 00:00:00 UTC -1: Interval (in sec) : 3600.00 -1: Number of requests : 9999 -1: --------------------------------------------- -1: Number of grids (including output grid) = 2 -1: -1: -1: Extension for grid 1 is --> uglo_100km -1: -1: Grid Particulars are : -1: Dimensions = 45166 1 -1: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Extension for grid 2 is --> ep_10m -1: -1: Grid Particulars are : -1: Dimensions = 511 301 -1: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -1: -1: -1: Preparing interpolation weights for output grid -1: Total number of wet points for interpolation 7439 -1: -1: -1: Variable: Grid Interpolation Map Units: 0.100E+01 -1: -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 301 | | -1: | | -1: | 0 0 | -1: 262 | 0 0 0 0 | -1: | 0 0 0 0 | -1: | 0 0 0 0 | -1: 223 | 0 0 0 | -1: | 0 | -1: | 0 | -1: 184 | | -1: | 0 0 0 | -1: | 0 0 0 0 | -1: 145 | 0 | -1: | | -1: | | -1: 106 | | -1: | | -1: | | -1: 67 | | -1: | 0 | -1: | | -1: 28 | 0 | -1: | | -1: | | -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: -1: -1: Interpolating fields .... -1: -1: Output group 1 -1: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -1: Output group 2 -1: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -1: Output group 3 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 4 -1: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -1: Output group 5 -1: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -1: Output group 6 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 7 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 8 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 9 -1: Output variables skipped -1: Output group 10 -1: Output variables skipped -1: ------------------------------------------------ -1: 1Current vel. -1: 1Wind speed -1: 1Ice concentration -1: 2Wave height -1: 2Mean wave period(+2) -1: 2Mean wave period(+1) -1: 2Peak frequency -1: 2Mean wave dir. a1b1 -1: 2Peak direction -1: 4Part. wave height -1: 4Part. peak period -1: 4Part. mean direction -1: 5Charnock parameter -1: ------------------------------------------------ -1: OUTPUT TIME : 2021/03/26 00:00:00 UTC -1: -1: End of file reached -1: -1: -1: *** End of Grid interpolation Routine *** -1: =============================================== -1: -1: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -1: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -1: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_ep_10m/out_grd.ep_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.ep_10m -1: + cmdfile.2[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032600 60 epacif 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + cmdfile.2[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib2_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grib2_sbs.sh[30]grdID=ep_10m -1: + wave_grib2_sbs.sh[31]GRIDNR=255 -1: + wave_grib2_sbs.sh[32]MODNR=11 -1: + wave_grib2_sbs.sh[33]valid_time=2021032600 -1: + wave_grib2_sbs.sh[34]fhr=60 -1: + wave_grib2_sbs.sh[35]grid_region=epacif -1: + wave_grib2_sbs.sh[36]grid_res=0p16 -1: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463 -1: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_ep_10m -1: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_ep_10m -1: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_ep_10m -1: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_ep_10m -1: ++ wave_grib2_sbs.sh[47]printf %03i 60 -1: + wave_grib2_sbs.sh[47]FH3=060 -1: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_epacif_0p16 -1: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.epacif.0p16.f060.grib2 -1: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f060.grib2 ]] -1: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mod_def.ep_10m ./mod_def.ww3 -1: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.ep_10m ./out_grd.ww3 -1: + wave_grib2_sbs.sh[73]ngrib=1 -1: + wave_grib2_sbs.sh[74]dtgrib=3600 -1: + wave_grib2_sbs.sh[75]tstart='20210326 000000' -1: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210326 000000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.ep_10m.inp.tmpl -1: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -1: $ WAVEWATCH-III gridded output input file -1: $ ---------------------------------------- -1: 20210326 000000 3600 1 -1: N -1: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -1: $ -1: 20210326 000000 7 11 255 0 0 -1: $ -1: $ end of input file -1: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[88]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.457148 ']' -1: ++ prep_step[4]echo gfs_ww3_grib.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[90]export err=0 -1: + wave_grib2_sbs.sh[90]err=0 -1: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[95]cat grib2_epacif_060.out -1: -1: *** WAVEWATCH III GRIB output postp. *** -1: ============================================== -1: -1: Comment character is '$' -1: -1: Grid name : East Pacific 10 min wave grid -1: -1: LINEIN: -1: 20210326 000000 3600 1 -1: -1: 20210326000000 3600 1 -1: GEN_PRO -99999 -1: -1: Output time data : -1: ----------------------------------------------------- -1: First time : 2021/03/26 00:00:00 UTC -1: Interval : 01:00:00 -1: Number of requests : 1 -1: Fields : Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: Requested output fields not yet available: -1: ----------------------------------------------------- -1: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -1: -1: Successfully requested output fields : -1: ----------------------------------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: -1: Additional GRIB parameters : -1: ----------------------------------------------------- -1: Run time : 2021/03/26 00:00:00 UTC -1: GRIB center ID : 7 -1: GRIB gen. proc. ID : 11 -1: GRIB grid ID : 255 -1: GRIB GDS parameter : 0 -1: Fields in file : -1: -------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: CHOSEN GRID TYPE: : LLRECTILINEAR -1: -1: -1: -1: Generating file -1: ----------------------------------------------------- -1: Data for 2021/03/26 00:00:00 UTC 0H forecast. -1: -1: End of program -1: ========================================= -1: WAVEWATCH III GRIB output -1: -1: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -1: + wave_grib2_sbs.sh[102][[ 60 -gt 0 ]] -1: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '60 hour fcst' -grib gfs.wave.t12z.epacif.0p16.f060.grib2 -1: 1:0:d=2021032312:SPC:surface:60 hour fcst: -1: 2:21699:d=2021032312:DIRC:surface:60 hour fcst: -1: 3:49883:d=2021032312:UOGRD:surface:60 hour fcst: -1: 4:71649:d=2021032312:VOGRD:surface:60 hour fcst: -1: 5:93725:d=2021032312:WIND:surface:60 hour fcst: -1: 6:118350:d=2021032312:WDIR:surface:60 hour fcst: -1: 7:145999:d=2021032312:UGRD:surface:60 hour fcst: -1: 8:170577:d=2021032312:VGRD:surface:60 hour fcst: -1: 9:195142:d=2021032312:ICEC:surface:60 hour fcst: -1: 10:214550:d=2021032312:HTSGW:surface:60 hour fcst: -1: 11:237406:d=2021032312:IMWF:surface:60 hour fcst: -1: 12:260119:d=2021032312:MWSPER:surface:60 hour fcst: -1: 13:282879:d=2021032312:PERPW:surface:60 hour fcst: -1: 14:305797:d=2021032312:WWSDIR:surface:60 hour fcst: -1: 15:332894:d=2021032312:DIRPW:surface:60 hour fcst: -1: 16:360009:d=2021032312:WVHGT:surface:60 hour fcst: -1: 17:382235:d=2021032312:SWELL:1 in sequence:60 hour fcst: -1: 18:404531:d=2021032312:SWELL:2 in sequence:60 hour fcst: -1: 19:425276:d=2021032312:SWELL:3 in sequence:60 hour fcst: -1: 20:444886:d=2021032312:WVPER:surface:60 hour fcst: -1: 21:467112:d=2021032312:SWPER:1 in sequence:60 hour fcst: -1: 22:489627:d=2021032312:SWPER:2 in sequence:60 hour fcst: -1: 23:510553:d=2021032312:SWPER:3 in sequence:60 hour fcst: -1: 24:530194:d=2021032312:WVDIR:surface:60 hour fcst: -1: 25:554863:d=2021032312:SWDIR:1 in sequence:60 hour fcst: -1: 26:580947:d=2021032312:SWDIR:2 in sequence:60 hour fcst: -1: 27:603692:d=2021032312:SWDIR:3 in sequence:60 hour fcst: -1: + wave_grib2_sbs.sh[104]err=0 -1: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.epacif.0p16.f060.grib2 -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f060.grib2 ]] -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f060.grib2.idx ]] -1: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.epacif.0p16.f060.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f060.grib2 -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f060.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f060.grib2 = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f060.grib2 ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f060.grib2 -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f060.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f060.grib2.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f060.grib2.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f060.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f060.grib2 -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.epacif.0p16.f060.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f060.grib2.idx -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f060.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f060.grib2.idx = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f060.grib2.idx ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f060.grib2.idx -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f060.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f060.grib2.idx.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f060.grib2.idx.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f060.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f060.grib2.idx -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.epacif.0p16.f060.grib2 and gfs.wave.t12z.epacif.0p16.f060.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_ep_10m to COM' -1: INFO: Copied gfs.wave.t12z.epacif.0p16.f060.grib2 and gfs.wave.t12z.epacif.0p16.f060.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_ep_10m to COM -1: + wave_grib2_sbs.sh[130][[ ep_10m == '' ]] -1: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -1: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.epacif.0p16.f060.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -1: INFO: gfs.wave.t12z.epacif.0p16.f060.grib2 is global.0p50 or SENDDBN is NO, no alert sent -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + cmdfile.6[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032600 3600. 9999 -5: + cmdfile.6[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grid_interp_sbs.sh[25]grdID=gnh_10m -5: + wave_grid_interp_sbs.sh[26]valid_time=2021032600 -5: + wave_grid_interp_sbs.sh[27]dt=3600. -5: + wave_grid_interp_sbs.sh[28]nst=9999 -5: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463 -5: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.uglo_100km ./out_grd.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mod_def.uglo_100km ./mod_def.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mod_def.gnh_10m ./mod_def.gnh_10m -5: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m'\''' -5: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m' -5: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[51]weights_found=1 -5: + wave_grid_interp_sbs.sh[59]ymdhms='20210326 000000' -5: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210326 000000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gnh_10m/g ww3_gint.inp.tmpl -5: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -5: $ Input file for interpolation of uglo_100km to gnh_10m -5: $------------------------------------------------ -5: $ Start Time 3600. NSteps -5: 20210326 000000 3600. 9999 -5: $ Total number of grids -5: 2 -5: $ Grid extensions -5: 'uglo_100km' -5: 'gnh_10m' -5: $ -5: 0 -5: $ -5: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[70]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.457148 ']' -5: ++ prep_step[4]echo gfs_ww3_gint.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -5: INFO: Executing 'gfs_ww3_gint.x' -5: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[73]cat grid_interp.gnh_10m.out -5: -5: *** WAVEWATCH III Grid interpolation *** -5: =============================================== -5: -5: Comment character is '$' -5: -5: Time Information : -5: --------------------------------------------- -5: Starting Time : 2021/03/26 00:00:00 UTC -5: Interval (in sec) : 3600.00 -5: Number of requests : 9999 -5: --------------------------------------------- -5: Number of grids (including output grid) = 2 -5: -5: -5: Extension for grid 1 is --> uglo_100km -5: -5: Grid Particulars are : -5: Dimensions = 45166 1 -5: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Extension for grid 2 is --> gnh_10m -5: -5: Grid Particulars are : -5: Dimensions = 2160 406 -5: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -5: -5: -5: Preparing interpolation weights for output grid -5: Total number of wet points for interpolation 571209 -5: -5: -5: Variable: Grid Interpolation Map Units: 0.100E+01 -5: -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 406 | | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: 355 | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 304 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 253 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 202 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 151 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 100 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 49 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: -5: -5: Interpolating fields .... -5: -5: Output group 1 -5: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -5: Output group 2 -5: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -5: Output group 3 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 4 -5: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -5: Output group 5 -5: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -5: Output group 6 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 7 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 8 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 9 -5: Output variables skipped -5: Output group 10 -5: Output variables skipped -5: ------------------------------------------------ -5: 1Current vel. -5: 1Wind speed -5: 1Ice concentration -5: 2Wave height -5: 2Mean wave period(+2) -5: 2Mean wave period(+1) -5: 2Peak frequency -5: 2Mean wave dir. a1b1 -5: 2Peak direction -5: 4Part. wave height -5: 4Part. peak period -5: 4Part. mean direction -5: 5Charnock parameter -5: ------------------------------------------------ -5: OUTPUT TIME : 2021/03/26 00:00:00 UTC -5: -5: End of file reached -5: -5: -5: *** End of Grid interpolation Routine *** -5: =============================================== -5: -5: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -5: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -5: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_gnh_10m/out_grd.gnh_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.gnh_10m -5: + cmdfile.6[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032600 60 global 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + cmdfile.6[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib2_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grib2_sbs.sh[30]grdID=gnh_10m -5: + wave_grib2_sbs.sh[31]GRIDNR=255 -5: + wave_grib2_sbs.sh[32]MODNR=11 -5: + wave_grib2_sbs.sh[33]valid_time=2021032600 -5: + wave_grib2_sbs.sh[34]fhr=60 -5: + wave_grib2_sbs.sh[35]grid_region=global -5: + wave_grib2_sbs.sh[36]grid_res=0p16 -5: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463 -5: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_gnh_10m -5: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_gnh_10m -5: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_gnh_10m -5: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_gnh_10m -5: ++ wave_grib2_sbs.sh[47]printf %03i 60 -5: + wave_grib2_sbs.sh[47]FH3=060 -5: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p16 -5: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p16.f060.grib2 -5: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f060.grib2 ]] -5: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mod_def.gnh_10m ./mod_def.ww3 -5: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.gnh_10m ./out_grd.ww3 -5: + wave_grib2_sbs.sh[73]ngrib=1 -5: + wave_grib2_sbs.sh[74]dtgrib=3600 -5: + wave_grib2_sbs.sh[75]tstart='20210326 000000' -5: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210326 000000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gnh_10m.inp.tmpl -5: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -5: $ WAVEWATCH-III gridded output input file -5: $ ---------------------------------------- -5: 20210326 000000 3600 1 -5: N -5: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -5: $ -5: 20210326 000000 7 11 255 0 0 -5: $ -5: $ end of input file -5: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[88]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.457148 ']' -5: ++ prep_step[4]echo gfs_ww3_grib.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[90]export err=0 -5: + wave_grib2_sbs.sh[90]err=0 -5: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[95]cat grib2_global_060.out -5: -5: *** WAVEWATCH III GRIB output postp. *** -5: ============================================== -5: -5: Comment character is '$' -5: -5: Grid name : GFSv16-wave N Hemisphere 1/6 d -5: -5: LINEIN: -5: 20210326 000000 3600 1 -5: -5: 20210326000000 3600 1 -5: GEN_PRO -99999 -5: -5: Output time data : -5: ----------------------------------------------------- -5: First time : 2021/03/26 00:00:00 UTC -5: Interval : 01:00:00 -5: Number of requests : 1 -5: Fields : Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: Requested output fields not yet available: -5: ----------------------------------------------------- -5: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -5: -5: Successfully requested output fields : -5: ----------------------------------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: -5: Additional GRIB parameters : -5: ----------------------------------------------------- -5: Run time : 2021/03/26 00:00:00 UTC -5: GRIB center ID : 7 -5: GRIB gen. proc. ID : 11 -5: GRIB grid ID : 255 -5: GRIB GDS parameter : 0 -5: Fields in file : -5: -------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: CHOSEN GRID TYPE: : LLRECTILINEAR -5: -5: -5: -5: Generating file -5: ----------------------------------------------------- -5: Data for 2021/03/26 00:00:00 UTC 0H forecast. -5: -5: End of program -5: ========================================= -5: WAVEWATCH III GRIB output -5: -5: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -5: + wave_grib2_sbs.sh[102][[ 60 -gt 0 ]] -5: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '60 hour fcst' -grib gfs.wave.t12z.global.0p16.f060.grib2 -5: 1:0:d=2021032312:SPC:surface:60 hour fcst: -5: 2:236426:d=2021032312:DIRC:surface:60 hour fcst: -5: 3:864798:d=2021032312:UOGRD:surface:60 hour fcst: -5: 4:1096149:d=2021032312:VOGRD:surface:60 hour fcst: -5: 5:1334557:d=2021032312:WIND:surface:60 hour fcst: -5: 6:1736127:d=2021032312:WDIR:surface:60 hour fcst: -5: 7:2378508:d=2021032312:UGRD:surface:60 hour fcst: -5: 8:2777146:d=2021032312:VGRD:surface:60 hour fcst: -5: 9:3176204:d=2021032312:ICEC:surface:60 hour fcst: -5: 10:3290877:d=2021032312:HTSGW:surface:60 hour fcst: -5: 11:3574303:d=2021032312:IMWF:surface:60 hour fcst: -5: 12:3883373:d=2021032312:MWSPER:surface:60 hour fcst: -5: 13:4194481:d=2021032312:PERPW:surface:60 hour fcst: -5: 14:4526830:d=2021032312:WWSDIR:surface:60 hour fcst: -5: 15:5120285:d=2021032312:DIRPW:surface:60 hour fcst: -5: 16:5740660:d=2021032312:WVHGT:surface:60 hour fcst: -5: 17:6021301:d=2021032312:SWELL:1 in sequence:60 hour fcst: -5: 18:6275116:d=2021032312:SWELL:2 in sequence:60 hour fcst: -5: 19:6443122:d=2021032312:SWELL:3 in sequence:60 hour fcst: -5: 20:6561895:d=2021032312:WVPER:surface:60 hour fcst: -5: 21:6882414:d=2021032312:SWPER:1 in sequence:60 hour fcst: -5: 22:7171339:d=2021032312:SWPER:2 in sequence:60 hour fcst: -5: 23:7369759:d=2021032312:SWPER:3 in sequence:60 hour fcst: -5: 24:7495509:d=2021032312:WVDIR:surface:60 hour fcst: -5: 25:8020228:d=2021032312:SWDIR:1 in sequence:60 hour fcst: -5: 26:8526796:d=2021032312:SWDIR:2 in sequence:60 hour fcst: -5: 27:8824383:d=2021032312:SWDIR:3 in sequence:60 hour fcst: -5: + wave_grib2_sbs.sh[104]err=0 -5: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p16.f060.grib2 -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f060.grib2 ]] -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f060.grib2.idx ]] -5: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p16.f060.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f060.grib2 -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f060.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f060.grib2 = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f060.grib2 ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f060.grib2 -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f060.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f060.grib2.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f060.grib2.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f060.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f060.grib2 -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p16.f060.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f060.grib2.idx -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f060.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f060.grib2.idx = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f060.grib2.idx ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f060.grib2.idx -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f060.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f060.grib2.idx.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f060.grib2.idx.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f060.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f060.grib2.idx -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p16.f060.grib2 and gfs.wave.t12z.global.0p16.f060.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_gnh_10m to COM' -5: INFO: Copied gfs.wave.t12z.global.0p16.f060.grib2 and gfs.wave.t12z.global.0p16.f060.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_gnh_10m to COM -5: + wave_grib2_sbs.sh[130][[ gnh_10m == '' ]] -5: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -5: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p16.f060.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -5: INFO: gfs.wave.t12z.global.0p16.f060.grib2 is global.0p50 or SENDDBN is NO, no alert sent -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + cmdfile.4[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032600 3600. 9999 -3: + cmdfile.4[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grid_interp_sbs.sh[25]grdID=glo_30m -3: + wave_grid_interp_sbs.sh[26]valid_time=2021032600 -3: + wave_grid_interp_sbs.sh[27]dt=3600. -3: + wave_grid_interp_sbs.sh[28]nst=9999 -3: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463 -3: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.uglo_100km ./out_grd.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mod_def.uglo_100km ./mod_def.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mod_def.glo_30m ./mod_def.glo_30m -3: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ]] -3: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m'\''' -3: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m' -3: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[51]weights_found=1 -3: + wave_grid_interp_sbs.sh[59]ymdhms='20210326 000000' -3: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210326 000000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/glo_30m/g ww3_gint.inp.tmpl -3: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -3: $ Input file for interpolation of uglo_100km to glo_30m -3: $------------------------------------------------ -3: $ Start Time 3600. NSteps -3: 20210326 000000 3600. 9999 -3: $ Total number of grids -3: 2 -3: $ Grid extensions -3: 'uglo_100km' -3: 'glo_30m' -3: $ -3: 0 -3: $ -3: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[70]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.457148 ']' -3: ++ prep_step[4]echo gfs_ww3_gint.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -3: INFO: Executing 'gfs_ww3_gint.x' -3: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[73]cat grid_interp.glo_30m.out -3: -3: *** WAVEWATCH III Grid interpolation *** -3: =============================================== -3: -3: Comment character is '$' -3: -3: Time Information : -3: --------------------------------------------- -3: Starting Time : 2021/03/26 00:00:00 UTC -3: Interval (in sec) : 3600.00 -3: Number of requests : 9999 -3: --------------------------------------------- -3: Number of grids (including output grid) = 2 -3: -3: -3: Extension for grid 1 is --> uglo_100km -3: -3: Grid Particulars are : -3: Dimensions = 45166 1 -3: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Extension for grid 2 is --> glo_30m -3: -3: Grid Particulars are : -3: Dimensions = 720 336 -3: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -3: -3: -3: Preparing interpolation weights for output grid -3: Total number of wet points for interpolation 167619 -3: -3: -3: Variable: Grid Interpolation Map Units: 0.100E+01 -3: -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 336 | | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 291 | 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 | -3: 246 | 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 201 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 156 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 111 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 66 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 21 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 | -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: -3: -3: Interpolating fields .... -3: -3: Output group 1 -3: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -3: Output group 2 -3: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -3: Output group 3 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 4 -3: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -3: Output group 5 -3: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -3: Output group 6 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 7 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 8 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 9 -3: Output variables skipped -3: Output group 10 -3: Output variables skipped -3: ------------------------------------------------ -3: 1Current vel. -3: 1Wind speed -3: 1Ice concentration -3: 2Wave height -3: 2Mean wave period(+2) -3: 2Mean wave period(+1) -3: 2Peak frequency -3: 2Mean wave dir. a1b1 -3: 2Peak direction -3: 4Part. wave height -3: 4Part. peak period -3: 4Part. mean direction -3: 5Charnock parameter -3: ------------------------------------------------ -3: OUTPUT TIME : 2021/03/26 00:00:00 UTC -3: -3: End of file reached -3: -3: -3: *** End of Grid interpolation Routine *** -3: =============================================== -3: -3: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -3: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -3: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_glo_30m/out_grd.glo_30m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.glo_30m -3: + cmdfile.4[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032600 60 global 0p50 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + cmdfile.4[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib2_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grib2_sbs.sh[30]grdID=glo_30m -3: + wave_grib2_sbs.sh[31]GRIDNR=255 -3: + wave_grib2_sbs.sh[32]MODNR=11 -3: + wave_grib2_sbs.sh[33]valid_time=2021032600 -3: + wave_grib2_sbs.sh[34]fhr=60 -3: + wave_grib2_sbs.sh[35]grid_region=global -3: + wave_grib2_sbs.sh[36]grid_res=0p50 -3: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463 -3: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_glo_30m -3: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_glo_30m -3: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_glo_30m -3: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_glo_30m -3: ++ wave_grib2_sbs.sh[47]printf %03i 60 -3: + wave_grib2_sbs.sh[47]FH3=060 -3: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p50 -3: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p50.f060.grib2 -3: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f060.grib2 ]] -3: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mod_def.glo_30m ./mod_def.ww3 -3: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.glo_30m ./out_grd.ww3 -3: + wave_grib2_sbs.sh[73]ngrib=1 -3: + wave_grib2_sbs.sh[74]dtgrib=3600 -3: + wave_grib2_sbs.sh[75]tstart='20210326 000000' -3: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210326 000000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.glo_30m.inp.tmpl -3: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -3: $ WAVEWATCH-III gridded output input file -3: $ ---------------------------------------- -3: 20210326 000000 3600 1 -3: N -3: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -3: $ -3: 20210326 000000 7 11 255 0 0 -3: $ -3: $ end of input file -3: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[88]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.457148 ']' -3: ++ prep_step[4]echo gfs_ww3_grib.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[90]export err=0 -3: + wave_grib2_sbs.sh[90]err=0 -3: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[95]cat grib2_global_060.out -3: -3: *** WAVEWATCH III GRIB output postp. *** -3: ============================================== -3: -3: Comment character is '$' -3: -3: Grid name : Global 30 min wave grid -3: -3: LINEIN: -3: 20210326 000000 3600 1 -3: -3: 20210326000000 3600 1 -3: GEN_PRO -99999 -3: -3: Output time data : -3: ----------------------------------------------------- -3: First time : 2021/03/26 00:00:00 UTC -3: Interval : 01:00:00 -3: Number of requests : 1 -3: Fields : Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: Requested output fields not yet available: -3: ----------------------------------------------------- -3: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -3: -3: Successfully requested output fields : -3: ----------------------------------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: -3: Additional GRIB parameters : -3: ----------------------------------------------------- -3: Run time : 2021/03/26 00:00:00 UTC -3: GRIB center ID : 7 -3: GRIB gen. proc. ID : 11 -3: GRIB grid ID : 255 -3: GRIB GDS parameter : 0 -3: Fields in file : -3: -------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: CHOSEN GRID TYPE: : LLRECTILINEAR -3: -3: -3: -3: Generating file -3: ----------------------------------------------------- -3: Data for 2021/03/26 00:00:00 UTC 0H forecast. -3: -3: End of program -3: ========================================= -3: WAVEWATCH III GRIB output -3: -3: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -3: + wave_grib2_sbs.sh[102][[ 60 -gt 0 ]] -3: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '60 hour fcst' -grib gfs.wave.t12z.global.0p50.f060.grib2 -3: 1:0:d=2021032312:SPC:surface:60 hour fcst: -3: 2:77806:d=2021032312:DIRC:surface:60 hour fcst: -3: 3:319453:d=2021032312:UOGRD:surface:60 hour fcst: -3: 4:395065:d=2021032312:VOGRD:surface:60 hour fcst: -3: 5:474636:d=2021032312:WIND:surface:60 hour fcst: -3: 6:632502:d=2021032312:WDIR:surface:60 hour fcst: -3: 7:868009:d=2021032312:UGRD:surface:60 hour fcst: -3: 8:1023388:d=2021032312:VGRD:surface:60 hour fcst: -3: 9:1180437:d=2021032312:ICEC:surface:60 hour fcst: -3: 10:1220798:d=2021032312:HTSGW:surface:60 hour fcst: -3: 11:1319579:d=2021032312:IMWF:surface:60 hour fcst: -3: 12:1427852:d=2021032312:MWSPER:surface:60 hour fcst: -3: 13:1537573:d=2021032312:PERPW:surface:60 hour fcst: -3: 14:1656968:d=2021032312:WWSDIR:surface:60 hour fcst: -3: 15:1861261:d=2021032312:DIRPW:surface:60 hour fcst: -3: 16:2074127:d=2021032312:WVHGT:surface:60 hour fcst: -3: 17:2173547:d=2021032312:SWELL:1 in sequence:60 hour fcst: -3: 18:2267151:d=2021032312:SWELL:2 in sequence:60 hour fcst: -3: 19:2326853:d=2021032312:SWELL:3 in sequence:60 hour fcst: -3: 20:2363471:d=2021032312:WVPER:surface:60 hour fcst: -3: 21:2476480:d=2021032312:SWPER:1 in sequence:60 hour fcst: -3: 22:2582951:d=2021032312:SWPER:2 in sequence:60 hour fcst: -3: 23:2655876:d=2021032312:SWPER:3 in sequence:60 hour fcst: -3: 24:2696522:d=2021032312:WVDIR:surface:60 hour fcst: -3: 25:2872678:d=2021032312:SWDIR:1 in sequence:60 hour fcst: -3: 26:3053647:d=2021032312:SWDIR:2 in sequence:60 hour fcst: -3: 27:3165358:d=2021032312:SWDIR:3 in sequence:60 hour fcst: -3: + wave_grib2_sbs.sh[104]err=0 -3: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p50.f060.grib2 -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f060.grib2 ]] -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f060.grib2.idx ]] -3: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p50.f060.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f060.grib2 -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f060.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f060.grib2 = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f060.grib2 ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f060.grib2 -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f060.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f060.grib2.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f060.grib2.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f060.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f060.grib2 -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p50.f060.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f060.grib2.idx -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f060.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f060.grib2.idx = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f060.grib2.idx ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f060.grib2.idx -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f060.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f060.grib2.idx.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f060.grib2.idx.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f060.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f060.grib2.idx -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p50.f060.grib2 and gfs.wave.t12z.global.0p50.f060.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_glo_30m to COM' -3: INFO: Copied gfs.wave.t12z.global.0p50.f060.grib2 and gfs.wave.t12z.global.0p50.f060.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_glo_30m to COM -3: + wave_grib2_sbs.sh[130][[ glo_30m == '' ]] -3: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -3: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p50.f060.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -3: INFO: gfs.wave.t12z.global.0p50.f060.grib2 is global.0p50 or SENDDBN is NO, no alert sent -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + cmdfile.3[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032600 3600. 9999 -2: + cmdfile.3[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grid_interp_sbs.sh[25]grdID=wc_10m -2: + wave_grid_interp_sbs.sh[26]valid_time=2021032600 -2: + wave_grid_interp_sbs.sh[27]dt=3600. -2: + wave_grid_interp_sbs.sh[28]nst=9999 -2: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463 -2: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.uglo_100km ./out_grd.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mod_def.uglo_100km ./mod_def.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mod_def.wc_10m ./mod_def.wc_10m -2: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ]] -2: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m'\''' -2: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m' -2: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[51]weights_found=1 -2: + wave_grid_interp_sbs.sh[59]ymdhms='20210326 000000' -2: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210326 000000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/wc_10m/g ww3_gint.inp.tmpl -2: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -2: $ Input file for interpolation of uglo_100km to wc_10m -2: $------------------------------------------------ -2: $ Start Time 3600. NSteps -2: 20210326 000000 3600. 9999 -2: $ Total number of grids -2: 2 -2: $ Grid extensions -2: 'uglo_100km' -2: 'wc_10m' -2: $ -2: 0 -2: $ -2: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[70]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.457148 ']' -2: ++ prep_step[4]echo gfs_ww3_gint.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -2: INFO: Executing 'gfs_ww3_gint.x' -2: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[73]cat grid_interp.wc_10m.out -2: -2: *** WAVEWATCH III Grid interpolation *** -2: =============================================== -2: -2: Comment character is '$' -2: -2: Time Information : -2: --------------------------------------------- -2: Starting Time : 2021/03/26 00:00:00 UTC -2: Interval (in sec) : 3600.00 -2: Number of requests : 9999 -2: --------------------------------------------- -2: Number of grids (including output grid) = 2 -2: -2: -2: Extension for grid 1 is --> uglo_100km -2: -2: Grid Particulars are : -2: Dimensions = 45166 1 -2: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Extension for grid 2 is --> wc_10m -2: -2: Grid Particulars are : -2: Dimensions = 241 151 -2: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -2: -2: -2: Preparing interpolation weights for output grid -2: Total number of wet points for interpolation 11044 -2: -2: -2: Variable: Grid Interpolation Map Units: 0.100E+01 -2: -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: +---------------------------------------------------------------------------------------------------------------+ -2: 151 | 0 0 0 0 | -2: | 0 0 0 0 0 | -2: | 0 0 0 0 0 | -2: 130 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 109 | 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 88 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 | -2: 67 | 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 | -2: 46 | 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: 25 | 0 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 0 | -2: | | -2: 4 | | -2: +---------------------------------------------------------------------------------------------------------------+ -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: -2: -2: Interpolating fields .... -2: -2: Output group 1 -2: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -2: Output group 2 -2: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -2: Output group 3 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 4 -2: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -2: Output group 5 -2: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -2: Output group 6 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 7 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 8 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 9 -2: Output variables skipped -2: Output group 10 -2: Output variables skipped -2: ------------------------------------------------ -2: 1Current vel. -2: 1Wind speed -2: 1Ice concentration -2: 2Wave height -2: 2Mean wave period(+2) -2: 2Mean wave period(+1) -2: 2Peak frequency -2: 2Mean wave dir. a1b1 -2: 2Peak direction -2: 4Part. wave height -2: 4Part. peak period -2: 4Part. mean direction -2: 5Charnock parameter -2: ------------------------------------------------ -2: OUTPUT TIME : 2021/03/26 00:00:00 UTC -2: -2: End of file reached -2: -2: -2: *** End of Grid interpolation Routine *** -2: =============================================== -2: -2: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -2: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -2: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_wc_10m/out_grd.wc_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.wc_10m -2: + cmdfile.3[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032600 60 wcoast 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + cmdfile.3[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib2_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grib2_sbs.sh[30]grdID=wc_10m -2: + wave_grib2_sbs.sh[31]GRIDNR=255 -2: + wave_grib2_sbs.sh[32]MODNR=11 -2: + wave_grib2_sbs.sh[33]valid_time=2021032600 -2: + wave_grib2_sbs.sh[34]fhr=60 -2: + wave_grib2_sbs.sh[35]grid_region=wcoast -2: + wave_grib2_sbs.sh[36]grid_res=0p16 -2: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463 -2: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_wc_10m -2: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_wc_10m -2: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_wc_10m -2: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_wc_10m -2: ++ wave_grib2_sbs.sh[47]printf %03i 60 -2: + wave_grib2_sbs.sh[47]FH3=060 -2: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_wcoast_0p16 -2: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.wcoast.0p16.f060.grib2 -2: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f060.grib2 ]] -2: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mod_def.wc_10m ./mod_def.ww3 -2: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.wc_10m ./out_grd.ww3 -2: + wave_grib2_sbs.sh[73]ngrib=1 -2: + wave_grib2_sbs.sh[74]dtgrib=3600 -2: + wave_grib2_sbs.sh[75]tstart='20210326 000000' -2: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210326 000000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.wc_10m.inp.tmpl -2: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -2: $ WAVEWATCH-III gridded output input file -2: $ ---------------------------------------- -2: 20210326 000000 3600 1 -2: N -2: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -2: $ -2: 20210326 000000 7 11 255 0 0 -2: $ -2: $ end of input file -2: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[88]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.457148 ']' -2: ++ prep_step[4]echo gfs_ww3_grib.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[90]export err=0 -2: + wave_grib2_sbs.sh[90]err=0 -2: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[95]cat grib2_wcoast_060.out -2: -2: *** WAVEWATCH III GRIB output postp. *** -2: ============================================== -2: -2: Comment character is '$' -2: -2: Grid name : West Coast 10 min wave grid -2: -2: LINEIN: -2: 20210326 000000 3600 1 -2: -2: 20210326000000 3600 1 -2: GEN_PRO -99999 -2: -2: Output time data : -2: ----------------------------------------------------- -2: First time : 2021/03/26 00:00:00 UTC -2: Interval : 01:00:00 -2: Number of requests : 1 -2: Fields : Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: Requested output fields not yet available: -2: ----------------------------------------------------- -2: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -2: -2: Successfully requested output fields : -2: ----------------------------------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: -2: Additional GRIB parameters : -2: ----------------------------------------------------- -2: Run time : 2021/03/26 00:00:00 UTC -2: GRIB center ID : 7 -2: GRIB gen. proc. ID : 11 -2: GRIB grid ID : 255 -2: GRIB GDS parameter : 0 -2: Fields in file : -2: -------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: CHOSEN GRID TYPE: : LLRECTILINEAR -2: -2: -2: -2: Generating file -2: ----------------------------------------------------- -2: Data for 2021/03/26 00:00:00 UTC 0H forecast. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: -2: End of program -2: ========================================= -2: WAVEWATCH III GRIB output -2: -2: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -2: + wave_grib2_sbs.sh[102][[ 60 -gt 0 ]] -2: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '60 hour fcst' -grib gfs.wave.t12z.wcoast.0p16.f060.grib2 -2: 1:0:d=2021032312:SPC:surface:60 hour fcst: -2: 2:7355:d=2021032312:DIRC:surface:60 hour fcst: -2: 3:23974:d=2021032312:UOGRD:surface:60 hour fcst: -2: 4:30991:d=2021032312:VOGRD:surface:60 hour fcst: -2: 5:38578:d=2021032312:WIND:surface:60 hour fcst: -2: 6:49710:d=2021032312:WDIR:surface:60 hour fcst: -2: 7:65543:d=2021032312:UGRD:surface:60 hour fcst: -2: 8:76858:d=2021032312:VGRD:surface:60 hour fcst: -2: 9:88277:d=2021032312:ICEC:surface:60 hour fcst: -2: 10:93007:d=2021032312:HTSGW:surface:60 hour fcst: -2: 11:102521:d=2021032312:IMWF:surface:60 hour fcst: -2: 12:112103:d=2021032312:MWSPER:surface:60 hour fcst: -2: 13:121797:d=2021032312:PERPW:surface:60 hour fcst: -2: 14:131763:d=2021032312:WWSDIR:surface:60 hour fcst: -2: 15:146666:d=2021032312:DIRPW:surface:60 hour fcst: -2: 16:161822:d=2021032312:WVHGT:surface:60 hour fcst: -2: 17:171514:d=2021032312:SWELL:1 in sequence:60 hour fcst: -2: 18:176926:d=2021032312:SWELL:2 in sequence:60 hour fcst: -2: 19:181656:d=2021032312:SWELL:3 in sequence:60 hour fcst: -2: 20:186386:d=2021032312:WVPER:surface:60 hour fcst: -2: 21:196488:d=2021032312:SWPER:1 in sequence:60 hour fcst: -2: 22:201822:d=2021032312:SWPER:2 in sequence:60 hour fcst: -2: 23:206552:d=2021032312:SWPER:3 in sequence:60 hour fcst: -2: 24:211282:d=2021032312:WVDIR:surface:60 hour fcst: -2: 25:226341:d=2021032312:SWDIR:1 in sequence:60 hour fcst: -2: 26:232201:d=2021032312:SWDIR:2 in sequence:60 hour fcst: -2: 27:236931:d=2021032312:SWDIR:3 in sequence:60 hour fcst: -2: + wave_grib2_sbs.sh[104]err=0 -2: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.wcoast.0p16.f060.grib2 -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f060.grib2 ]] -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f060.grib2.idx ]] -2: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.wcoast.0p16.f060.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f060.grib2 -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f060.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f060.grib2 = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f060.grib2 ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f060.grib2 -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f060.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f060.grib2.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f060.grib2.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f060.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f060.grib2 -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.wcoast.0p16.f060.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f060.grib2.idx -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f060.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f060.grib2.idx = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f060.grib2.idx ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f060.grib2.idx -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f060.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f060.grib2.idx.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f060.grib2.idx.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f060.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f060.grib2.idx -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.wcoast.0p16.f060.grib2 and gfs.wave.t12z.wcoast.0p16.f060.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_wc_10m to COM' -2: INFO: Copied gfs.wave.t12z.wcoast.0p16.f060.grib2 and gfs.wave.t12z.wcoast.0p16.f060.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_wc_10m to COM -2: + wave_grib2_sbs.sh[130][[ wc_10m == '' ]] -2: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -2: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.wcoast.0p16.f060.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -2: INFO: gfs.wave.t12z.wcoast.0p16.f060.grib2 is global.0p50 or SENDDBN is NO, no alert sent -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + cmdfile.1[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032600 3600. 9999 -0: + cmdfile.1[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grid_interp_sbs.sh[25]grdID=at_10m -0: + wave_grid_interp_sbs.sh[26]valid_time=2021032600 -0: + wave_grid_interp_sbs.sh[27]dt=3600. -0: + wave_grid_interp_sbs.sh[28]nst=9999 -0: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463 -0: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.uglo_100km ./out_grd.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mod_def.uglo_100km ./mod_def.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mod_def.at_10m ./mod_def.at_10m -0: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ]] -0: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m'\''' -0: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m' -0: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[51]weights_found=1 -0: + wave_grid_interp_sbs.sh[59]ymdhms='20210326 000000' -0: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210326 000000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/at_10m/g ww3_gint.inp.tmpl -0: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -0: $ Input file for interpolation of uglo_100km to at_10m -0: $------------------------------------------------ -0: $ Start Time 3600. NSteps -0: 20210326 000000 3600. 9999 -0: $ Total number of grids -0: 2 -0: $ Grid extensions -0: 'uglo_100km' -0: 'at_10m' -0: $ -0: 0 -0: $ -0: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[70]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.457148 ']' -0: ++ prep_step[4]echo gfs_ww3_gint.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -0: INFO: Executing 'gfs_ww3_gint.x' -0: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[73]cat grid_interp.at_10m.out -0: -0: *** WAVEWATCH III Grid interpolation *** -0: =============================================== -0: -0: Comment character is '$' -0: -0: Time Information : -0: --------------------------------------------- -0: Starting Time : 2021/03/26 00:00:00 UTC -0: Interval (in sec) : 3600.00 -0: Number of requests : 9999 -0: --------------------------------------------- -0: Number of grids (including output grid) = 2 -0: -0: -0: Extension for grid 1 is --> uglo_100km -0: -0: Grid Particulars are : -0: Dimensions = 45166 1 -0: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Extension for grid 2 is --> at_10m -0: -0: Grid Particulars are : -0: Dimensions = 301 331 -0: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -0: -0: -0: Preparing interpolation weights for output grid -0: Total number of wet points for interpolation 29591 -0: -0: -0: Variable: Grid Interpolation Map Units: 0.100E+01 -0: -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 331 | | -0: | | -0: | | -0: 289 | | -0: | | -0: | 0 0 0 | -0: 247 | 0 0 0 0 | -0: | 0 0 0 0 0 | -0: | 0 0 0 0 | -0: 205 | 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 163 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 121 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 79 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 | -0: | 0 0 | -0: 37 | 0 | -0: | | -0: | | -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: -0: -0: Interpolating fields .... -0: -0: Output group 1 -0: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -0: Output group 2 -0: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -0: Output group 3 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 4 -0: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -0: Output group 5 -0: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -0: Output group 6 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 7 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 8 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 9 -0: Output variables skipped -0: Output group 10 -0: Output variables skipped -0: ------------------------------------------------ -0: 1Current vel. -0: 1Wind speed -0: 1Ice concentration -0: 2Wave height -0: 2Mean wave period(+2) -0: 2Mean wave period(+1) -0: 2Peak frequency -0: 2Mean wave dir. a1b1 -0: 2Peak direction -0: 4Part. wave height -0: 4Part. peak period -0: 4Part. mean direction -0: 5Charnock parameter -0: ------------------------------------------------ -0: OUTPUT TIME : 2021/03/26 00:00:00 UTC -0: -0: End of file reached -0: -0: -0: *** End of Grid interpolation Routine *** -0: =============================================== -0: -0: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -0: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -0: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_at_10m/out_grd.at_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.at_10m -0: + cmdfile.1[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032600 60 atlocn 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + cmdfile.1[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib2_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grib2_sbs.sh[30]grdID=at_10m -0: + wave_grib2_sbs.sh[31]GRIDNR=255 -0: + wave_grib2_sbs.sh[32]MODNR=11 -0: + wave_grib2_sbs.sh[33]valid_time=2021032600 -0: + wave_grib2_sbs.sh[34]fhr=60 -0: + wave_grib2_sbs.sh[35]grid_region=atlocn -0: + wave_grib2_sbs.sh[36]grid_res=0p16 -0: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463 -0: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_at_10m -0: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_at_10m -0: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_at_10m -0: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_at_10m -0: ++ wave_grib2_sbs.sh[47]printf %03i 60 -0: + wave_grib2_sbs.sh[47]FH3=060 -0: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_atlocn_0p16 -0: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.atlocn.0p16.f060.grib2 -0: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f060.grib2 ]] -0: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mod_def.at_10m ./mod_def.ww3 -0: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.at_10m ./out_grd.ww3 -0: + wave_grib2_sbs.sh[73]ngrib=1 -0: + wave_grib2_sbs.sh[74]dtgrib=3600 -0: + wave_grib2_sbs.sh[75]tstart='20210326 000000' -0: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210326 000000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.at_10m.inp.tmpl -0: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -0: $ WAVEWATCH-III gridded output input file -0: $ ---------------------------------------- -0: 20210326 000000 3600 1 -0: N -0: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -0: $ -0: 20210326 000000 7 11 255 0 0 -0: $ -0: $ end of input file -0: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[88]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.457148 ']' -0: ++ prep_step[4]echo gfs_ww3_grib.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[90]export err=0 -0: + wave_grib2_sbs.sh[90]err=0 -0: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[95]cat grib2_atlocn_060.out -0: -0: *** WAVEWATCH III GRIB output postp. *** -0: ============================================== -0: -0: Comment character is '$' -0: -0: Grid name : NW Atlantic 10 min wave grid -0: -0: LINEIN: -0: 20210326 000000 3600 1 -0: -0: 20210326000000 3600 1 -0: GEN_PRO -99999 -0: -0: Output time data : -0: ----------------------------------------------------- -0: First time : 2021/03/26 00:00:00 UTC -0: Interval : 01:00:00 -0: Number of requests : 1 -0: Fields : Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: Requested output fields not yet available: -0: ----------------------------------------------------- -0: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -0: -0: Successfully requested output fields : -0: ----------------------------------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: -0: Additional GRIB parameters : -0: ----------------------------------------------------- -0: Run time : 2021/03/26 00:00:00 UTC -0: GRIB center ID : 7 -0: GRIB gen. proc. ID : 11 -0: GRIB grid ID : 255 -0: GRIB GDS parameter : 0 -0: Fields in file : -0: -------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: CHOSEN GRID TYPE: : LLRECTILINEAR -0: -0: -0: -0: Generating file -0: ----------------------------------------------------- -0: Data for 2021/03/26 00:00:00 UTC 0H forecast. -0: -0: End of program -0: ========================================= -0: WAVEWATCH III GRIB output -0: -0: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -0: + wave_grib2_sbs.sh[102][[ 60 -gt 0 ]] -0: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '60 hour fcst' -grib gfs.wave.t12z.atlocn.0p16.f060.grib2 -0: 1:0:d=2021032312:SPC:surface:60 hour fcst: -0: 2:20355:d=2021032312:DIRC:surface:60 hour fcst: -0: 3:60505:d=2021032312:UOGRD:surface:60 hour fcst: -0: 4:80426:d=2021032312:VOGRD:surface:60 hour fcst: -0: 5:100557:d=2021032312:WIND:surface:60 hour fcst: -0: 6:131263:d=2021032312:WDIR:surface:60 hour fcst: -0: 7:173010:d=2021032312:UGRD:surface:60 hour fcst: -0: 8:203415:d=2021032312:VGRD:surface:60 hour fcst: -0: 9:233703:d=2021032312:ICEC:surface:60 hour fcst: -0: 10:246338:d=2021032312:HTSGW:surface:60 hour fcst: -0: 11:271111:d=2021032312:IMWF:surface:60 hour fcst: -0: 12:297552:d=2021032312:MWSPER:surface:60 hour fcst: -0: 13:324148:d=2021032312:PERPW:surface:60 hour fcst: -0: 14:352253:d=2021032312:WWSDIR:surface:60 hour fcst: -0: 15:392321:d=2021032312:DIRPW:surface:60 hour fcst: -0: 16:434124:d=2021032312:WVHGT:surface:60 hour fcst: -0: 17:458648:d=2021032312:SWELL:1 in sequence:60 hour fcst: -0: 18:479353:d=2021032312:SWELL:2 in sequence:60 hour fcst: -0: 19:495444:d=2021032312:SWELL:3 in sequence:60 hour fcst: -0: 20:508936:d=2021032312:WVPER:surface:60 hour fcst: -0: 21:536242:d=2021032312:SWPER:1 in sequence:60 hour fcst: -0: 22:558673:d=2021032312:SWPER:2 in sequence:60 hour fcst: -0: 23:576571:d=2021032312:SWPER:3 in sequence:60 hour fcst: -0: 24:590524:d=2021032312:WVDIR:surface:60 hour fcst: -0: 25:628933:d=2021032312:SWDIR:1 in sequence:60 hour fcst: -0: 26:661159:d=2021032312:SWDIR:2 in sequence:60 hour fcst: -0: 27:684139:d=2021032312:SWDIR:3 in sequence:60 hour fcst: -0: + wave_grib2_sbs.sh[104]err=0 -0: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.atlocn.0p16.f060.grib2 -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f060.grib2 ]] -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f060.grib2.idx ]] -0: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.atlocn.0p16.f060.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f060.grib2 -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f060.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f060.grib2 = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f060.grib2 ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f060.grib2 -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f060.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f060.grib2.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f060.grib2.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f060.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f060.grib2 -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.atlocn.0p16.f060.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f060.grib2.idx -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f060.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f060.grib2.idx = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f060.grib2.idx ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f060.grib2.idx -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f060.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f060.grib2.idx.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f060.grib2.idx.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f060.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f060.grib2.idx -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.atlocn.0p16.f060.grib2 and gfs.wave.t12z.atlocn.0p16.f060.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_at_10m to COM' -0: INFO: Copied gfs.wave.t12z.atlocn.0p16.f060.grib2 and gfs.wave.t12z.atlocn.0p16.f060.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_at_10m to COM -0: + wave_grib2_sbs.sh[130][[ at_10m == '' ]] -0: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -0: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.atlocn.0p16.f060.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -0: INFO: gfs.wave.t12z.atlocn.0p16.f060.grib2 is global.0p50 or SENDDBN is NO, no alert sent -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + cmdfile.5[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032600 3600. 9999 -4: + cmdfile.5[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grid_interp_sbs.sh[25]grdID=aoc_9km -4: + wave_grid_interp_sbs.sh[26]valid_time=2021032600 -4: + wave_grid_interp_sbs.sh[27]dt=3600. -4: + wave_grid_interp_sbs.sh[28]nst=9999 -4: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463 -4: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.uglo_100km ./out_grd.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mod_def.uglo_100km ./mod_def.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mod_def.aoc_9km ./mod_def.aoc_9km -4: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km'\''' -4: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km' -4: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[51]weights_found=1 -4: + wave_grid_interp_sbs.sh[59]ymdhms='20210326 000000' -4: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210326 000000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/aoc_9km/g ww3_gint.inp.tmpl -4: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -4: $ Input file for interpolation of uglo_100km to aoc_9km -4: $------------------------------------------------ -4: $ Start Time 3600. NSteps -4: 20210326 000000 3600. 9999 -4: $ Total number of grids -4: 2 -4: $ Grid extensions -4: 'uglo_100km' -4: 'aoc_9km' -4: $ -4: 0 -4: $ -4: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[70]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.457148 ']' -4: ++ prep_step[4]echo gfs_ww3_gint.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -4: INFO: Executing 'gfs_ww3_gint.x' -4: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[73]cat grid_interp.aoc_9km.out -4: -4: *** WAVEWATCH III Grid interpolation *** -4: =============================================== -4: -4: Comment character is '$' -4: -4: Time Information : -4: --------------------------------------------- -4: Starting Time : 2021/03/26 00:00:00 UTC -4: Interval (in sec) : 3600.00 -4: Number of requests : 9999 -4: --------------------------------------------- -4: Number of grids (including output grid) = 2 -4: -4: -4: Extension for grid 1 is --> uglo_100km -4: -4: Grid Particulars are : -4: Dimensions = 45166 1 -4: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Extension for grid 2 is --> aoc_9km -4: -4: Grid Particulars are : -4: Dimensions = 1006 1006 -4: Grid Type = 2 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -4: -4: -4: Preparing interpolation weights for output grid -4: Total number of wet points for interpolation 360052 -4: -4: -4: Variable: Grid Interpolation Map Units: 0.100E+01 -4: -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: *** | | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: 880 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 | -4: 754 | 0 0 0 | -4: | 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 628 | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 | -4: 502 | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 376 | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 250 | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 124 | 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 | -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: -4: -4: Interpolating fields .... -4: -4: Output group 1 -4: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -4: Output group 2 -4: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -4: Output group 3 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 4 -4: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -4: Output group 5 -4: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -4: Output group 6 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 7 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 8 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 9 -4: Output variables skipped -4: Output group 10 -4: Output variables skipped -4: ------------------------------------------------ -4: 1Current vel. -4: 1Wind speed -4: 1Ice concentration -4: 2Wave height -4: 2Mean wave period(+2) -4: 2Mean wave period(+1) -4: 2Peak frequency -4: 2Mean wave dir. a1b1 -4: 2Peak direction -4: 4Part. wave height -4: 4Part. peak period -4: 4Part. mean direction -4: 5Charnock parameter -4: ------------------------------------------------ -4: OUTPUT TIME : 2021/03/26 00:00:00 UTC -4: -4: End of file reached -4: -4: -4: *** End of Grid interpolation Routine *** -4: =============================================== -4: -4: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -4: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -4: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_aoc_9km/out_grd.aoc_9km /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.aoc_9km -4: + cmdfile.5[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032600 60 arctic 9km 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + cmdfile.5[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib2_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grib2_sbs.sh[30]grdID=aoc_9km -4: + wave_grib2_sbs.sh[31]GRIDNR=255 -4: + wave_grib2_sbs.sh[32]MODNR=11 -4: + wave_grib2_sbs.sh[33]valid_time=2021032600 -4: + wave_grib2_sbs.sh[34]fhr=60 -4: + wave_grib2_sbs.sh[35]grid_region=arctic -4: + wave_grib2_sbs.sh[36]grid_res=9km -4: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463 -4: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_aoc_9km -4: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_aoc_9km -4: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_aoc_9km -4: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_aoc_9km -4: ++ wave_grib2_sbs.sh[47]printf %03i 60 -4: + wave_grib2_sbs.sh[47]FH3=060 -4: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_arctic_9km -4: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.arctic.9km.f060.grib2 -4: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f060.grib2 ]] -4: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mod_def.aoc_9km ./mod_def.ww3 -4: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.aoc_9km ./out_grd.ww3 -4: + wave_grib2_sbs.sh[73]ngrib=1 -4: + wave_grib2_sbs.sh[74]dtgrib=3600 -4: + wave_grib2_sbs.sh[75]tstart='20210326 000000' -4: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210326 000000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.aoc_9km.inp.tmpl -4: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -4: $ WAVEWATCH-III gridded output input file -4: $ ---------------------------------------- -4: 20210326 000000 3600 1 -4: N -4: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -4: $ -4: 20210326 000000 7 11 255 0 20 -4: $ -4: 70 0 9.0 9.0 64 -4: $ 60 0 8.64919046313 8.64919046313 64 -4: $ end of input file -4: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[88]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.457148 ']' -4: ++ prep_step[4]echo gfs_ww3_grib.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[90]export err=0 -4: + wave_grib2_sbs.sh[90]err=0 -4: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[95]cat grib2_arctic_060.out -4: -4: *** WAVEWATCH III GRIB output postp. *** -4: ============================================== -4: -4: Comment character is '$' -4: -4: Grid name : Arctic Ocean PolarStereo 9km -4: -4: LINEIN: -4: 20210326 000000 3600 1 -4: -4: 20210326000000 3600 1 -4: GEN_PRO -99999 -4: -4: Output time data : -4: ----------------------------------------------------- -4: First time : 2021/03/26 00:00:00 UTC -4: Interval : 01:00:00 -4: Number of requests : 1 -4: Fields : Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: Requested output fields not yet available: -4: ----------------------------------------------------- -4: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -4: -4: Successfully requested output fields : -4: ----------------------------------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: -4: Additional GRIB parameters : -4: ----------------------------------------------------- -4: Run time : 2021/03/26 00:00:00 UTC -4: GRIB center ID : 7 -4: GRIB gen. proc. ID : 11 -4: GRIB grid ID : 255 -4: GRIB GDS parameter : 0 -4: Fields in file : -4: -------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: CHOSEN GRID TYPE: : POLARSTEREO -4: -4: -4: -4: Generating file -4: ----------------------------------------------------- -4: Data for 2021/03/26 00:00:00 UTC 0H forecast. -4: -4: End of program -4: ========================================= -4: WAVEWATCH III GRIB output -4: -4: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -4: + wave_grib2_sbs.sh[102][[ 60 -gt 0 ]] -4: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '60 hour fcst' -grib gfs.wave.t12z.arctic.9km.f060.grib2 -4: 1:0:d=2021032312:SPC:surface:60 hour fcst: -4: 2:192321:d=2021032312:DIRC:surface:60 hour fcst: -4: 3:659753:d=2021032312:UOGRD:surface:60 hour fcst: -4: 4:849316:d=2021032312:VOGRD:surface:60 hour fcst: -4: 5:1043332:d=2021032312:WIND:surface:60 hour fcst: -4: 6:1343851:d=2021032312:WDIR:surface:60 hour fcst: -4: 7:1772176:d=2021032312:UGRD:surface:60 hour fcst: -4: 8:2069460:d=2021032312:VGRD:surface:60 hour fcst: -4: 9:2369361:d=2021032312:ICEC:surface:60 hour fcst: -4: 10:2539807:d=2021032312:HTSGW:surface:60 hour fcst: -4: 11:2750142:d=2021032312:IMWF:surface:60 hour fcst: -4: 12:2964710:d=2021032312:MWSPER:surface:60 hour fcst: -4: 13:3181190:d=2021032312:PERPW:surface:60 hour fcst: -4: 14:3407265:d=2021032312:WWSDIR:surface:60 hour fcst: -4: 15:3708312:d=2021032312:DIRPW:surface:60 hour fcst: -4: 16:4016128:d=2021032312:WVHGT:surface:60 hour fcst: -4: 17:4223015:d=2021032312:SWELL:1 in sequence:60 hour fcst: -4: 18:4413860:d=2021032312:SWELL:2 in sequence:60 hour fcst: -4: 19:4564583:d=2021032312:SWELL:3 in sequence:60 hour fcst: -4: 20:4692301:d=2021032312:WVPER:surface:60 hour fcst: -4: 21:4909633:d=2021032312:SWPER:1 in sequence:60 hour fcst: -4: 22:5109863:d=2021032312:SWPER:2 in sequence:60 hour fcst: -4: 23:5268140:d=2021032312:SWPER:3 in sequence:60 hour fcst: -4: 24:5396316:d=2021032312:WVDIR:surface:60 hour fcst: -4: 25:5676511:d=2021032312:SWDIR:1 in sequence:60 hour fcst: -4: 26:5945620:d=2021032312:SWDIR:2 in sequence:60 hour fcst: -4: 27:6133073:d=2021032312:SWDIR:3 in sequence:60 hour fcst: -4: + wave_grib2_sbs.sh[104]err=0 -4: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.arctic.9km.f060.grib2 -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f060.grib2 ]] -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f060.grib2.idx ]] -4: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.arctic.9km.f060.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f060.grib2 -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f060.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f060.grib2 = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f060.grib2 ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f060.grib2 -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f060.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f060.grib2.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f060.grib2.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f060.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f060.grib2 -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.arctic.9km.f060.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f060.grib2.idx -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f060.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f060.grib2.idx = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f060.grib2.idx ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f060.grib2.idx -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f060.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f060.grib2.idx.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f060.grib2.idx.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f060.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f060.grib2.idx -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.arctic.9km.f060.grib2 and gfs.wave.t12z.arctic.9km.f060.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_aoc_9km to COM' -4: INFO: Copied gfs.wave.t12z.arctic.9km.f060.grib2 and gfs.wave.t12z.arctic.9km.f060.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_aoc_9km to COM -4: + wave_grib2_sbs.sh[130][[ aoc_9km == '' ]] -4: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -4: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.arctic.9km.f060.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -4: INFO: gfs.wave.t12z.arctic.9km.f060.grib2 is global.0p50 or SENDDBN is NO, no alert sent -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + cmdfile.7[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032600 3600. 9999 -6: + cmdfile.7[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grid_interp_sbs.sh[25]grdID=gsh_15m -6: + wave_grid_interp_sbs.sh[26]valid_time=2021032600 -6: + wave_grid_interp_sbs.sh[27]dt=3600. -6: + wave_grid_interp_sbs.sh[28]nst=9999 -6: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463 -6: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.uglo_100km ./out_grd.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mod_def.uglo_100km ./mod_def.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mod_def.gsh_15m ./mod_def.gsh_15m -6: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m'\''' -6: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m' -6: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[51]weights_found=1 -6: + wave_grid_interp_sbs.sh[59]ymdhms='20210326 000000' -6: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210326 000000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gsh_15m/g ww3_gint.inp.tmpl -6: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -6: $ Input file for interpolation of uglo_100km to gsh_15m -6: $------------------------------------------------ -6: $ Start Time 3600. NSteps -6: 20210326 000000 3600. 9999 -6: $ Total number of grids -6: 2 -6: $ Grid extensions -6: 'uglo_100km' -6: 'gsh_15m' -6: $ -6: 0 -6: $ -6: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[70]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.457148 ']' -6: ++ prep_step[4]echo gfs_ww3_gint.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]env -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: ++ prep_step[12]unset FORT01 -6: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -6: INFO: Executing 'gfs_ww3_gint.x' -6: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[73]cat grid_interp.gsh_15m.out -6: -6: *** WAVEWATCH III Grid interpolation *** -6: =============================================== -6: -6: Comment character is '$' -6: -6: Time Information : -6: --------------------------------------------- -6: Starting Time : 2021/03/26 00:00:00 UTC -6: Interval (in sec) : 3600.00 -6: Number of requests : 9999 -6: --------------------------------------------- -6: Number of grids (including output grid) = 2 -6: -6: -6: Extension for grid 1 is --> uglo_100km -6: -6: Grid Particulars are : -6: Dimensions = 45166 1 -6: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Extension for grid 2 is --> gsh_15m -6: -6: Grid Particulars are : -6: Dimensions = 1440 277 -6: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -6: -6: -6: Preparing interpolation weights for output grid -6: Total number of wet points for interpolation 317192 -6: -6: -6: Variable: Grid Interpolation Map Units: 0.100E+01 -6: -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 277 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 241 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 205 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 169 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 133 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 97 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 61 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 | -6: 25 | 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 | -6: | | -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: -6: -6: Interpolating fields .... -6: -6: Output group 1 -6: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -6: Output group 2 -6: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -6: Output group 3 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 4 -6: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -6: Output group 5 -6: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -6: Output group 6 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 7 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 8 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 9 -6: Output variables skipped -6: Output group 10 -6: Output variables skipped -6: ------------------------------------------------ -6: 1Current vel. -6: 1Wind speed -6: 1Ice concentration -6: 2Wave height -6: 2Mean wave period(+2) -6: 2Mean wave period(+1) -6: 2Peak frequency -6: 2Mean wave dir. a1b1 -6: 2Peak direction -6: 4Part. wave height -6: 4Part. peak period -6: 4Part. mean direction -6: 5Charnock parameter -6: ------------------------------------------------ -6: OUTPUT TIME : 2021/03/26 00:00:00 UTC -6: -6: End of file reached -6: -6: -6: *** End of Grid interpolation Routine *** -6: =============================================== -6: -6: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -6: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -6: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grid_interp_gsh_15m/out_grd.gsh_15m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.gsh_15m -6: + cmdfile.7[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032600 60 gsouth 0p25 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + cmdfile.7[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib2_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grib2_sbs.sh[30]grdID=gsh_15m -6: + wave_grib2_sbs.sh[31]GRIDNR=255 -6: + wave_grib2_sbs.sh[32]MODNR=11 -6: + wave_grib2_sbs.sh[33]valid_time=2021032600 -6: + wave_grib2_sbs.sh[34]fhr=60 -6: + wave_grib2_sbs.sh[35]grid_region=gsouth -6: + wave_grib2_sbs.sh[36]grid_res=0p25 -6: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463 -6: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_gsh_15m -6: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_gsh_15m -6: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_gsh_15m -6: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_gsh_15m -6: ++ wave_grib2_sbs.sh[47]printf %03i 60 -6: + wave_grib2_sbs.sh[47]FH3=060 -6: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -6: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.gsouth.0p25.f060.grib2 -6: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f060.grib2 ]] -6: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/mod_def.gsh_15m ./mod_def.ww3 -6: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/out_grd.gsh_15m ./out_grd.ww3 -6: + wave_grib2_sbs.sh[73]ngrib=1 -6: + wave_grib2_sbs.sh[74]dtgrib=3600 -6: + wave_grib2_sbs.sh[75]tstart='20210326 000000' -6: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210326 000000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gsh_15m.inp.tmpl -6: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -6: $ WAVEWATCH-III gridded output input file -6: $ ---------------------------------------- -6: 20210326 000000 3600 1 -6: N -6: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -6: $ -6: 20210326 000000 7 11 255 0 0 -6: $ -6: $ end of input file -6: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[88]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.457148 ']' -6: ++ prep_step[4]echo gfs_ww3_grib.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[90]export err=0 -6: + wave_grib2_sbs.sh[90]err=0 -6: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[95]cat grib2_gsouth_060.out -6: -6: *** WAVEWATCH III GRIB output postp. *** -6: ============================================== -6: -6: Comment character is '$' -6: -6: Grid name : GFSv16-wave S Hemisphere 1/4 d -6: -6: LINEIN: -6: 20210326 000000 3600 1 -6: -6: 20210326000000 3600 1 -6: GEN_PRO -99999 -6: -6: Output time data : -6: ----------------------------------------------------- -6: First time : 2021/03/26 00:00:00 UTC -6: Interval : 01:00:00 -6: Number of requests : 1 -6: Fields : Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: Requested output fields not yet available: -6: ----------------------------------------------------- -6: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -6: -6: Successfully requested output fields : -6: ----------------------------------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: -6: Additional GRIB parameters : -6: ----------------------------------------------------- -6: Run time : 2021/03/26 00:00:00 UTC -6: GRIB center ID : 7 -6: GRIB gen. proc. ID : 11 -6: GRIB grid ID : 255 -6: GRIB GDS parameter : 0 -6: Fields in file : -6: -------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: CHOSEN GRID TYPE: : LLRECTILINEAR -6: -6: -6: -6: Generating file -6: ----------------------------------------------------- -6: Data for 2021/03/26 00:00:00 UTC 0H forecast. -6: -6: End of program -6: ========================================= -6: WAVEWATCH III GRIB output -6: -6: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -6: + wave_grib2_sbs.sh[102][[ 60 -gt 0 ]] -6: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '60 hour fcst' -grib gfs.wave.t12z.gsouth.0p25.f060.grib2 -6: 1:0:d=2021032312:SPC:surface:60 hour fcst: -6: 2:117412:d=2021032312:DIRC:surface:60 hour fcst: -6: 3:499313:d=2021032312:UOGRD:surface:60 hour fcst: -6: 4:612749:d=2021032312:VOGRD:surface:60 hour fcst: -6: 5:734175:d=2021032312:WIND:surface:60 hour fcst: -6: 6:979158:d=2021032312:WDIR:surface:60 hour fcst: -6: 7:1357275:d=2021032312:UGRD:surface:60 hour fcst: -6: 8:1596353:d=2021032312:VGRD:surface:60 hour fcst: -6: 9:1838768:d=2021032312:ICEC:surface:60 hour fcst: -6: 10:1899917:d=2021032312:HTSGW:surface:60 hour fcst: -6: 11:2057600:d=2021032312:IMWF:surface:60 hour fcst: -6: 12:2227317:d=2021032312:MWSPER:surface:60 hour fcst: -6: 13:2398103:d=2021032312:PERPW:surface:60 hour fcst: -6: 14:2588135:d=2021032312:WWSDIR:surface:60 hour fcst: -6: 15:2929993:d=2021032312:DIRPW:surface:60 hour fcst: -6: 16:3288271:d=2021032312:WVHGT:surface:60 hour fcst: -6: 17:3452324:d=2021032312:SWELL:1 in sequence:60 hour fcst: -6: 18:3615057:d=2021032312:SWELL:2 in sequence:60 hour fcst: -6: 19:3723089:d=2021032312:SWELL:3 in sequence:60 hour fcst: -6: 20:3787338:d=2021032312:WVPER:surface:60 hour fcst: -6: 21:3972625:d=2021032312:SWPER:1 in sequence:60 hour fcst: -6: 22:4156762:d=2021032312:SWPER:2 in sequence:60 hour fcst: -6: 23:4290702:d=2021032312:SWPER:3 in sequence:60 hour fcst: -6: 24:4364686:d=2021032312:WVDIR:surface:60 hour fcst: -6: 25:4662847:d=2021032312:SWDIR:1 in sequence:60 hour fcst: -6: 26:4994683:d=2021032312:SWDIR:2 in sequence:60 hour fcst: -6: 27:5217398:d=2021032312:SWDIR:3 in sequence:60 hour fcst: -6: + wave_grib2_sbs.sh[104]err=0 -6: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.gsouth.0p25.f060.grib2 -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f060.grib2 ]] -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f060.grib2.idx ]] -6: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.gsouth.0p25.f060.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f060.grib2 -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f060.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f060.grib2 = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f060.grib2 ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f060.grib2 -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f060.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f060.grib2.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f060.grib2.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f060.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f060.grib2 -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.gsouth.0p25.f060.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f060.grib2.idx -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f060.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f060.grib2.idx = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f060.grib2.idx ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f060.grib2.idx -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f060.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f060.grib2.idx.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f060.grib2.idx.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f060.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f060.grib2.idx -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.gsouth.0p25.f060.grib2 and gfs.wave.t12z.gsouth.0p25.f060.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_gsh_15m to COM' -6: INFO: Copied gfs.wave.t12z.gsouth.0p25.f060.grib2 and gfs.wave.t12z.gsouth.0p25.f060.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463/grib_gsh_15m to COM -6: + wave_grib2_sbs.sh[130][[ gsh_15m == '' ]] -6: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -6: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.gsouth.0p25.f060.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -6: INFO: gfs.wave.t12z.gsouth.0p25.f060.grib2 is global.0p50 or SENDDBN is NO, no alert sent -+ run_mpmd.sh[113]exit 0 -+ run_mpmd.sh[1]postamble run_mpmd.sh 1753758301 0 -+ preamble.sh[62]set +x -End run_mpmd.sh at 03:05:12 with error code 0 (time elapsed: 00:00:11) -+ exgfs_wave_post_gridded_sbs.sh[122]true -+ exgfs_wave_post_gridded_sbs.sh[123]export err=0 -+ exgfs_wave_post_gridded_sbs.sh[123]err=0 -+ exgfs_wave_post_gridded_sbs.sh[124][[ 0 -ne 0 ]] -+ exgfs_wave_post_gridded_sbs.sh[130]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ exgfs_wave_post_gridded_sbs.sh[131]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ exgfs_wave_post_gridded_sbs.sh[132]gribchk=gfs.wave.t12z.gsouth.0p25.f060.grib2 -+ exgfs_wave_post_gridded_sbs.sh[133][[ ! -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f060.grib2 ]] -+ exgfs_wave_post_gridded_sbs.sh[138]exit 0 -+ JGLOBAL_WAVE_POST_SBS[28]true -+ JGLOBAL_WAVE_POST_SBS[29]export err=0 -+ JGLOBAL_WAVE_POST_SBS[29]err=0 -+ JGLOBAL_WAVE_POST_SBS[30][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[37]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312 -+ JGLOBAL_WAVE_POST_SBS[38][[ NO != \Y\E\S ]] -+ JGLOBAL_WAVE_POST_SBS[39]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f060.453463 -+ JGLOBAL_WAVE_POST_SBS[42]exit 0 -+ JGLOBAL_WAVE_POST_SBS[1]postamble /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS 1753758289 0 -+ preamble.sh[62]set +x -End /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at 03:05:12 with error code 0 (time elapsed: 00:00:23) -Begin /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at Tue Jul 29 03:05:12 UTC 2025 -++ jjob_header.sh[46]OPTIND=1 -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[50]env_job=wavepostsbs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[49]read -ra configs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[61]shift 4 -++ jjob_header.sh[63][[ -z wavepostsbs ]] -++ jjob_header.sh[71]export DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463 -++ jjob_header.sh[71]DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463 -++ jjob_header.sh[72][[ YES == \Y\E\S ]] -++ jjob_header.sh[73]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463 -++ jjob_header.sh[75]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463 -++ jjob_header.sh[76]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463 -++ jjob_header.sh[85]export pid=460139 -++ jjob_header.sh[85]pid=460139 -++ jjob_header.sh[86]export pgmout=OUTPUT.460139 -++ jjob_header.sh[86]pgmout=OUTPUT.460139 -++ jjob_header.sh[87]export pgmerr=errfile -++ jjob_header.sh[87]pgmerr=errfile -++ jjob_header.sh[90]export pgm= -++ jjob_header.sh[90]pgm= -++ jjob_header.sh[96]export cycle=t12z -++ jjob_header.sh[96]cycle=t12z -++ jjob_header.sh[97]setpdy.sh -+ setpdy.sh[20]'[' /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463 == /home/mterry ']' -+ setpdy.sh[25][[ ! t12z =~ t??z ]] -+ setpdy.sh[30]case $# in -+ setpdy.sh[31]dates_before_PDY=7 -+ setpdy.sh[32]dates_after_PDY=7 -+ setpdy.sh[50]COMDATEROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+ setpdy.sh[53]'[' -z 20210323 ']' -+ setpdy.sh[57]sed 's/[0-9]\{8\}/20210323/' /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z -sed: can't read /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z: No such file or directory -++ jjob_header.sh[97]true -++ jjob_header.sh[98]source ./PDY -/work2/noaa/global/mterry/global-workflow_forked/ush/jjob_header.sh: line 98: ./PDY: No such file or directory -++ jjob_header.sh[98]true -++ jjob_header.sh[104]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[104]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.base -+++ config.base[6]echo 'BEGIN: config.base' -BEGIN: config.base -+++ config.base[9]export machine=HERCULES -+++ config.base[9]machine=HERCULES -+++ config.base[12]export RUN_ENVIR=emc -+++ config.base[12]RUN_ENVIR=emc -+++ config.base[15]export ACCOUNT=fv3-cpu -+++ config.base[15]ACCOUNT=fv3-cpu -+++ config.base[16]export QUEUE=batch -+++ config.base[16]QUEUE=batch -+++ config.base[17]export QUEUE_SERVICE=batch -+++ config.base[17]QUEUE_SERVICE=batch -+++ config.base[18]export QUEUE_DTN=batch -+++ config.base[18]QUEUE_DTN=batch -+++ config.base[19]export PARTITION_BATCH=hercules -+++ config.base[19]PARTITION_BATCH=hercules -+++ config.base[20]export PARTITION_SERVICE=service -+++ config.base[20]PARTITION_SERVICE=service -+++ config.base[21]export PARTITION_DTN= -+++ config.base[21]PARTITION_DTN= -+++ config.base[22]export RESERVATION= -+++ config.base[22]RESERVATION= -+++ config.base[23]export CLUSTERS= -+++ config.base[23]CLUSTERS= -+++ config.base[24]export CLUSTERS_SERVICE= -+++ config.base[24]CLUSTERS_SERVICE= -+++ config.base[25]export CLUSTERS_DTN= -+++ config.base[25]CLUSTERS_DTN= -+++ config.base[28]export HPSS_PROJECT=emc-global -+++ config.base[28]HPSS_PROJECT=emc-global -+++ config.base[31]export HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[31]HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[32]export EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[32]EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[33]export FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[33]FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[34]export PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[34]PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[35]export SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[35]SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[36]export USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[36]USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[38]export FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[38]FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[39]export FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[39]FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[40]export FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[40]FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[41]export FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[41]FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[42]export FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[42]FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[43]export FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[43]FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[44]export FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[44]FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[45]export FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[45]FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[50]export PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[50]PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[51]export COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[51]COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[52]export COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[52]COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[53]export DMPDIR=/work/noaa/rstprod/dump -+++ config.base[53]DMPDIR=/work/noaa/rstprod/dump -+++ config.base[57]export COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[57]COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[58]export COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[58]COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[59]export COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[59]COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[62]export HOMEDIR=/work2/noaa/global/mterry -+++ config.base[62]HOMEDIR=/work2/noaa/global/mterry -+++ config.base[63]export STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[63]STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]export PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[65]export NOSCRUB=/work2/noaa/global/mterry -+++ config.base[65]NOSCRUB=/work2/noaa/global/mterry -+++ config.base[68]export BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[68]BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[71]export BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[71]BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[74]export DO_PREP_SFC=NO -+++ config.base[74]DO_PREP_SFC=NO -+++ config.base[77]export DO_GOES=NO -+++ config.base[77]DO_GOES=NO -+++ config.base[78]export DO_BUFRSND=NO -+++ config.base[78]DO_BUFRSND=NO -+++ config.base[79]export DO_GEMPAK=NO -+++ config.base[79]DO_GEMPAK=NO -+++ config.base[80]export DO_AWIPS=NO -+++ config.base[80]DO_AWIPS=NO -+++ config.base[81]export DO_NPOESS=NO -+++ config.base[81]DO_NPOESS=NO -+++ config.base[82]export DO_TRACKER=YES -+++ config.base[82]DO_TRACKER=YES -+++ config.base[83]export DO_GENESIS=YES -+++ config.base[83]DO_GENESIS=YES -+++ config.base[84]export DO_GENESIS_FSU=NO -+++ config.base[84]DO_GENESIS_FSU=NO -+++ config.base[85]export DO_VERFOZN=YES -+++ config.base[85]DO_VERFOZN=YES -+++ config.base[86]export DO_VERFRAD=YES -+++ config.base[86]DO_VERFRAD=YES -+++ config.base[87]export DO_VMINMON=YES -+++ config.base[87]DO_VMINMON=YES -+++ config.base[88]export DO_ANLSTAT=NO -+++ config.base[88]DO_ANLSTAT=NO -+++ config.base[91]export MODE=forecast-only -+++ config.base[91]MODE=forecast-only -+++ config.base[92]export DO_TEST_MODE=YES -+++ config.base[92]DO_TEST_MODE=YES -+++ config.base[101]export FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[101]FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[102]export HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[102]HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[103]export HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[103]HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[106]export NMV=/bin/mv -+++ config.base[106]NMV=/bin/mv -+++ config.base[107]export 'NLN=/bin/ln -sf' -+++ config.base[107]NLN='/bin/ln -sf' -+++ config.base[108]export VERBOSE=YES -+++ config.base[108]VERBOSE=YES -+++ config.base[109]export KEEPDATA=NO -+++ config.base[109]KEEPDATA=NO -+++ config.base[110]export DEBUG_POSTSCRIPT=NO -+++ config.base[110]DEBUG_POSTSCRIPT=NO -+++ config.base[111]export CHGRP_RSTPROD=YES -+++ config.base[111]CHGRP_RSTPROD=YES -+++ config.base[112]export 'CHGRP_CMD=chgrp rstprod' -+++ config.base[112]CHGRP_CMD='chgrp rstprod' -+++ config.base[113]export NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[113]NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[114]export NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[114]NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[117]export BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[117]BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[120]export SDATE=2021032312 -+++ config.base[120]SDATE=2021032312 -+++ config.base[121]export EDATE=2021032312 -+++ config.base[121]EDATE=2021032312 -+++ config.base[122]export EXP_WARM_START=.false. -+++ config.base[122]EXP_WARM_START=.false. -+++ config.base[123]export assim_freq=6 -+++ config.base[123]assim_freq=6 -+++ config.base[124]export PSLOT=C48_S2SW -+++ config.base[124]PSLOT=C48_S2SW -+++ config.base[125]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[125]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[126]export ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[126]ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[127]export DUMP_SUFFIX= -+++ config.base[127]DUMP_SUFFIX= -+++ config.base[128][[ 2021032312 -ge 2019092100 ]] -+++ config.base[128][[ 2021032312 -le 2019110700 ]] -+++ config.base[131]export ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[131]ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[132]export ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[132]ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[133]export FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[133]FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[136]export envir=prod -+++ config.base[136]envir=prod -+++ config.base[137]export NET=gfs -+++ config.base[137]NET=gfs -+++ config.base[138]export RUN=gfs -+++ config.base[138]RUN=gfs -+++ config.base[141]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.com -++++ config.com[4]echo 'BEGIN: config.com' -BEGIN: config.com -++++ config.com[38][[ emc == \n\c\o ]] -++++ config.com[43]COM_OBSPROC_TMPL='${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos' -++++ config.com[44]COM_RTOFS_TMPL='${DMPDIR}' -++++ config.com[45]COM_TCVITAL_TMPL='${DMPDIR}/${RUN}.${YMD}/${HH}/atmos' -++++ config.com[47]declare -rx 'COM_OBS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/obs' -++++ config.com[48]declare -rx COM_OBSPROC_TMPL COM_RTOFS_TMPL -++++ config.com[50]COM_BASE='${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}' -++++ config.com[52]declare -rx 'COM_TOP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}' -++++ config.com[54]declare -rx 'COM_CONF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf' -++++ config.com[55]declare -rx 'COM_OBS_JEDI=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi' -++++ config.com[57]declare -rx 'COM_ATMOS_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input' -++++ config.com[58]declare -rx 'COM_ATMOS_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart' -++++ config.com[59]declare -rx 'COM_ATMOS_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos' -++++ config.com[60]declare -rx 'COM_SNOW_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow' -++++ config.com[61]declare -rx 'COM_SNOW_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon' -++++ config.com[62]declare -rx 'COM_ATMOS_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history' -++++ config.com[63]declare -rx 'COM_ATMOS_MASTER_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master' -++++ config.com[64]declare -rx 'COM_ATMOS_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2' -++++ config.com[65]declare -rx 'COM_ATMOS_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}' -++++ config.com[66]declare -rx 'COM_ATMOS_BUFR_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr' -++++ config.com[67]declare -rx 'COM_ATMOS_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}' -++++ config.com[68]declare -rx 'COM_ATMOS_GENESIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital' -++++ config.com[69]declare -rx 'COM_ATMOS_TRACK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks' -++++ config.com[70]declare -rx 'COM_ATMOS_GOES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim' -++++ config.com[71]declare -rx 'COM_ATMOS_IMAGERY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery' -++++ config.com[72]declare -rx 'COM_ATMOS_OZNMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon' -++++ config.com[73]declare -rx 'COM_ATMOS_RADMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon' -++++ config.com[74]declare -rx 'COM_ATMOS_MINMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon' -++++ config.com[75]declare -rx 'COM_ATMOS_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon' -++++ config.com[76]declare -rx 'COM_ATMOS_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo' -++++ config.com[78]declare -rx 'COM_WAVE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart' -++++ config.com[79]declare -rx 'COM_WAVE_PREP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep' -++++ config.com[80]declare -rx 'COM_WAVE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history' -++++ config.com[81]declare -rx 'COM_WAVE_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded' -++++ config.com[82]declare -rx 'COM_WAVE_GRID_RES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}' -++++ config.com[83]declare -rx 'COM_WAVE_STATION_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station' -++++ config.com[84]declare -rx 'COM_WAVE_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak' -++++ config.com[85]declare -rx 'COM_WAVE_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo' -++++ config.com[87]declare -rx 'COM_OCEAN_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history' -++++ config.com[88]declare -rx 'COM_OCEAN_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart' -++++ config.com[89]declare -rx 'COM_OCEAN_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input' -++++ config.com[90]declare -rx 'COM_OCEAN_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean' -++++ config.com[91]declare -rx 'COM_OCEAN_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon' -++++ config.com[92]declare -rx 'COM_OCEAN_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf' -++++ config.com[93]declare -rx 'COM_OCEAN_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean' -++++ config.com[94]declare -rx 'COM_OCEAN_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf' -++++ config.com[95]declare -rx 'COM_OCEAN_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2' -++++ config.com[96]declare -rx 'COM_OCEAN_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}' -++++ config.com[98]declare -rx 'COM_ICE_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice' -++++ config.com[99]declare -rx 'COM_ICE_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf' -++++ config.com[100]declare -rx 'COM_ICE_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon' -++++ config.com[101]declare -rx 'COM_ICE_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice' -++++ config.com[102]declare -rx 'COM_ICE_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input' -++++ config.com[103]declare -rx 'COM_ICE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history' -++++ config.com[104]declare -rx 'COM_ICE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart' -++++ config.com[105]declare -rx 'COM_ICE_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf' -++++ config.com[106]declare -rx 'COM_ICE_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2' -++++ config.com[107]declare -rx 'COM_ICE_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}' -++++ config.com[109]declare -rx 'COM_CHEM_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history' -++++ config.com[110]declare -rx 'COM_CHEM_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem' -++++ config.com[111]declare -rx 'COM_CHEM_BMAT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix' -++++ config.com[112]declare -rx 'COM_CHEM_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon' -++++ config.com[114]declare -rx 'COM_MED_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart' -+++ config.base[143]export LOGSCRIPT= -+++ config.base[143]LOGSCRIPT= -+++ config.base[145]export 'REDOUT=1>' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wave -+++ config.wave[6]echo 'BEGIN: config.wave' -BEGIN: config.wave -+++ config.wave[13]export RUNRSTwave=gdas -+++ config.wave[13]RUNRSTwave=gdas -+++ config.wave[16]export MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[16]MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[19]case "${waveGRD}" in -+++ config.wave[64]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[64]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[65]export wavepostGRD= -+++ config.wave[65]wavepostGRD= -+++ config.wave[66]export waveuoutpGRD=uglo_100km -+++ config.wave[66]waveuoutpGRD=uglo_100km -+++ config.wave[75]export WAVEWND_DID= -+++ config.wave[75]WAVEWND_DID= -+++ config.wave[76]export WAVEWND_FID= -+++ config.wave[76]WAVEWND_FID= -+++ config.wave[79][[ gfs == \g\f\s ]] -+++ config.wave[80]export FHMAX_WAV=120 -+++ config.wave[80]FHMAX_WAV=120 -+++ config.wave[82]export WAVHINDH=0 -+++ config.wave[82]WAVHINDH=0 -+++ config.wave[83]export FHMAX_WAV_IBP=180 -+++ config.wave[83]FHMAX_WAV_IBP=180 -+++ config.wave[84](( FHMAX_WAV < FHMAX_WAV_IBP )) -+++ config.wave[84]export FHMAX_WAV_IBP=120 -+++ config.wave[84]FHMAX_WAV_IBP=120 -+++ config.wave[87]export DTFLD_WAV=3600 -+++ config.wave[87]DTFLD_WAV=3600 -+++ config.wave[88]export DTPNT_WAV=3600 -+++ config.wave[88]DTPNT_WAV=3600 -+++ config.wave[89]export FHINCP_WAV=1 -+++ config.wave[89]FHINCP_WAV=1 -+++ config.wave[92]export 'OUTPARS_WAV=WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[92]OUTPARS_WAV='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[95][[ gfs == \g\d\a\s ]] -+++ config.wave[99](( INTERVAL_GFS > 0 )) -+++ config.wave[100]export WAVHCYC=6 -+++ config.wave[100]WAVHCYC=6 -+++ config.wave[101]export FHMAX_WAV_CUR=192 -+++ config.wave[101]FHMAX_WAV_CUR=192 -+++ config.wave[109]export RSTTYPE_WAV=T -+++ config.wave[109]RSTTYPE_WAV=T -+++ config.wave[110][[ gfs != gfs ]] -+++ config.wave[115]rst_dt_gfs=43200 -+++ config.wave[116][[ 43200 -gt 0 ]] -+++ config.wave[117]export DT_1_RST_WAV=0 -+++ config.wave[117]DT_1_RST_WAV=0 -+++ config.wave[120]export DT_2_RST_WAV=43200 -+++ config.wave[120]DT_2_RST_WAV=43200 -+++ config.wave[126]export RSTIOFF_WAV=0 -+++ config.wave[126]RSTIOFF_WAV=0 -+++ config.wave[131]export RUNMEM=-1 -+++ config.wave[131]RUNMEM=-1 -+++ config.wave[134](( RUNMEM == -1 )) -+++ config.wave[136]export waveMEMB= -+++ config.wave[136]waveMEMB= -+++ config.wave[143]export WW3ATMINP=CPL -+++ config.wave[143]WW3ATMINP=CPL -+++ config.wave[144][[ YES == \Y\E\S ]] -+++ config.wave[145]export WW3ICEINP=CPL -+++ config.wave[145]WW3ICEINP=CPL -+++ config.wave[146]export WAVEICE_FID= -+++ config.wave[146]WAVEICE_FID= -+++ config.wave[152][[ YES == \Y\E\S ]] -+++ config.wave[153]export WW3CURINP=CPL -+++ config.wave[153]WW3CURINP=CPL -+++ config.wave[154]export WAVECUR_FID= -+++ config.wave[154]WAVECUR_FID= -+++ config.wave[161]export WW3ATMIENS=F -+++ config.wave[161]WW3ATMIENS=F -+++ config.wave[162]export WW3ICEIENS=F -+++ config.wave[162]WW3ICEIENS=F -+++ config.wave[163]export WW3CURIENS=F -+++ config.wave[163]WW3CURIENS=F -+++ config.wave[165]export GOFILETYPE=1 -+++ config.wave[165]GOFILETYPE=1 -+++ config.wave[166]export POFILETYPE=1 -+++ config.wave[166]POFILETYPE=1 -+++ config.wave[170]export FUNIPNT=T -+++ config.wave[170]FUNIPNT=T -+++ config.wave[172]export IOSRV=1 -+++ config.wave[172]IOSRV=1 -+++ config.wave[174]export FPNTPROC=T -+++ config.wave[174]FPNTPROC=T -+++ config.wave[176]export FGRDPROC=F -+++ config.wave[176]FGRDPROC=F -+++ config.wave[178]export FLAGMASKCOMP=F -+++ config.wave[178]FLAGMASKCOMP=F -+++ config.wave[180]export FLAGMASKOUT=F -+++ config.wave[180]FLAGMASKOUT=F -+++ config.wave[182]echo 'END: config.wave' -END: config.wave -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wavepostsbs -+++ config.wavepostsbs[6]echo 'BEGIN: config.wavepostsbs' -BEGIN: config.wavepostsbs -+++ config.wavepostsbs[9]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources wavepostsbs -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=wavepostsbs -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[156]ntasks=8 -++++ config.resources[157]threads_per_task=1 -++++ config.resources[158]tasks_per_node=80 -++++ config.resources[159]NTASKS=8 -++++ config.resources[160]memory=20GB -++++ config.resources[161]walltime=00:15:00 -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export NTASKS -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export memory -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.wavepostsbs[12]export MAX_TASKS=25 -+++ config.wavepostsbs[12]MAX_TASKS=25 -+++ config.wavepostsbs[15]export WAV_SUBGRBSRC= -+++ config.wavepostsbs[15]WAV_SUBGRBSRC= -+++ config.wavepostsbs[16]export WAV_SUBGRB= -+++ config.wavepostsbs[16]WAV_SUBGRB= -+++ config.wavepostsbs[19]export DOFLD_WAV=YES -+++ config.wavepostsbs[19]DOFLD_WAV=YES -+++ config.wavepostsbs[20]export DOPNT_WAV=YES -+++ config.wavepostsbs[20]DOPNT_WAV=YES -+++ config.wavepostsbs[21]export DOGRB_WAV=YES -+++ config.wavepostsbs[21]DOGRB_WAV=YES -+++ config.wavepostsbs[22][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+++ config.wavepostsbs[23]export DOGRI_WAV=YES -+++ config.wavepostsbs[23]DOGRI_WAV=YES -+++ config.wavepostsbs[27]export DOSPC_WAV=YES -+++ config.wavepostsbs[27]DOSPC_WAV=YES -+++ config.wavepostsbs[28]export DOBLL_WAV=YES -+++ config.wavepostsbs[28]DOBLL_WAV=YES -+++ config.wavepostsbs[30]echo 'END: config.wavepostsbs' -END: config.wavepostsbs -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env wavepostsbs -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=wavepostsbs -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 20971520 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 8 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[34]max_threads_per_task=1 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 1 ]] -+++ HERCULES.env[40][[ 1 -gt 1 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 8' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[63]export USE_CFP=YES -+++ HERCULES.env[63]USE_CFP=YES -+++ HERCULES.env[64][[ wavepostsbs == \w\a\v\e\p\r\e\p ]] -+++ HERCULES.env[65]export 'wavempexec=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[65]wavempexec='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[66]export 'wave_mpmd=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[66]wave_mpmd='--multi-prog --output=mpmd.%j.%t.out' -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[4]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ JGLOBAL_WAVE_POST_SBS[7]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[7]HH=12 -+ JGLOBAL_WAVE_POST_SBS[7]declare_from_tmpl -rx COMIN_WAVE_PREP:COM_WAVE_PREP_TMPL COMIN_WAVE_HISTORY:COM_WAVE_HISTORY_TMPL COMOUT_WAVE_GRID:COM_WAVE_GRID_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_WAVE_PREP=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep -declare_from_tmpl :: COMIN_WAVE_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history -declare_from_tmpl :: COMOUT_WAVE_GRID=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded -+ JGLOBAL_WAVE_POST_SBS[12][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n '' ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_atlocn_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_atlocn_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_atlocn_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_epacif_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_epacif_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_epacif_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_wcoast_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_wcoast_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_wcoast_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p50 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p50:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p50=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p50 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=arctic.9km -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_arctic_9km:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_arctic_9km=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_arctic_9km -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_gsouth_0p25:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_gsouth_0p25=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 ]] -+ JGLOBAL_WAVE_POST_SBS[28]/work2/noaa/global/mterry/global-workflow_forked/scripts/exgfs_wave_post_gridded_sbs.sh -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ exgfs_wave_post_gridded_sbs.sh[24]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ exgfs_wave_post_gridded_sbs.sh[26]DOGRI_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[27]DOGRB_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[29]export waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[29]waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[30]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[30]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[31]export wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[31]wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[33]cat - INFO: Grid information: - INFO: Native wave grids: uglo_100km - INFO: Interpolated grids: at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m - INFO: Post-process grids: -++ exgfs_wave_post_gridded_sbs.sh[40]printf %03i 63 -+ exgfs_wave_post_gridded_sbs.sh[40]fhr3=063 -++ exgfs_wave_post_gridded_sbs.sh[41]date -u -d '20210323 12 + 63 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[41]valid_time=2021032603 -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[49]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f063.bin ./out_grd.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f063.bin ./out_grd.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[52][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[62]cat - INFO: Summary: - INFO: Grid interp: DOGRI_WAV="YES" - INFO: Grib files: DOGRB_WAV="YES" - INFO: Fields to be included in grib files: - INFO: OUTPARS_WAV="WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA" -+ exgfs_wave_post_gridded_sbs.sh[70][[ YES == \N\O ]] -+ exgfs_wave_post_gridded_sbs.sh[76]rm -f 'cmdfile.*' cmdfile -+ exgfs_wave_post_gridded_sbs.sh[77]count=0 -+ exgfs_wave_post_gridded_sbs.sh[80][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[81]dt_int=3600. -+ exgfs_wave_post_gridded_sbs.sh[82]n_int=9999 -++ exgfs_wave_post_gridded_sbs.sh[83]date -u -d '20210326 03 - 0 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[83]ymdh_int=2021032603 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=1 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032603 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032603 63 atlocn 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib2_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib2_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=2 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032603 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032603 63 epacif 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib2_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib2_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=3 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032603 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032603 63 wcoast 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib2_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib2_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=4 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032603 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032603 63 global 0p50 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib2_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib2_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=5 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032603 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032603 63 arctic 9km '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib2_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib2_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=6 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032603 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032603 63 global 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib2_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib2_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=7 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032603 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032603 63 gsouth 0p25 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib2_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib2_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[100][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[113][[ 8 -lt 7 ]] -+ exgfs_wave_post_gridded_sbs.sh[121]echo 'INFO: Running MPMD job with 7 commands' -INFO: Running MPMD job with 7 commands -+ exgfs_wave_post_gridded_sbs.sh[122]/work2/noaa/global/mterry/global-workflow_forked/ush/run_mpmd.sh /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/cmdfile -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ run_mpmd.sh[31]source /work2/noaa/global/mterry/global-workflow_forked/ush/preamble.sh -++ preamble.sh[20]set +x -Begin run_mpmd.sh at Tue Jul 29 03:05:24 UTC 2025 -+ run_mpmd.sh[33]cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/cmdfile -+ run_mpmd.sh[36][[ YES != \Y\E\S ]] -+ run_mpmd.sh[46]export OMP_NUM_THREADS=1 -+ run_mpmd.sh[46]OMP_NUM_THREADS=1 -++ run_mpmd.sh[49]wc -l -+ run_mpmd.sh[49]nprocs=7 -+ run_mpmd.sh[52]mpmd_cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mpmd_cmdfile -+ run_mpmd.sh[53][[ -s /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mpmd_cmdfile ]] -+ run_mpmd.sh[55]cat - INFO: Executing MPMD job, STDOUT redirected for each process separately - INFO: On failure, logs for each job will be available in /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mpmd.proc_num.out - INFO: The proc_num corresponds to the line in '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mpmd_cmdfile' -+ run_mpmd.sh[61][[ srun -l --export=ALL --hint=nomultithread =~ ^srun.* ]] -+ run_mpmd.sh[65]nm=0 -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '0 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/cmdfile.1' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '1 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/cmdfile.2' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '2 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/cmdfile.3' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '3 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/cmdfile.4' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '4 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/cmdfile.5' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '5 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/cmdfile.6' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '6 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/cmdfile.7' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[72]set +e -+ run_mpmd.sh[74]srun -l --export=ALL --hint=nomultithread --multi-prog --output=mpmd.%j.%t.out -n 7 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mpmd_cmdfile -+ run_mpmd.sh[75]err=0 -+ run_mpmd.sh[76]set_strict -+ preamble.sh[35][[ YES == \Y\E\S ]] -+ preamble.sh[37]set -eu -+ run_mpmd.sh[103][[ 0 -eq 0 ]] -+ run_mpmd.sh[104]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mpmd_cmdfile -++ run_mpmd.sh[105]find . -name 'mpmd.*.out' -+ run_mpmd.sh[105]out_files='./mpmd.5951794.1.out -./mpmd.5951794.5.out -./mpmd.5951794.3.out -./mpmd.5951794.2.out -./mpmd.5951794.0.out -./mpmd.5951794.4.out -./mpmd.5951794.6.out' -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951794.1.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951794.1.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951794.5.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951794.5.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951794.3.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951794.3.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951794.2.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951794.2.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951794.0.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951794.0.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951794.4.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951794.4.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951794.6.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951794.6.out -+ run_mpmd.sh[110]cat mpmd.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + cmdfile.2[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032603 3600. 9999 -1: + cmdfile.2[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grid_interp_sbs.sh[25]grdID=ep_10m -1: + wave_grid_interp_sbs.sh[26]valid_time=2021032603 -1: + wave_grid_interp_sbs.sh[27]dt=3600. -1: + wave_grid_interp_sbs.sh[28]nst=9999 -1: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463 -1: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.uglo_100km ./out_grd.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mod_def.uglo_100km ./mod_def.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mod_def.ep_10m ./mod_def.ep_10m -1: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ]] -1: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m'\''' -1: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m' -1: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[51]weights_found=1 -1: + wave_grid_interp_sbs.sh[59]ymdhms='20210326 030000' -1: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210326 030000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/ep_10m/g ww3_gint.inp.tmpl -1: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -1: $ Input file for interpolation of uglo_100km to ep_10m -1: $------------------------------------------------ -1: $ Start Time 3600. NSteps -1: 20210326 030000 3600. 9999 -1: $ Total number of grids -1: 2 -1: $ Grid extensions -1: 'uglo_100km' -1: 'ep_10m' -1: $ -1: 0 -1: $ -1: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[70]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.460139 ']' -1: ++ prep_step[4]echo gfs_ww3_gint.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -1: INFO: Executing 'gfs_ww3_gint.x' -1: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[73]cat grid_interp.ep_10m.out -1: -1: *** WAVEWATCH III Grid interpolation *** -1: =============================================== -1: -1: Comment character is '$' -1: -1: Time Information : -1: --------------------------------------------- -1: Starting Time : 2021/03/26 03:00:00 UTC -1: Interval (in sec) : 3600.00 -1: Number of requests : 9999 -1: --------------------------------------------- -1: Number of grids (including output grid) = 2 -1: -1: -1: Extension for grid 1 is --> uglo_100km -1: -1: Grid Particulars are : -1: Dimensions = 45166 1 -1: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Extension for grid 2 is --> ep_10m -1: -1: Grid Particulars are : -1: Dimensions = 511 301 -1: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -1: -1: -1: Preparing interpolation weights for output grid -1: Total number of wet points for interpolation 7439 -1: -1: -1: Variable: Grid Interpolation Map Units: 0.100E+01 -1: -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 301 | | -1: | | -1: | 0 0 | -1: 262 | 0 0 0 0 | -1: | 0 0 0 0 | -1: | 0 0 0 0 | -1: 223 | 0 0 0 | -1: | 0 | -1: | 0 | -1: 184 | | -1: | 0 0 0 | -1: | 0 0 0 0 | -1: 145 | 0 | -1: | | -1: | | -1: 106 | | -1: | | -1: | | -1: 67 | | -1: | 0 | -1: | | -1: 28 | 0 | -1: | | -1: | | -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: -1: -1: Interpolating fields .... -1: -1: Output group 1 -1: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -1: Output group 2 -1: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -1: Output group 3 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 4 -1: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -1: Output group 5 -1: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -1: Output group 6 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 7 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 8 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 9 -1: Output variables skipped -1: Output group 10 -1: Output variables skipped -1: ------------------------------------------------ -1: 1Current vel. -1: 1Wind speed -1: 1Ice concentration -1: 2Wave height -1: 2Mean wave period(+2) -1: 2Mean wave period(+1) -1: 2Peak frequency -1: 2Mean wave dir. a1b1 -1: 2Peak direction -1: 4Part. wave height -1: 4Part. peak period -1: 4Part. mean direction -1: 5Charnock parameter -1: ------------------------------------------------ -1: OUTPUT TIME : 2021/03/26 03:00:00 UTC -1: -1: End of file reached -1: -1: -1: *** End of Grid interpolation Routine *** -1: =============================================== -1: -1: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -1: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -1: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_ep_10m/out_grd.ep_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.ep_10m -1: + cmdfile.2[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032603 63 epacif 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + cmdfile.2[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib2_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grib2_sbs.sh[30]grdID=ep_10m -1: + wave_grib2_sbs.sh[31]GRIDNR=255 -1: + wave_grib2_sbs.sh[32]MODNR=11 -1: + wave_grib2_sbs.sh[33]valid_time=2021032603 -1: + wave_grib2_sbs.sh[34]fhr=63 -1: + wave_grib2_sbs.sh[35]grid_region=epacif -1: + wave_grib2_sbs.sh[36]grid_res=0p16 -1: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463 -1: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_ep_10m -1: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_ep_10m -1: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_ep_10m -1: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_ep_10m -1: ++ wave_grib2_sbs.sh[47]printf %03i 63 -1: + wave_grib2_sbs.sh[47]FH3=063 -1: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_epacif_0p16 -1: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.epacif.0p16.f063.grib2 -1: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f063.grib2 ]] -1: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mod_def.ep_10m ./mod_def.ww3 -1: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.ep_10m ./out_grd.ww3 -1: + wave_grib2_sbs.sh[73]ngrib=1 -1: + wave_grib2_sbs.sh[74]dtgrib=3600 -1: + wave_grib2_sbs.sh[75]tstart='20210326 030000' -1: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210326 030000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.ep_10m.inp.tmpl -1: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -1: $ WAVEWATCH-III gridded output input file -1: $ ---------------------------------------- -1: 20210326 030000 3600 1 -1: N -1: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -1: $ -1: 20210326 030000 7 11 255 0 0 -1: $ -1: $ end of input file -1: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[88]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.460139 ']' -1: ++ prep_step[4]echo gfs_ww3_grib.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[90]export err=0 -1: + wave_grib2_sbs.sh[90]err=0 -1: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[95]cat grib2_epacif_063.out -1: -1: *** WAVEWATCH III GRIB output postp. *** -1: ============================================== -1: -1: Comment character is '$' -1: -1: Grid name : East Pacific 10 min wave grid -1: -1: LINEIN: -1: 20210326 030000 3600 1 -1: -1: 20210326030000 3600 1 -1: GEN_PRO -99999 -1: -1: Output time data : -1: ----------------------------------------------------- -1: First time : 2021/03/26 03:00:00 UTC -1: Interval : 01:00:00 -1: Number of requests : 1 -1: Fields : Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: Requested output fields not yet available: -1: ----------------------------------------------------- -1: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -1: -1: Successfully requested output fields : -1: ----------------------------------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: -1: Additional GRIB parameters : -1: ----------------------------------------------------- -1: Run time : 2021/03/26 03:00:00 UTC -1: GRIB center ID : 7 -1: GRIB gen. proc. ID : 11 -1: GRIB grid ID : 255 -1: GRIB GDS parameter : 0 -1: Fields in file : -1: -------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: CHOSEN GRID TYPE: : LLRECTILINEAR -1: -1: -1: -1: Generating file -1: ----------------------------------------------------- -1: Data for 2021/03/26 03:00:00 UTC 0H forecast. -1: -1: End of program -1: ========================================= -1: WAVEWATCH III GRIB output -1: -1: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -1: + wave_grib2_sbs.sh[102][[ 63 -gt 0 ]] -1: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '63 hour fcst' -grib gfs.wave.t12z.epacif.0p16.f063.grib2 -1: 1:0:d=2021032312:SPC:surface:63 hour fcst: -1: 2:21946:d=2021032312:DIRC:surface:63 hour fcst: -1: 3:50115:d=2021032312:UOGRD:surface:63 hour fcst: -1: 4:71872:d=2021032312:VOGRD:surface:63 hour fcst: -1: 5:93987:d=2021032312:WIND:surface:63 hour fcst: -1: 6:118604:d=2021032312:WDIR:surface:63 hour fcst: -1: 7:146271:d=2021032312:UGRD:surface:63 hour fcst: -1: 8:170873:d=2021032312:VGRD:surface:63 hour fcst: -1: 9:195328:d=2021032312:ICEC:surface:63 hour fcst: -1: 10:214736:d=2021032312:HTSGW:surface:63 hour fcst: -1: 11:237540:d=2021032312:IMWF:surface:63 hour fcst: -1: 12:260178:d=2021032312:MWSPER:surface:63 hour fcst: -1: 13:282767:d=2021032312:PERPW:surface:63 hour fcst: -1: 14:305543:d=2021032312:WWSDIR:surface:63 hour fcst: -1: 15:332443:d=2021032312:DIRPW:surface:63 hour fcst: -1: 16:359348:d=2021032312:WVHGT:surface:63 hour fcst: -1: 17:381001:d=2021032312:SWELL:1 in sequence:63 hour fcst: -1: 18:403502:d=2021032312:SWELL:2 in sequence:63 hour fcst: -1: 19:424357:d=2021032312:SWELL:3 in sequence:63 hour fcst: -1: 20:443935:d=2021032312:WVPER:surface:63 hour fcst: -1: 21:465512:d=2021032312:SWPER:1 in sequence:63 hour fcst: -1: 22:488252:d=2021032312:SWPER:2 in sequence:63 hour fcst: -1: 23:509246:d=2021032312:SWPER:3 in sequence:63 hour fcst: -1: 24:528855:d=2021032312:WVDIR:surface:63 hour fcst: -1: 25:552612:d=2021032312:SWDIR:1 in sequence:63 hour fcst: -1: 26:578927:d=2021032312:SWDIR:2 in sequence:63 hour fcst: -1: 27:601917:d=2021032312:SWDIR:3 in sequence:63 hour fcst: -1: + wave_grib2_sbs.sh[104]err=0 -1: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.epacif.0p16.f063.grib2 -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f063.grib2 ]] -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f063.grib2.idx ]] -1: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.epacif.0p16.f063.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f063.grib2 -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f063.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f063.grib2 = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f063.grib2 ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f063.grib2 -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f063.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f063.grib2.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f063.grib2.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f063.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f063.grib2 -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.epacif.0p16.f063.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f063.grib2.idx -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f063.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f063.grib2.idx = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f063.grib2.idx ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f063.grib2.idx -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f063.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f063.grib2.idx.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f063.grib2.idx.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f063.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f063.grib2.idx -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.epacif.0p16.f063.grib2 and gfs.wave.t12z.epacif.0p16.f063.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_ep_10m to COM' -1: INFO: Copied gfs.wave.t12z.epacif.0p16.f063.grib2 and gfs.wave.t12z.epacif.0p16.f063.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_ep_10m to COM -1: + wave_grib2_sbs.sh[130][[ ep_10m == '' ]] -1: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -1: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.epacif.0p16.f063.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -1: INFO: gfs.wave.t12z.epacif.0p16.f063.grib2 is global.0p50 or SENDDBN is NO, no alert sent -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + cmdfile.6[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032603 3600. 9999 -5: + cmdfile.6[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grid_interp_sbs.sh[25]grdID=gnh_10m -5: + wave_grid_interp_sbs.sh[26]valid_time=2021032603 -5: + wave_grid_interp_sbs.sh[27]dt=3600. -5: + wave_grid_interp_sbs.sh[28]nst=9999 -5: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463 -5: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.uglo_100km ./out_grd.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mod_def.uglo_100km ./mod_def.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mod_def.gnh_10m ./mod_def.gnh_10m -5: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m'\''' -5: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m' -5: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[51]weights_found=1 -5: + wave_grid_interp_sbs.sh[59]ymdhms='20210326 030000' -5: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210326 030000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gnh_10m/g ww3_gint.inp.tmpl -5: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -5: $ Input file for interpolation of uglo_100km to gnh_10m -5: $------------------------------------------------ -5: $ Start Time 3600. NSteps -5: 20210326 030000 3600. 9999 -5: $ Total number of grids -5: 2 -5: $ Grid extensions -5: 'uglo_100km' -5: 'gnh_10m' -5: $ -5: 0 -5: $ -5: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[70]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.460139 ']' -5: ++ prep_step[4]echo gfs_ww3_gint.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -5: INFO: Executing 'gfs_ww3_gint.x' -5: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[73]cat grid_interp.gnh_10m.out -5: -5: *** WAVEWATCH III Grid interpolation *** -5: =============================================== -5: -5: Comment character is '$' -5: -5: Time Information : -5: --------------------------------------------- -5: Starting Time : 2021/03/26 03:00:00 UTC -5: Interval (in sec) : 3600.00 -5: Number of requests : 9999 -5: --------------------------------------------- -5: Number of grids (including output grid) = 2 -5: -5: -5: Extension for grid 1 is --> uglo_100km -5: -5: Grid Particulars are : -5: Dimensions = 45166 1 -5: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Extension for grid 2 is --> gnh_10m -5: -5: Grid Particulars are : -5: Dimensions = 2160 406 -5: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -5: -5: -5: Preparing interpolation weights for output grid -5: Total number of wet points for interpolation 571209 -5: -5: -5: Variable: Grid Interpolation Map Units: 0.100E+01 -5: -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 406 | | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: 355 | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 304 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 253 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 202 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 151 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 100 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 49 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: -5: -5: Interpolating fields .... -5: -5: Output group 1 -5: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -5: Output group 2 -5: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -5: Output group 3 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 4 -5: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -5: Output group 5 -5: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -5: Output group 6 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 7 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 8 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 9 -5: Output variables skipped -5: Output group 10 -5: Output variables skipped -5: ------------------------------------------------ -5: 1Current vel. -5: 1Wind speed -5: 1Ice concentration -5: 2Wave height -5: 2Mean wave period(+2) -5: 2Mean wave period(+1) -5: 2Peak frequency -5: 2Mean wave dir. a1b1 -5: 2Peak direction -5: 4Part. wave height -5: 4Part. peak period -5: 4Part. mean direction -5: 5Charnock parameter -5: ------------------------------------------------ -5: OUTPUT TIME : 2021/03/26 03:00:00 UTC -5: -5: End of file reached -5: -5: -5: *** End of Grid interpolation Routine *** -5: =============================================== -5: -5: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -5: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -5: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_gnh_10m/out_grd.gnh_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.gnh_10m -5: + cmdfile.6[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032603 63 global 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + cmdfile.6[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib2_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grib2_sbs.sh[30]grdID=gnh_10m -5: + wave_grib2_sbs.sh[31]GRIDNR=255 -5: + wave_grib2_sbs.sh[32]MODNR=11 -5: + wave_grib2_sbs.sh[33]valid_time=2021032603 -5: + wave_grib2_sbs.sh[34]fhr=63 -5: + wave_grib2_sbs.sh[35]grid_region=global -5: + wave_grib2_sbs.sh[36]grid_res=0p16 -5: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463 -5: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_gnh_10m -5: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_gnh_10m -5: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_gnh_10m -5: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_gnh_10m -5: ++ wave_grib2_sbs.sh[47]printf %03i 63 -5: + wave_grib2_sbs.sh[47]FH3=063 -5: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p16 -5: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p16.f063.grib2 -5: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f063.grib2 ]] -5: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mod_def.gnh_10m ./mod_def.ww3 -5: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.gnh_10m ./out_grd.ww3 -5: + wave_grib2_sbs.sh[73]ngrib=1 -5: + wave_grib2_sbs.sh[74]dtgrib=3600 -5: + wave_grib2_sbs.sh[75]tstart='20210326 030000' -5: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210326 030000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gnh_10m.inp.tmpl -5: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -5: $ WAVEWATCH-III gridded output input file -5: $ ---------------------------------------- -5: 20210326 030000 3600 1 -5: N -5: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -5: $ -5: 20210326 030000 7 11 255 0 0 -5: $ -5: $ end of input file -5: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[88]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.460139 ']' -5: ++ prep_step[4]echo gfs_ww3_grib.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[90]export err=0 -5: + wave_grib2_sbs.sh[90]err=0 -5: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[95]cat grib2_global_063.out -5: -5: *** WAVEWATCH III GRIB output postp. *** -5: ============================================== -5: -5: Comment character is '$' -5: -5: Grid name : GFSv16-wave N Hemisphere 1/6 d -5: -5: LINEIN: -5: 20210326 030000 3600 1 -5: -5: 20210326030000 3600 1 -5: GEN_PRO -99999 -5: -5: Output time data : -5: ----------------------------------------------------- -5: First time : 2021/03/26 03:00:00 UTC -5: Interval : 01:00:00 -5: Number of requests : 1 -5: Fields : Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: Requested output fields not yet available: -5: ----------------------------------------------------- -5: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -5: -5: Successfully requested output fields : -5: ----------------------------------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: -5: Additional GRIB parameters : -5: ----------------------------------------------------- -5: Run time : 2021/03/26 03:00:00 UTC -5: GRIB center ID : 7 -5: GRIB gen. proc. ID : 11 -5: GRIB grid ID : 255 -5: GRIB GDS parameter : 0 -5: Fields in file : -5: -------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: CHOSEN GRID TYPE: : LLRECTILINEAR -5: -5: -5: -5: Generating file -5: ----------------------------------------------------- -5: Data for 2021/03/26 03:00:00 UTC 0H forecast. -5: -5: End of program -5: ========================================= -5: WAVEWATCH III GRIB output -5: -5: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -5: + wave_grib2_sbs.sh[102][[ 63 -gt 0 ]] -5: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '63 hour fcst' -grib gfs.wave.t12z.global.0p16.f063.grib2 -5: 1:0:d=2021032312:SPC:surface:63 hour fcst: -5: 2:236703:d=2021032312:DIRC:surface:63 hour fcst: -5: 3:863548:d=2021032312:UOGRD:surface:63 hour fcst: -5: 4:1095227:d=2021032312:VOGRD:surface:63 hour fcst: -5: 5:1334687:d=2021032312:WIND:surface:63 hour fcst: -5: 6:1735295:d=2021032312:WDIR:surface:63 hour fcst: -5: 7:2375163:d=2021032312:UGRD:surface:63 hour fcst: -5: 8:2772051:d=2021032312:VGRD:surface:63 hour fcst: -5: 9:3169874:d=2021032312:ICEC:surface:63 hour fcst: -5: 10:3284558:d=2021032312:HTSGW:surface:63 hour fcst: -5: 11:3567760:d=2021032312:IMWF:surface:63 hour fcst: -5: 12:3876875:d=2021032312:MWSPER:surface:63 hour fcst: -5: 13:4187825:d=2021032312:PERPW:surface:63 hour fcst: -5: 14:4519489:d=2021032312:WWSDIR:surface:63 hour fcst: -5: 15:5111938:d=2021032312:DIRPW:surface:63 hour fcst: -5: 16:5730360:d=2021032312:WVHGT:surface:63 hour fcst: -5: 17:6009256:d=2021032312:SWELL:1 in sequence:63 hour fcst: -5: 18:6265544:d=2021032312:SWELL:2 in sequence:63 hour fcst: -5: 19:6436711:d=2021032312:SWELL:3 in sequence:63 hour fcst: -5: 20:6555629:d=2021032312:WVPER:surface:63 hour fcst: -5: 21:6874175:d=2021032312:SWPER:1 in sequence:63 hour fcst: -5: 22:7166368:d=2021032312:SWPER:2 in sequence:63 hour fcst: -5: 23:7368833:d=2021032312:SWPER:3 in sequence:63 hour fcst: -5: 24:7495386:d=2021032312:WVDIR:surface:63 hour fcst: -5: 25:8015026:d=2021032312:SWDIR:1 in sequence:63 hour fcst: -5: 26:8529713:d=2021032312:SWDIR:2 in sequence:63 hour fcst: -5: 27:8836376:d=2021032312:SWDIR:3 in sequence:63 hour fcst: -5: + wave_grib2_sbs.sh[104]err=0 -5: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p16.f063.grib2 -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f063.grib2 ]] -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f063.grib2.idx ]] -5: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p16.f063.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f063.grib2 -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f063.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f063.grib2 = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f063.grib2 ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f063.grib2 -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f063.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f063.grib2.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f063.grib2.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f063.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f063.grib2 -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p16.f063.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f063.grib2.idx -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f063.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f063.grib2.idx = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f063.grib2.idx ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f063.grib2.idx -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f063.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f063.grib2.idx.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f063.grib2.idx.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f063.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f063.grib2.idx -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p16.f063.grib2 and gfs.wave.t12z.global.0p16.f063.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_gnh_10m to COM' -5: INFO: Copied gfs.wave.t12z.global.0p16.f063.grib2 and gfs.wave.t12z.global.0p16.f063.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_gnh_10m to COM -5: + wave_grib2_sbs.sh[130][[ gnh_10m == '' ]] -5: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -5: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p16.f063.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -5: INFO: gfs.wave.t12z.global.0p16.f063.grib2 is global.0p50 or SENDDBN is NO, no alert sent -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + cmdfile.4[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032603 3600. 9999 -3: + cmdfile.4[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grid_interp_sbs.sh[25]grdID=glo_30m -3: + wave_grid_interp_sbs.sh[26]valid_time=2021032603 -3: + wave_grid_interp_sbs.sh[27]dt=3600. -3: + wave_grid_interp_sbs.sh[28]nst=9999 -3: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463 -3: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.uglo_100km ./out_grd.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mod_def.uglo_100km ./mod_def.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mod_def.glo_30m ./mod_def.glo_30m -3: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ]] -3: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m'\''' -3: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m' -3: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[51]weights_found=1 -3: + wave_grid_interp_sbs.sh[59]ymdhms='20210326 030000' -3: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210326 030000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/glo_30m/g ww3_gint.inp.tmpl -3: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -3: $ Input file for interpolation of uglo_100km to glo_30m -3: $------------------------------------------------ -3: $ Start Time 3600. NSteps -3: 20210326 030000 3600. 9999 -3: $ Total number of grids -3: 2 -3: $ Grid extensions -3: 'uglo_100km' -3: 'glo_30m' -3: $ -3: 0 -3: $ -3: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[70]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.460139 ']' -3: ++ prep_step[4]echo gfs_ww3_gint.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -3: INFO: Executing 'gfs_ww3_gint.x' -3: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[73]cat grid_interp.glo_30m.out -3: -3: *** WAVEWATCH III Grid interpolation *** -3: =============================================== -3: -3: Comment character is '$' -3: -3: Time Information : -3: --------------------------------------------- -3: Starting Time : 2021/03/26 03:00:00 UTC -3: Interval (in sec) : 3600.00 -3: Number of requests : 9999 -3: --------------------------------------------- -3: Number of grids (including output grid) = 2 -3: -3: -3: Extension for grid 1 is --> uglo_100km -3: -3: Grid Particulars are : -3: Dimensions = 45166 1 -3: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Extension for grid 2 is --> glo_30m -3: -3: Grid Particulars are : -3: Dimensions = 720 336 -3: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -3: -3: -3: Preparing interpolation weights for output grid -3: Total number of wet points for interpolation 167619 -3: -3: -3: Variable: Grid Interpolation Map Units: 0.100E+01 -3: -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 336 | | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 291 | 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 | -3: 246 | 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 201 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 156 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 111 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 66 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 21 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 | -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: -3: -3: Interpolating fields .... -3: -3: Output group 1 -3: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -3: Output group 2 -3: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -3: Output group 3 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 4 -3: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -3: Output group 5 -3: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -3: Output group 6 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 7 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 8 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 9 -3: Output variables skipped -3: Output group 10 -3: Output variables skipped -3: ------------------------------------------------ -3: 1Current vel. -3: 1Wind speed -3: 1Ice concentration -3: 2Wave height -3: 2Mean wave period(+2) -3: 2Mean wave period(+1) -3: 2Peak frequency -3: 2Mean wave dir. a1b1 -3: 2Peak direction -3: 4Part. wave height -3: 4Part. peak period -3: 4Part. mean direction -3: 5Charnock parameter -3: ------------------------------------------------ -3: OUTPUT TIME : 2021/03/26 03:00:00 UTC -3: -3: End of file reached -3: -3: -3: *** End of Grid interpolation Routine *** -3: =============================================== -3: -3: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -3: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -3: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_glo_30m/out_grd.glo_30m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.glo_30m -3: + cmdfile.4[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032603 63 global 0p50 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + cmdfile.4[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib2_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grib2_sbs.sh[30]grdID=glo_30m -3: + wave_grib2_sbs.sh[31]GRIDNR=255 -3: + wave_grib2_sbs.sh[32]MODNR=11 -3: + wave_grib2_sbs.sh[33]valid_time=2021032603 -3: + wave_grib2_sbs.sh[34]fhr=63 -3: + wave_grib2_sbs.sh[35]grid_region=global -3: + wave_grib2_sbs.sh[36]grid_res=0p50 -3: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463 -3: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_glo_30m -3: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_glo_30m -3: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_glo_30m -3: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_glo_30m -3: ++ wave_grib2_sbs.sh[47]printf %03i 63 -3: + wave_grib2_sbs.sh[47]FH3=063 -3: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p50 -3: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p50.f063.grib2 -3: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f063.grib2 ]] -3: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mod_def.glo_30m ./mod_def.ww3 -3: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.glo_30m ./out_grd.ww3 -3: + wave_grib2_sbs.sh[73]ngrib=1 -3: + wave_grib2_sbs.sh[74]dtgrib=3600 -3: + wave_grib2_sbs.sh[75]tstart='20210326 030000' -3: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210326 030000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.glo_30m.inp.tmpl -3: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -3: $ WAVEWATCH-III gridded output input file -3: $ ---------------------------------------- -3: 20210326 030000 3600 1 -3: N -3: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -3: $ -3: 20210326 030000 7 11 255 0 0 -3: $ -3: $ end of input file -3: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[88]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.460139 ']' -3: ++ prep_step[4]echo gfs_ww3_grib.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[90]export err=0 -3: + wave_grib2_sbs.sh[90]err=0 -3: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[95]cat grib2_global_063.out -3: -3: *** WAVEWATCH III GRIB output postp. *** -3: ============================================== -3: -3: Comment character is '$' -3: -3: Grid name : Global 30 min wave grid -3: -3: LINEIN: -3: 20210326 030000 3600 1 -3: -3: 20210326030000 3600 1 -3: GEN_PRO -99999 -3: -3: Output time data : -3: ----------------------------------------------------- -3: First time : 2021/03/26 03:00:00 UTC -3: Interval : 01:00:00 -3: Number of requests : 1 -3: Fields : Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: Requested output fields not yet available: -3: ----------------------------------------------------- -3: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -3: -3: Successfully requested output fields : -3: ----------------------------------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: -3: Additional GRIB parameters : -3: ----------------------------------------------------- -3: Run time : 2021/03/26 03:00:00 UTC -3: GRIB center ID : 7 -3: GRIB gen. proc. ID : 11 -3: GRIB grid ID : 255 -3: GRIB GDS parameter : 0 -3: Fields in file : -3: -------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: CHOSEN GRID TYPE: : LLRECTILINEAR -3: -3: -3: -3: Generating file -3: ----------------------------------------------------- -3: Data for 2021/03/26 03:00:00 UTC 0H forecast. -3: -3: End of program -3: ========================================= -3: WAVEWATCH III GRIB output -3: -3: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -3: + wave_grib2_sbs.sh[102][[ 63 -gt 0 ]] -3: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '63 hour fcst' -grib gfs.wave.t12z.global.0p50.f063.grib2 -3: 1:0:d=2021032312:SPC:surface:63 hour fcst: -3: 2:77859:d=2021032312:DIRC:surface:63 hour fcst: -3: 3:319215:d=2021032312:UOGRD:surface:63 hour fcst: -3: 4:394668:d=2021032312:VOGRD:surface:63 hour fcst: -3: 5:474487:d=2021032312:WIND:surface:63 hour fcst: -3: 6:632334:d=2021032312:WDIR:surface:63 hour fcst: -3: 7:867572:d=2021032312:UGRD:surface:63 hour fcst: -3: 8:1022299:d=2021032312:VGRD:surface:63 hour fcst: -3: 9:1179037:d=2021032312:ICEC:surface:63 hour fcst: -3: 10:1219350:d=2021032312:HTSGW:surface:63 hour fcst: -3: 11:1318083:d=2021032312:IMWF:surface:63 hour fcst: -3: 12:1426368:d=2021032312:MWSPER:surface:63 hour fcst: -3: 13:1536035:d=2021032312:PERPW:surface:63 hour fcst: -3: 14:1655010:d=2021032312:WWSDIR:surface:63 hour fcst: -3: 15:1858915:d=2021032312:DIRPW:surface:63 hour fcst: -3: 16:2071140:d=2021032312:WVHGT:surface:63 hour fcst: -3: 17:2169829:d=2021032312:SWELL:1 in sequence:63 hour fcst: -3: 18:2264372:d=2021032312:SWELL:2 in sequence:63 hour fcst: -3: 19:2325411:d=2021032312:SWELL:3 in sequence:63 hour fcst: -3: 20:2362510:d=2021032312:WVPER:surface:63 hour fcst: -3: 21:2474292:d=2021032312:SWPER:1 in sequence:63 hour fcst: -3: 22:2581594:d=2021032312:SWPER:2 in sequence:63 hour fcst: -3: 23:2656237:d=2021032312:SWPER:3 in sequence:63 hour fcst: -3: 24:2697739:d=2021032312:WVDIR:surface:63 hour fcst: -3: 25:2871778:d=2021032312:SWDIR:1 in sequence:63 hour fcst: -3: 26:3054810:d=2021032312:SWDIR:2 in sequence:63 hour fcst: -3: 27:3169711:d=2021032312:SWDIR:3 in sequence:63 hour fcst: -3: + wave_grib2_sbs.sh[104]err=0 -3: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p50.f063.grib2 -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f063.grib2 ]] -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f063.grib2.idx ]] -3: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p50.f063.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f063.grib2 -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f063.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f063.grib2 = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f063.grib2 ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f063.grib2 -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f063.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f063.grib2.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f063.grib2.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f063.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f063.grib2 -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p50.f063.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f063.grib2.idx -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f063.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f063.grib2.idx = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f063.grib2.idx ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f063.grib2.idx -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f063.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f063.grib2.idx.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f063.grib2.idx.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f063.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f063.grib2.idx -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p50.f063.grib2 and gfs.wave.t12z.global.0p50.f063.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_glo_30m to COM' -3: INFO: Copied gfs.wave.t12z.global.0p50.f063.grib2 and gfs.wave.t12z.global.0p50.f063.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_glo_30m to COM -3: + wave_grib2_sbs.sh[130][[ glo_30m == '' ]] -3: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -3: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p50.f063.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -3: INFO: gfs.wave.t12z.global.0p50.f063.grib2 is global.0p50 or SENDDBN is NO, no alert sent -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + cmdfile.3[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032603 3600. 9999 -2: + cmdfile.3[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grid_interp_sbs.sh[25]grdID=wc_10m -2: + wave_grid_interp_sbs.sh[26]valid_time=2021032603 -2: + wave_grid_interp_sbs.sh[27]dt=3600. -2: + wave_grid_interp_sbs.sh[28]nst=9999 -2: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463 -2: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.uglo_100km ./out_grd.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mod_def.uglo_100km ./mod_def.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mod_def.wc_10m ./mod_def.wc_10m -2: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ]] -2: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m'\''' -2: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m' -2: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[51]weights_found=1 -2: + wave_grid_interp_sbs.sh[59]ymdhms='20210326 030000' -2: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210326 030000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/wc_10m/g ww3_gint.inp.tmpl -2: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -2: $ Input file for interpolation of uglo_100km to wc_10m -2: $------------------------------------------------ -2: $ Start Time 3600. NSteps -2: 20210326 030000 3600. 9999 -2: $ Total number of grids -2: 2 -2: $ Grid extensions -2: 'uglo_100km' -2: 'wc_10m' -2: $ -2: 0 -2: $ -2: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[70]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.460139 ']' -2: ++ prep_step[4]echo gfs_ww3_gint.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -2: INFO: Executing 'gfs_ww3_gint.x' -2: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[73]cat grid_interp.wc_10m.out -2: -2: *** WAVEWATCH III Grid interpolation *** -2: =============================================== -2: -2: Comment character is '$' -2: -2: Time Information : -2: --------------------------------------------- -2: Starting Time : 2021/03/26 03:00:00 UTC -2: Interval (in sec) : 3600.00 -2: Number of requests : 9999 -2: --------------------------------------------- -2: Number of grids (including output grid) = 2 -2: -2: -2: Extension for grid 1 is --> uglo_100km -2: -2: Grid Particulars are : -2: Dimensions = 45166 1 -2: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Extension for grid 2 is --> wc_10m -2: -2: Grid Particulars are : -2: Dimensions = 241 151 -2: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -2: -2: -2: Preparing interpolation weights for output grid -2: Total number of wet points for interpolation 11044 -2: -2: -2: Variable: Grid Interpolation Map Units: 0.100E+01 -2: -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: +---------------------------------------------------------------------------------------------------------------+ -2: 151 | 0 0 0 0 | -2: | 0 0 0 0 0 | -2: | 0 0 0 0 0 | -2: 130 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 109 | 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 88 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 | -2: 67 | 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 | -2: 46 | 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: 25 | 0 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 0 | -2: | | -2: 4 | | -2: +---------------------------------------------------------------------------------------------------------------+ -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: -2: -2: Interpolating fields .... -2: -2: Output group 1 -2: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -2: Output group 2 -2: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -2: Output group 3 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 4 -2: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -2: Output group 5 -2: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -2: Output group 6 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 7 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 8 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 9 -2: Output variables skipped -2: Output group 10 -2: Output variables skipped -2: ------------------------------------------------ -2: 1Current vel. -2: 1Wind speed -2: 1Ice concentration -2: 2Wave height -2: 2Mean wave period(+2) -2: 2Mean wave period(+1) -2: 2Peak frequency -2: 2Mean wave dir. a1b1 -2: 2Peak direction -2: 4Part. wave height -2: 4Part. peak period -2: 4Part. mean direction -2: 5Charnock parameter -2: ------------------------------------------------ -2: OUTPUT TIME : 2021/03/26 03:00:00 UTC -2: -2: End of file reached -2: -2: -2: *** End of Grid interpolation Routine *** -2: =============================================== -2: -2: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -2: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -2: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_wc_10m/out_grd.wc_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.wc_10m -2: + cmdfile.3[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032603 63 wcoast 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + cmdfile.3[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib2_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grib2_sbs.sh[30]grdID=wc_10m -2: + wave_grib2_sbs.sh[31]GRIDNR=255 -2: + wave_grib2_sbs.sh[32]MODNR=11 -2: + wave_grib2_sbs.sh[33]valid_time=2021032603 -2: + wave_grib2_sbs.sh[34]fhr=63 -2: + wave_grib2_sbs.sh[35]grid_region=wcoast -2: + wave_grib2_sbs.sh[36]grid_res=0p16 -2: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463 -2: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_wc_10m -2: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_wc_10m -2: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_wc_10m -2: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_wc_10m -2: ++ wave_grib2_sbs.sh[47]printf %03i 63 -2: + wave_grib2_sbs.sh[47]FH3=063 -2: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_wcoast_0p16 -2: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.wcoast.0p16.f063.grib2 -2: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f063.grib2 ]] -2: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mod_def.wc_10m ./mod_def.ww3 -2: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.wc_10m ./out_grd.ww3 -2: + wave_grib2_sbs.sh[73]ngrib=1 -2: + wave_grib2_sbs.sh[74]dtgrib=3600 -2: + wave_grib2_sbs.sh[75]tstart='20210326 030000' -2: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210326 030000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.wc_10m.inp.tmpl -2: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -2: $ WAVEWATCH-III gridded output input file -2: $ ---------------------------------------- -2: 20210326 030000 3600 1 -2: N -2: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -2: $ -2: 20210326 030000 7 11 255 0 0 -2: $ -2: $ end of input file -2: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[88]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.460139 ']' -2: ++ prep_step[4]echo gfs_ww3_grib.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[90]export err=0 -2: + wave_grib2_sbs.sh[90]err=0 -2: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[95]cat grib2_wcoast_063.out -2: -2: *** WAVEWATCH III GRIB output postp. *** -2: ============================================== -2: -2: Comment character is '$' -2: -2: Grid name : West Coast 10 min wave grid -2: -2: LINEIN: -2: 20210326 030000 3600 1 -2: -2: 20210326030000 3600 1 -2: GEN_PRO -99999 -2: -2: Output time data : -2: ----------------------------------------------------- -2: First time : 2021/03/26 03:00:00 UTC -2: Interval : 01:00:00 -2: Number of requests : 1 -2: Fields : Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: Requested output fields not yet available: -2: ----------------------------------------------------- -2: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -2: -2: Successfully requested output fields : -2: ----------------------------------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: -2: Additional GRIB parameters : -2: ----------------------------------------------------- -2: Run time : 2021/03/26 03:00:00 UTC -2: GRIB center ID : 7 -2: GRIB gen. proc. ID : 11 -2: GRIB grid ID : 255 -2: GRIB GDS parameter : 0 -2: Fields in file : -2: -------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: CHOSEN GRID TYPE: : LLRECTILINEAR -2: -2: -2: -2: Generating file -2: ----------------------------------------------------- -2: Data for 2021/03/26 03:00:00 UTC 0H forecast. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: -2: End of program -2: ========================================= -2: WAVEWATCH III GRIB output -2: -2: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -2: + wave_grib2_sbs.sh[102][[ 63 -gt 0 ]] -2: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '63 hour fcst' -grib gfs.wave.t12z.wcoast.0p16.f063.grib2 -2: 1:0:d=2021032312:SPC:surface:63 hour fcst: -2: 2:7297:d=2021032312:DIRC:surface:63 hour fcst: -2: 3:23858:d=2021032312:UOGRD:surface:63 hour fcst: -2: 4:30827:d=2021032312:VOGRD:surface:63 hour fcst: -2: 5:38429:d=2021032312:WIND:surface:63 hour fcst: -2: 6:49688:d=2021032312:WDIR:surface:63 hour fcst: -2: 7:65592:d=2021032312:UGRD:surface:63 hour fcst: -2: 8:76919:d=2021032312:VGRD:surface:63 hour fcst: -2: 9:88447:d=2021032312:ICEC:surface:63 hour fcst: -2: 10:93177:d=2021032312:HTSGW:surface:63 hour fcst: -2: 11:102702:d=2021032312:IMWF:surface:63 hour fcst: -2: 12:112247:d=2021032312:MWSPER:surface:63 hour fcst: -2: 13:121874:d=2021032312:PERPW:surface:63 hour fcst: -2: 14:131765:d=2021032312:WWSDIR:surface:63 hour fcst: -2: 15:146798:d=2021032312:DIRPW:surface:63 hour fcst: -2: 16:161971:d=2021032312:WVHGT:surface:63 hour fcst: -2: 17:171593:d=2021032312:SWELL:1 in sequence:63 hour fcst: -2: 18:176885:d=2021032312:SWELL:2 in sequence:63 hour fcst: -2: 19:181615:d=2021032312:SWELL:3 in sequence:63 hour fcst: -2: 20:186345:d=2021032312:WVPER:surface:63 hour fcst: -2: 21:196267:d=2021032312:SWPER:1 in sequence:63 hour fcst: -2: 22:201527:d=2021032312:SWPER:2 in sequence:63 hour fcst: -2: 23:206257:d=2021032312:SWPER:3 in sequence:63 hour fcst: -2: 24:210987:d=2021032312:WVDIR:surface:63 hour fcst: -2: 25:226143:d=2021032312:SWDIR:1 in sequence:63 hour fcst: -2: 26:231737:d=2021032312:SWDIR:2 in sequence:63 hour fcst: -2: 27:236467:d=2021032312:SWDIR:3 in sequence:63 hour fcst: -2: + wave_grib2_sbs.sh[104]err=0 -2: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.wcoast.0p16.f063.grib2 -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f063.grib2 ]] -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f063.grib2.idx ]] -2: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.wcoast.0p16.f063.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f063.grib2 -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f063.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f063.grib2 = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f063.grib2 ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f063.grib2 -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f063.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f063.grib2.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f063.grib2.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f063.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f063.grib2 -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.wcoast.0p16.f063.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f063.grib2.idx -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f063.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f063.grib2.idx = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f063.grib2.idx ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f063.grib2.idx -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f063.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f063.grib2.idx.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f063.grib2.idx.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f063.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f063.grib2.idx -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.wcoast.0p16.f063.grib2 and gfs.wave.t12z.wcoast.0p16.f063.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_wc_10m to COM' -2: INFO: Copied gfs.wave.t12z.wcoast.0p16.f063.grib2 and gfs.wave.t12z.wcoast.0p16.f063.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_wc_10m to COM -2: + wave_grib2_sbs.sh[130][[ wc_10m == '' ]] -2: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -2: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.wcoast.0p16.f063.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -2: INFO: gfs.wave.t12z.wcoast.0p16.f063.grib2 is global.0p50 or SENDDBN is NO, no alert sent -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + cmdfile.1[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032603 3600. 9999 -0: + cmdfile.1[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grid_interp_sbs.sh[25]grdID=at_10m -0: + wave_grid_interp_sbs.sh[26]valid_time=2021032603 -0: + wave_grid_interp_sbs.sh[27]dt=3600. -0: + wave_grid_interp_sbs.sh[28]nst=9999 -0: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463 -0: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.uglo_100km ./out_grd.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mod_def.uglo_100km ./mod_def.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mod_def.at_10m ./mod_def.at_10m -0: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ]] -0: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m'\''' -0: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m' -0: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[51]weights_found=1 -0: + wave_grid_interp_sbs.sh[59]ymdhms='20210326 030000' -0: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210326 030000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/at_10m/g ww3_gint.inp.tmpl -0: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -0: $ Input file for interpolation of uglo_100km to at_10m -0: $------------------------------------------------ -0: $ Start Time 3600. NSteps -0: 20210326 030000 3600. 9999 -0: $ Total number of grids -0: 2 -0: $ Grid extensions -0: 'uglo_100km' -0: 'at_10m' -0: $ -0: 0 -0: $ -0: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[70]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.460139 ']' -0: ++ prep_step[4]echo gfs_ww3_gint.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -0: INFO: Executing 'gfs_ww3_gint.x' -0: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[73]cat grid_interp.at_10m.out -0: -0: *** WAVEWATCH III Grid interpolation *** -0: =============================================== -0: -0: Comment character is '$' -0: -0: Time Information : -0: --------------------------------------------- -0: Starting Time : 2021/03/26 03:00:00 UTC -0: Interval (in sec) : 3600.00 -0: Number of requests : 9999 -0: --------------------------------------------- -0: Number of grids (including output grid) = 2 -0: -0: -0: Extension for grid 1 is --> uglo_100km -0: -0: Grid Particulars are : -0: Dimensions = 45166 1 -0: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Extension for grid 2 is --> at_10m -0: -0: Grid Particulars are : -0: Dimensions = 301 331 -0: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -0: -0: -0: Preparing interpolation weights for output grid -0: Total number of wet points for interpolation 29591 -0: -0: -0: Variable: Grid Interpolation Map Units: 0.100E+01 -0: -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 331 | | -0: | | -0: | | -0: 289 | | -0: | | -0: | 0 0 0 | -0: 247 | 0 0 0 0 | -0: | 0 0 0 0 0 | -0: | 0 0 0 0 | -0: 205 | 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 163 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 121 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 79 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 | -0: | 0 0 | -0: 37 | 0 | -0: | | -0: | | -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: -0: -0: Interpolating fields .... -0: -0: Output group 1 -0: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -0: Output group 2 -0: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -0: Output group 3 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 4 -0: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -0: Output group 5 -0: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -0: Output group 6 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 7 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 8 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 9 -0: Output variables skipped -0: Output group 10 -0: Output variables skipped -0: ------------------------------------------------ -0: 1Current vel. -0: 1Wind speed -0: 1Ice concentration -0: 2Wave height -0: 2Mean wave period(+2) -0: 2Mean wave period(+1) -0: 2Peak frequency -0: 2Mean wave dir. a1b1 -0: 2Peak direction -0: 4Part. wave height -0: 4Part. peak period -0: 4Part. mean direction -0: 5Charnock parameter -0: ------------------------------------------------ -0: OUTPUT TIME : 2021/03/26 03:00:00 UTC -0: -0: End of file reached -0: -0: -0: *** End of Grid interpolation Routine *** -0: =============================================== -0: -0: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -0: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -0: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_at_10m/out_grd.at_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.at_10m -0: + cmdfile.1[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032603 63 atlocn 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + cmdfile.1[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib2_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grib2_sbs.sh[30]grdID=at_10m -0: + wave_grib2_sbs.sh[31]GRIDNR=255 -0: + wave_grib2_sbs.sh[32]MODNR=11 -0: + wave_grib2_sbs.sh[33]valid_time=2021032603 -0: + wave_grib2_sbs.sh[34]fhr=63 -0: + wave_grib2_sbs.sh[35]grid_region=atlocn -0: + wave_grib2_sbs.sh[36]grid_res=0p16 -0: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463 -0: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_at_10m -0: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_at_10m -0: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_at_10m -0: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_at_10m -0: ++ wave_grib2_sbs.sh[47]printf %03i 63 -0: + wave_grib2_sbs.sh[47]FH3=063 -0: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_atlocn_0p16 -0: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.atlocn.0p16.f063.grib2 -0: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f063.grib2 ]] -0: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mod_def.at_10m ./mod_def.ww3 -0: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.at_10m ./out_grd.ww3 -0: + wave_grib2_sbs.sh[73]ngrib=1 -0: + wave_grib2_sbs.sh[74]dtgrib=3600 -0: + wave_grib2_sbs.sh[75]tstart='20210326 030000' -0: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210326 030000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.at_10m.inp.tmpl -0: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -0: $ WAVEWATCH-III gridded output input file -0: $ ---------------------------------------- -0: 20210326 030000 3600 1 -0: N -0: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -0: $ -0: 20210326 030000 7 11 255 0 0 -0: $ -0: $ end of input file -0: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[88]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.460139 ']' -0: ++ prep_step[4]echo gfs_ww3_grib.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[90]export err=0 -0: + wave_grib2_sbs.sh[90]err=0 -0: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[95]cat grib2_atlocn_063.out -0: -0: *** WAVEWATCH III GRIB output postp. *** -0: ============================================== -0: -0: Comment character is '$' -0: -0: Grid name : NW Atlantic 10 min wave grid -0: -0: LINEIN: -0: 20210326 030000 3600 1 -0: -0: 20210326030000 3600 1 -0: GEN_PRO -99999 -0: -0: Output time data : -0: ----------------------------------------------------- -0: First time : 2021/03/26 03:00:00 UTC -0: Interval : 01:00:00 -0: Number of requests : 1 -0: Fields : Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: Requested output fields not yet available: -0: ----------------------------------------------------- -0: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -0: -0: Successfully requested output fields : -0: ----------------------------------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: -0: Additional GRIB parameters : -0: ----------------------------------------------------- -0: Run time : 2021/03/26 03:00:00 UTC -0: GRIB center ID : 7 -0: GRIB gen. proc. ID : 11 -0: GRIB grid ID : 255 -0: GRIB GDS parameter : 0 -0: Fields in file : -0: -------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: CHOSEN GRID TYPE: : LLRECTILINEAR -0: -0: -0: -0: Generating file -0: ----------------------------------------------------- -0: Data for 2021/03/26 03:00:00 UTC 0H forecast. -0: -0: End of program -0: ========================================= -0: WAVEWATCH III GRIB output -0: -0: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -0: + wave_grib2_sbs.sh[102][[ 63 -gt 0 ]] -0: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '63 hour fcst' -grib gfs.wave.t12z.atlocn.0p16.f063.grib2 -0: 1:0:d=2021032312:SPC:surface:63 hour fcst: -0: 2:20375:d=2021032312:DIRC:surface:63 hour fcst: -0: 3:60400:d=2021032312:UOGRD:surface:63 hour fcst: -0: 4:80150:d=2021032312:VOGRD:surface:63 hour fcst: -0: 5:100475:d=2021032312:WIND:surface:63 hour fcst: -0: 6:131359:d=2021032312:WDIR:surface:63 hour fcst: -0: 7:172703:d=2021032312:UGRD:surface:63 hour fcst: -0: 8:203130:d=2021032312:VGRD:surface:63 hour fcst: -0: 9:233565:d=2021032312:ICEC:surface:63 hour fcst: -0: 10:246200:d=2021032312:HTSGW:surface:63 hour fcst: -0: 11:271003:d=2021032312:IMWF:surface:63 hour fcst: -0: 12:297421:d=2021032312:MWSPER:surface:63 hour fcst: -0: 13:324060:d=2021032312:PERPW:surface:63 hour fcst: -0: 14:352103:d=2021032312:WWSDIR:surface:63 hour fcst: -0: 15:392106:d=2021032312:DIRPW:surface:63 hour fcst: -0: 16:433724:d=2021032312:WVHGT:surface:63 hour fcst: -0: 17:458219:d=2021032312:SWELL:1 in sequence:63 hour fcst: -0: 18:479128:d=2021032312:SWELL:2 in sequence:63 hour fcst: -0: 19:495143:d=2021032312:SWELL:3 in sequence:63 hour fcst: -0: 20:508572:d=2021032312:WVPER:surface:63 hour fcst: -0: 21:536031:d=2021032312:SWPER:1 in sequence:63 hour fcst: -0: 22:559013:d=2021032312:SWPER:2 in sequence:63 hour fcst: -0: 23:576688:d=2021032312:SWPER:3 in sequence:63 hour fcst: -0: 24:590463:d=2021032312:WVDIR:surface:63 hour fcst: -0: 25:628834:d=2021032312:SWDIR:1 in sequence:63 hour fcst: -0: 26:661657:d=2021032312:SWDIR:2 in sequence:63 hour fcst: -0: 27:683815:d=2021032312:SWDIR:3 in sequence:63 hour fcst: -0: + wave_grib2_sbs.sh[104]err=0 -0: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.atlocn.0p16.f063.grib2 -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f063.grib2 ]] -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f063.grib2.idx ]] -0: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.atlocn.0p16.f063.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f063.grib2 -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f063.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f063.grib2 = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f063.grib2 ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f063.grib2 -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f063.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f063.grib2.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f063.grib2.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f063.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f063.grib2 -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.atlocn.0p16.f063.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f063.grib2.idx -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f063.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f063.grib2.idx = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f063.grib2.idx ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f063.grib2.idx -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f063.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f063.grib2.idx.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f063.grib2.idx.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f063.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f063.grib2.idx -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.atlocn.0p16.f063.grib2 and gfs.wave.t12z.atlocn.0p16.f063.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_at_10m to COM' -0: INFO: Copied gfs.wave.t12z.atlocn.0p16.f063.grib2 and gfs.wave.t12z.atlocn.0p16.f063.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_at_10m to COM -0: + wave_grib2_sbs.sh[130][[ at_10m == '' ]] -0: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -0: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.atlocn.0p16.f063.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -0: INFO: gfs.wave.t12z.atlocn.0p16.f063.grib2 is global.0p50 or SENDDBN is NO, no alert sent -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + cmdfile.5[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032603 3600. 9999 -4: + cmdfile.5[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grid_interp_sbs.sh[25]grdID=aoc_9km -4: + wave_grid_interp_sbs.sh[26]valid_time=2021032603 -4: + wave_grid_interp_sbs.sh[27]dt=3600. -4: + wave_grid_interp_sbs.sh[28]nst=9999 -4: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463 -4: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.uglo_100km ./out_grd.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mod_def.uglo_100km ./mod_def.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mod_def.aoc_9km ./mod_def.aoc_9km -4: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km'\''' -4: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km' -4: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[51]weights_found=1 -4: + wave_grid_interp_sbs.sh[59]ymdhms='20210326 030000' -4: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210326 030000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/aoc_9km/g ww3_gint.inp.tmpl -4: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -4: $ Input file for interpolation of uglo_100km to aoc_9km -4: $------------------------------------------------ -4: $ Start Time 3600. NSteps -4: 20210326 030000 3600. 9999 -4: $ Total number of grids -4: 2 -4: $ Grid extensions -4: 'uglo_100km' -4: 'aoc_9km' -4: $ -4: 0 -4: $ -4: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[70]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.460139 ']' -4: ++ prep_step[4]echo gfs_ww3_gint.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -4: INFO: Executing 'gfs_ww3_gint.x' -4: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[73]cat grid_interp.aoc_9km.out -4: -4: *** WAVEWATCH III Grid interpolation *** -4: =============================================== -4: -4: Comment character is '$' -4: -4: Time Information : -4: --------------------------------------------- -4: Starting Time : 2021/03/26 03:00:00 UTC -4: Interval (in sec) : 3600.00 -4: Number of requests : 9999 -4: --------------------------------------------- -4: Number of grids (including output grid) = 2 -4: -4: -4: Extension for grid 1 is --> uglo_100km -4: -4: Grid Particulars are : -4: Dimensions = 45166 1 -4: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Extension for grid 2 is --> aoc_9km -4: -4: Grid Particulars are : -4: Dimensions = 1006 1006 -4: Grid Type = 2 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -4: -4: -4: Preparing interpolation weights for output grid -4: Total number of wet points for interpolation 360052 -4: -4: -4: Variable: Grid Interpolation Map Units: 0.100E+01 -4: -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: *** | | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: 880 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 | -4: 754 | 0 0 0 | -4: | 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 628 | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 | -4: 502 | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 376 | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 250 | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 124 | 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 | -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: -4: -4: Interpolating fields .... -4: -4: Output group 1 -4: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -4: Output group 2 -4: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -4: Output group 3 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 4 -4: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -4: Output group 5 -4: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -4: Output group 6 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 7 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 8 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 9 -4: Output variables skipped -4: Output group 10 -4: Output variables skipped -4: ------------------------------------------------ -4: 1Current vel. -4: 1Wind speed -4: 1Ice concentration -4: 2Wave height -4: 2Mean wave period(+2) -4: 2Mean wave period(+1) -4: 2Peak frequency -4: 2Mean wave dir. a1b1 -4: 2Peak direction -4: 4Part. wave height -4: 4Part. peak period -4: 4Part. mean direction -4: 5Charnock parameter -4: ------------------------------------------------ -4: OUTPUT TIME : 2021/03/26 03:00:00 UTC -4: -4: End of file reached -4: -4: -4: *** End of Grid interpolation Routine *** -4: =============================================== -4: -4: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -4: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -4: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_aoc_9km/out_grd.aoc_9km /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.aoc_9km -4: + cmdfile.5[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032603 63 arctic 9km 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + cmdfile.5[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib2_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grib2_sbs.sh[30]grdID=aoc_9km -4: + wave_grib2_sbs.sh[31]GRIDNR=255 -4: + wave_grib2_sbs.sh[32]MODNR=11 -4: + wave_grib2_sbs.sh[33]valid_time=2021032603 -4: + wave_grib2_sbs.sh[34]fhr=63 -4: + wave_grib2_sbs.sh[35]grid_region=arctic -4: + wave_grib2_sbs.sh[36]grid_res=9km -4: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463 -4: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_aoc_9km -4: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_aoc_9km -4: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_aoc_9km -4: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_aoc_9km -4: ++ wave_grib2_sbs.sh[47]printf %03i 63 -4: + wave_grib2_sbs.sh[47]FH3=063 -4: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_arctic_9km -4: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.arctic.9km.f063.grib2 -4: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f063.grib2 ]] -4: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mod_def.aoc_9km ./mod_def.ww3 -4: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.aoc_9km ./out_grd.ww3 -4: + wave_grib2_sbs.sh[73]ngrib=1 -4: + wave_grib2_sbs.sh[74]dtgrib=3600 -4: + wave_grib2_sbs.sh[75]tstart='20210326 030000' -4: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210326 030000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.aoc_9km.inp.tmpl -4: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -4: $ WAVEWATCH-III gridded output input file -4: $ ---------------------------------------- -4: 20210326 030000 3600 1 -4: N -4: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -4: $ -4: 20210326 030000 7 11 255 0 20 -4: $ -4: 70 0 9.0 9.0 64 -4: $ 60 0 8.64919046313 8.64919046313 64 -4: $ end of input file -4: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[88]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.460139 ']' -4: ++ prep_step[4]echo gfs_ww3_grib.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[90]export err=0 -4: + wave_grib2_sbs.sh[90]err=0 -4: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[95]cat grib2_arctic_063.out -4: -4: *** WAVEWATCH III GRIB output postp. *** -4: ============================================== -4: -4: Comment character is '$' -4: -4: Grid name : Arctic Ocean PolarStereo 9km -4: -4: LINEIN: -4: 20210326 030000 3600 1 -4: -4: 20210326030000 3600 1 -4: GEN_PRO -99999 -4: -4: Output time data : -4: ----------------------------------------------------- -4: First time : 2021/03/26 03:00:00 UTC -4: Interval : 01:00:00 -4: Number of requests : 1 -4: Fields : Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: Requested output fields not yet available: -4: ----------------------------------------------------- -4: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -4: -4: Successfully requested output fields : -4: ----------------------------------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: -4: Additional GRIB parameters : -4: ----------------------------------------------------- -4: Run time : 2021/03/26 03:00:00 UTC -4: GRIB center ID : 7 -4: GRIB gen. proc. ID : 11 -4: GRIB grid ID : 255 -4: GRIB GDS parameter : 0 -4: Fields in file : -4: -------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: CHOSEN GRID TYPE: : POLARSTEREO -4: -4: -4: -4: Generating file -4: ----------------------------------------------------- -4: Data for 2021/03/26 03:00:00 UTC 0H forecast. -4: -4: End of program -4: ========================================= -4: WAVEWATCH III GRIB output -4: -4: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -4: + wave_grib2_sbs.sh[102][[ 63 -gt 0 ]] -4: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '63 hour fcst' -grib gfs.wave.t12z.arctic.9km.f063.grib2 -4: 1:0:d=2021032312:SPC:surface:63 hour fcst: -4: 2:192171:d=2021032312:DIRC:surface:63 hour fcst: -4: 3:659248:d=2021032312:UOGRD:surface:63 hour fcst: -4: 4:849424:d=2021032312:VOGRD:surface:63 hour fcst: -4: 5:1043435:d=2021032312:WIND:surface:63 hour fcst: -4: 6:1344814:d=2021032312:WDIR:surface:63 hour fcst: -4: 7:1771651:d=2021032312:UGRD:surface:63 hour fcst: -4: 8:2068735:d=2021032312:VGRD:surface:63 hour fcst: -4: 9:2370395:d=2021032312:ICEC:surface:63 hour fcst: -4: 10:2541025:d=2021032312:HTSGW:surface:63 hour fcst: -4: 11:2751612:d=2021032312:IMWF:surface:63 hour fcst: -4: 12:2966106:d=2021032312:MWSPER:surface:63 hour fcst: -4: 13:3182502:d=2021032312:PERPW:surface:63 hour fcst: -4: 14:3407627:d=2021032312:WWSDIR:surface:63 hour fcst: -4: 15:3708271:d=2021032312:DIRPW:surface:63 hour fcst: -4: 16:4014980:d=2021032312:WVHGT:surface:63 hour fcst: -4: 17:4222395:d=2021032312:SWELL:1 in sequence:63 hour fcst: -4: 18:4414184:d=2021032312:SWELL:2 in sequence:63 hour fcst: -4: 19:4566045:d=2021032312:SWELL:3 in sequence:63 hour fcst: -4: 20:4694273:d=2021032312:WVPER:surface:63 hour fcst: -4: 21:4911332:d=2021032312:SWPER:1 in sequence:63 hour fcst: -4: 22:5111461:d=2021032312:SWPER:2 in sequence:63 hour fcst: -4: 23:5270457:d=2021032312:SWPER:3 in sequence:63 hour fcst: -4: 24:5399284:d=2021032312:WVDIR:surface:63 hour fcst: -4: 25:5679626:d=2021032312:SWDIR:1 in sequence:63 hour fcst: -4: 26:5950466:d=2021032312:SWDIR:2 in sequence:63 hour fcst: -4: 27:6138838:d=2021032312:SWDIR:3 in sequence:63 hour fcst: -4: + wave_grib2_sbs.sh[104]err=0 -4: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.arctic.9km.f063.grib2 -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f063.grib2 ]] -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f063.grib2.idx ]] -4: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.arctic.9km.f063.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f063.grib2 -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f063.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f063.grib2 = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f063.grib2 ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f063.grib2 -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f063.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f063.grib2.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f063.grib2.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f063.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f063.grib2 -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.arctic.9km.f063.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f063.grib2.idx -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f063.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f063.grib2.idx = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f063.grib2.idx ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f063.grib2.idx -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f063.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f063.grib2.idx.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f063.grib2.idx.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f063.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f063.grib2.idx -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.arctic.9km.f063.grib2 and gfs.wave.t12z.arctic.9km.f063.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_aoc_9km to COM' -4: INFO: Copied gfs.wave.t12z.arctic.9km.f063.grib2 and gfs.wave.t12z.arctic.9km.f063.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_aoc_9km to COM -4: + wave_grib2_sbs.sh[130][[ aoc_9km == '' ]] -4: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -4: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.arctic.9km.f063.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -4: INFO: gfs.wave.t12z.arctic.9km.f063.grib2 is global.0p50 or SENDDBN is NO, no alert sent -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + cmdfile.7[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032603 3600. 9999 -6: + cmdfile.7[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grid_interp_sbs.sh[25]grdID=gsh_15m -6: + wave_grid_interp_sbs.sh[26]valid_time=2021032603 -6: + wave_grid_interp_sbs.sh[27]dt=3600. -6: + wave_grid_interp_sbs.sh[28]nst=9999 -6: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463 -6: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.uglo_100km ./out_grd.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mod_def.uglo_100km ./mod_def.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mod_def.gsh_15m ./mod_def.gsh_15m -6: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m'\''' -6: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m' -6: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[51]weights_found=1 -6: + wave_grid_interp_sbs.sh[59]ymdhms='20210326 030000' -6: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210326 030000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gsh_15m/g ww3_gint.inp.tmpl -6: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -6: $ Input file for interpolation of uglo_100km to gsh_15m -6: $------------------------------------------------ -6: $ Start Time 3600. NSteps -6: 20210326 030000 3600. 9999 -6: $ Total number of grids -6: 2 -6: $ Grid extensions -6: 'uglo_100km' -6: 'gsh_15m' -6: $ -6: 0 -6: $ -6: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[70]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.460139 ']' -6: ++ prep_step[4]echo gfs_ww3_gint.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -6: INFO: Executing 'gfs_ww3_gint.x' -6: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[73]cat grid_interp.gsh_15m.out -6: -6: *** WAVEWATCH III Grid interpolation *** -6: =============================================== -6: -6: Comment character is '$' -6: -6: Time Information : -6: --------------------------------------------- -6: Starting Time : 2021/03/26 03:00:00 UTC -6: Interval (in sec) : 3600.00 -6: Number of requests : 9999 -6: --------------------------------------------- -6: Number of grids (including output grid) = 2 -6: -6: -6: Extension for grid 1 is --> uglo_100km -6: -6: Grid Particulars are : -6: Dimensions = 45166 1 -6: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Extension for grid 2 is --> gsh_15m -6: -6: Grid Particulars are : -6: Dimensions = 1440 277 -6: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -6: -6: -6: Preparing interpolation weights for output grid -6: Total number of wet points for interpolation 317192 -6: -6: -6: Variable: Grid Interpolation Map Units: 0.100E+01 -6: -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 277 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 241 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 205 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 169 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 133 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 97 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 61 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 | -6: 25 | 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 | -6: | | -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: -6: -6: Interpolating fields .... -6: -6: Output group 1 -6: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -6: Output group 2 -6: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -6: Output group 3 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 4 -6: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -6: Output group 5 -6: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -6: Output group 6 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 7 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 8 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 9 -6: Output variables skipped -6: Output group 10 -6: Output variables skipped -6: ------------------------------------------------ -6: 1Current vel. -6: 1Wind speed -6: 1Ice concentration -6: 2Wave height -6: 2Mean wave period(+2) -6: 2Mean wave period(+1) -6: 2Peak frequency -6: 2Mean wave dir. a1b1 -6: 2Peak direction -6: 4Part. wave height -6: 4Part. peak period -6: 4Part. mean direction -6: 5Charnock parameter -6: ------------------------------------------------ -6: OUTPUT TIME : 2021/03/26 03:00:00 UTC -6: -6: End of file reached -6: -6: -6: *** End of Grid interpolation Routine *** -6: =============================================== -6: -6: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -6: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -6: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grid_interp_gsh_15m/out_grd.gsh_15m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.gsh_15m -6: + cmdfile.7[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032603 63 gsouth 0p25 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + cmdfile.7[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib2_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grib2_sbs.sh[30]grdID=gsh_15m -6: + wave_grib2_sbs.sh[31]GRIDNR=255 -6: + wave_grib2_sbs.sh[32]MODNR=11 -6: + wave_grib2_sbs.sh[33]valid_time=2021032603 -6: + wave_grib2_sbs.sh[34]fhr=63 -6: + wave_grib2_sbs.sh[35]grid_region=gsouth -6: + wave_grib2_sbs.sh[36]grid_res=0p25 -6: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463 -6: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_gsh_15m -6: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_gsh_15m -6: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_gsh_15m -6: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_gsh_15m -6: ++ wave_grib2_sbs.sh[47]printf %03i 63 -6: + wave_grib2_sbs.sh[47]FH3=063 -6: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -6: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.gsouth.0p25.f063.grib2 -6: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f063.grib2 ]] -6: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/mod_def.gsh_15m ./mod_def.ww3 -6: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/out_grd.gsh_15m ./out_grd.ww3 -6: + wave_grib2_sbs.sh[73]ngrib=1 -6: + wave_grib2_sbs.sh[74]dtgrib=3600 -6: + wave_grib2_sbs.sh[75]tstart='20210326 030000' -6: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210326 030000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gsh_15m.inp.tmpl -6: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -6: $ WAVEWATCH-III gridded output input file -6: $ ---------------------------------------- -6: 20210326 030000 3600 1 -6: N -6: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -6: $ -6: 20210326 030000 7 11 255 0 0 -6: $ -6: $ end of input file -6: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[88]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.460139 ']' -6: ++ prep_step[4]echo gfs_ww3_grib.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[90]export err=0 -6: + wave_grib2_sbs.sh[90]err=0 -6: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[95]cat grib2_gsouth_063.out -6: -6: *** WAVEWATCH III GRIB output postp. *** -6: ============================================== -6: -6: Comment character is '$' -6: -6: Grid name : GFSv16-wave S Hemisphere 1/4 d -6: -6: LINEIN: -6: 20210326 030000 3600 1 -6: -6: 20210326030000 3600 1 -6: GEN_PRO -99999 -6: -6: Output time data : -6: ----------------------------------------------------- -6: First time : 2021/03/26 03:00:00 UTC -6: Interval : 01:00:00 -6: Number of requests : 1 -6: Fields : Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: Requested output fields not yet available: -6: ----------------------------------------------------- -6: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -6: -6: Successfully requested output fields : -6: ----------------------------------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: -6: Additional GRIB parameters : -6: ----------------------------------------------------- -6: Run time : 2021/03/26 03:00:00 UTC -6: GRIB center ID : 7 -6: GRIB gen. proc. ID : 11 -6: GRIB grid ID : 255 -6: GRIB GDS parameter : 0 -6: Fields in file : -6: -------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: CHOSEN GRID TYPE: : LLRECTILINEAR -6: -6: -6: -6: Generating file -6: ----------------------------------------------------- -6: Data for 2021/03/26 03:00:00 UTC 0H forecast. -6: -6: End of program -6: ========================================= -6: WAVEWATCH III GRIB output -6: -6: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -6: + wave_grib2_sbs.sh[102][[ 63 -gt 0 ]] -6: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '63 hour fcst' -grib gfs.wave.t12z.gsouth.0p25.f063.grib2 -6: 1:0:d=2021032312:SPC:surface:63 hour fcst: -6: 2:117594:d=2021032312:DIRC:surface:63 hour fcst: -6: 3:497381:d=2021032312:UOGRD:surface:63 hour fcst: -6: 4:610488:d=2021032312:VOGRD:surface:63 hour fcst: -6: 5:731351:d=2021032312:WIND:surface:63 hour fcst: -6: 6:975879:d=2021032312:WDIR:surface:63 hour fcst: -6: 7:1353575:d=2021032312:UGRD:surface:63 hour fcst: -6: 8:1591531:d=2021032312:VGRD:surface:63 hour fcst: -6: 9:1833223:d=2021032312:ICEC:surface:63 hour fcst: -6: 10:1894356:d=2021032312:HTSGW:surface:63 hour fcst: -6: 11:2051650:d=2021032312:IMWF:surface:63 hour fcst: -6: 12:2221015:d=2021032312:MWSPER:surface:63 hour fcst: -6: 13:2391452:d=2021032312:PERPW:surface:63 hour fcst: -6: 14:2580933:d=2021032312:WWSDIR:surface:63 hour fcst: -6: 15:2921440:d=2021032312:DIRPW:surface:63 hour fcst: -6: 16:3277806:d=2021032312:WVHGT:surface:63 hour fcst: -6: 17:3440469:d=2021032312:SWELL:1 in sequence:63 hour fcst: -6: 18:3605052:d=2021032312:SWELL:2 in sequence:63 hour fcst: -6: 19:3715257:d=2021032312:SWELL:3 in sequence:63 hour fcst: -6: 20:3780521:d=2021032312:WVPER:surface:63 hour fcst: -6: 21:3963309:d=2021032312:SWPER:1 in sequence:63 hour fcst: -6: 22:4148570:d=2021032312:SWPER:2 in sequence:63 hour fcst: -6: 23:4285488:d=2021032312:SWPER:3 in sequence:63 hour fcst: -6: 24:4361230:d=2021032312:WVDIR:surface:63 hour fcst: -6: 25:4654638:d=2021032312:SWDIR:1 in sequence:63 hour fcst: -6: 26:4988238:d=2021032312:SWDIR:2 in sequence:63 hour fcst: -6: 27:5215704:d=2021032312:SWDIR:3 in sequence:63 hour fcst: -6: + wave_grib2_sbs.sh[104]err=0 -6: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.gsouth.0p25.f063.grib2 -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f063.grib2 ]] -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f063.grib2.idx ]] -6: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.gsouth.0p25.f063.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f063.grib2 -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f063.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f063.grib2 = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f063.grib2 ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f063.grib2 -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f063.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f063.grib2.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f063.grib2.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f063.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f063.grib2 -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.gsouth.0p25.f063.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f063.grib2.idx -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f063.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f063.grib2.idx = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f063.grib2.idx ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f063.grib2.idx -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f063.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f063.grib2.idx.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f063.grib2.idx.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f063.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f063.grib2.idx -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.gsouth.0p25.f063.grib2 and gfs.wave.t12z.gsouth.0p25.f063.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_gsh_15m to COM' -6: INFO: Copied gfs.wave.t12z.gsouth.0p25.f063.grib2 and gfs.wave.t12z.gsouth.0p25.f063.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463/grib_gsh_15m to COM -6: + wave_grib2_sbs.sh[130][[ gsh_15m == '' ]] -6: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -6: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.gsouth.0p25.f063.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -6: INFO: gfs.wave.t12z.gsouth.0p25.f063.grib2 is global.0p50 or SENDDBN is NO, no alert sent -+ run_mpmd.sh[113]exit 0 -+ run_mpmd.sh[1]postamble run_mpmd.sh 1753758324 0 -+ preamble.sh[62]set +x -End run_mpmd.sh at 03:05:34 with error code 0 (time elapsed: 00:00:10) -+ exgfs_wave_post_gridded_sbs.sh[122]true -+ exgfs_wave_post_gridded_sbs.sh[123]export err=0 -+ exgfs_wave_post_gridded_sbs.sh[123]err=0 -+ exgfs_wave_post_gridded_sbs.sh[124][[ 0 -ne 0 ]] -+ exgfs_wave_post_gridded_sbs.sh[130]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ exgfs_wave_post_gridded_sbs.sh[131]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ exgfs_wave_post_gridded_sbs.sh[132]gribchk=gfs.wave.t12z.gsouth.0p25.f063.grib2 -+ exgfs_wave_post_gridded_sbs.sh[133][[ ! -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f063.grib2 ]] -+ exgfs_wave_post_gridded_sbs.sh[138]exit 0 -+ JGLOBAL_WAVE_POST_SBS[28]true -+ JGLOBAL_WAVE_POST_SBS[29]export err=0 -+ JGLOBAL_WAVE_POST_SBS[29]err=0 -+ JGLOBAL_WAVE_POST_SBS[30][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[37]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312 -+ JGLOBAL_WAVE_POST_SBS[38][[ NO != \Y\E\S ]] -+ JGLOBAL_WAVE_POST_SBS[39]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f063.453463 -+ JGLOBAL_WAVE_POST_SBS[42]exit 0 -+ JGLOBAL_WAVE_POST_SBS[1]postamble /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS 1753758312 0 -+ preamble.sh[62]set +x -End /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at 03:05:35 with error code 0 (time elapsed: 00:00:23) diff --git a/ci/error_logs/HECULES_PR_LOGS/C48_S2SW_logs_2021032312_gfs_wavepostsbs_f093-f099.log b/ci/error_logs/HECULES_PR_LOGS/C48_S2SW_logs_2021032312_gfs_wavepostsbs_f093-f099.log deleted file mode 100644 index 5d325f51..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/C48_S2SW_logs_2021032312_gfs_wavepostsbs_f093-f099.log +++ /dev/null @@ -1,12866 +0,0 @@ -+ source /work2/noaa/global/mterry/global-workflow_forked/ush/load_ufswm_modules.sh -++ [[ NO == \N\O ]] -++ echo 'Loading modules quietly...' -Loading modules quietly... -++ set +x -Running "module reset". Resetting modules to system default. The following $MODULEPATH directories have been removed: None - -Currently Loaded Modules: - 1) contrib/0.1 25) git-lfs/3.1.2 - 2) intel-oneapi-compilers/2023.1.0 26) crtm/2.4.0.1 - 3) stack-intel/2021.9.0 27) g2/3.5.1 - 4) intel-oneapi-mpi/2021.9.0 28) g2tmpl/1.13.0 - 5) stack-intel-oneapi-mpi/2021.9.0 29) ip/4.3.0 - 6) nghttp2/1.57.0 30) sp/2.5.0 - 7) curl/8.4.0 31) w3emc/2.10.0 - 8) cmake/3.23.1 32) gftl/1.10.0 - 9) libjpeg/2.1.0 33) gftl-shared/1.6.1 - 10) jasper/2.0.32 34) fargparse/1.5.0 - 11) zlib/1.2.13 35) pigz/2.7 - 12) libpng/1.6.37 36) tar/1.34 - 13) hdf5/1.14.0 37) gettext/0.21.1 - 14) snappy/1.1.10 38) libxcrypt/4.4.35 - 15) zstd/1.5.2 39) sqlite/3.43.2 - 16) c-blosc/1.21.5 40) util-linux-uuid/2.38.1 - 17) netcdf-c/4.9.2 41) python/3.10.13 - 18) netcdf-fortran/4.6.1 42) mapl/2.40.3-esmf-8.6.0 - 19) parallel-netcdf/1.12.2 43) scotch/7.0.4 - 20) parallelio/2.5.10 44) ufs_common - 21) esmf/8.6.0 45) nccmp/1.9.0.1 - 22) fms/2024.01.02 46) ufs_hercules.intel - 23) bacio/2.4.1 47) prod_util/2.1.1 - 24) crtm-fix/2.4.0.1_emc 48) wgrib2/3.1.1 - - - -Begin /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at Tue Jul 29 03:04:26 UTC 2025 -++ jjob_header.sh[46]OPTIND=1 -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[50]env_job=wavepostsbs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[49]read -ra configs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[61]shift 4 -++ jjob_header.sh[63][[ -z wavepostsbs ]] -++ jjob_header.sh[71]export DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676 -++ jjob_header.sh[71]DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676 -++ jjob_header.sh[72][[ YES == \Y\E\S ]] -++ jjob_header.sh[73]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676 -++ jjob_header.sh[75]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676 -++ jjob_header.sh[76]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676 -++ jjob_header.sh[85]export pid=315371 -++ jjob_header.sh[85]pid=315371 -++ jjob_header.sh[86]export pgmout=OUTPUT.315371 -++ jjob_header.sh[86]pgmout=OUTPUT.315371 -++ jjob_header.sh[87]export pgmerr=errfile -++ jjob_header.sh[87]pgmerr=errfile -++ jjob_header.sh[90]export pgm= -++ jjob_header.sh[90]pgm= -++ jjob_header.sh[96]export cycle=t12z -++ jjob_header.sh[96]cycle=t12z -++ jjob_header.sh[97]setpdy.sh -+ setpdy.sh[20]'[' /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676 == /home/mterry ']' -+ setpdy.sh[25][[ ! t12z =~ t??z ]] -+ setpdy.sh[30]case $# in -+ setpdy.sh[31]dates_before_PDY=7 -+ setpdy.sh[32]dates_after_PDY=7 -+ setpdy.sh[50]COMDATEROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+ setpdy.sh[53]'[' -z 20210323 ']' -+ setpdy.sh[57]sed 's/[0-9]\{8\}/20210323/' /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z -sed: can't read /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z: No such file or directory -++ jjob_header.sh[97]true -++ jjob_header.sh[98]source ./PDY -/work2/noaa/global/mterry/global-workflow_forked/ush/jjob_header.sh: line 98: ./PDY: No such file or directory -++ jjob_header.sh[98]true -++ jjob_header.sh[104]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[104]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.base -+++ config.base[6]echo 'BEGIN: config.base' -BEGIN: config.base -+++ config.base[9]export machine=HERCULES -+++ config.base[9]machine=HERCULES -+++ config.base[12]export RUN_ENVIR=emc -+++ config.base[12]RUN_ENVIR=emc -+++ config.base[15]export ACCOUNT=fv3-cpu -+++ config.base[15]ACCOUNT=fv3-cpu -+++ config.base[16]export QUEUE=batch -+++ config.base[16]QUEUE=batch -+++ config.base[17]export QUEUE_SERVICE=batch -+++ config.base[17]QUEUE_SERVICE=batch -+++ config.base[18]export QUEUE_DTN=batch -+++ config.base[18]QUEUE_DTN=batch -+++ config.base[19]export PARTITION_BATCH=hercules -+++ config.base[19]PARTITION_BATCH=hercules -+++ config.base[20]export PARTITION_SERVICE=service -+++ config.base[20]PARTITION_SERVICE=service -+++ config.base[21]export PARTITION_DTN= -+++ config.base[21]PARTITION_DTN= -+++ config.base[22]export RESERVATION= -+++ config.base[22]RESERVATION= -+++ config.base[23]export CLUSTERS= -+++ config.base[23]CLUSTERS= -+++ config.base[24]export CLUSTERS_SERVICE= -+++ config.base[24]CLUSTERS_SERVICE= -+++ config.base[25]export CLUSTERS_DTN= -+++ config.base[25]CLUSTERS_DTN= -+++ config.base[28]export HPSS_PROJECT=emc-global -+++ config.base[28]HPSS_PROJECT=emc-global -+++ config.base[31]export HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[31]HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[32]export EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[32]EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[33]export FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[33]FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[34]export PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[34]PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[35]export SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[35]SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[36]export USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[36]USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[38]export FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[38]FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[39]export FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[39]FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[40]export FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[40]FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[41]export FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[41]FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[42]export FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[42]FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[43]export FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[43]FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[44]export FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[44]FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[45]export FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[45]FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[50]export PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[50]PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[51]export COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[51]COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[52]export COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[52]COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[53]export DMPDIR=/work/noaa/rstprod/dump -+++ config.base[53]DMPDIR=/work/noaa/rstprod/dump -+++ config.base[57]export COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[57]COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[58]export COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[58]COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[59]export COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[59]COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[62]export HOMEDIR=/work2/noaa/global/mterry -+++ config.base[62]HOMEDIR=/work2/noaa/global/mterry -+++ config.base[63]export STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[63]STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]export PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[65]export NOSCRUB=/work2/noaa/global/mterry -+++ config.base[65]NOSCRUB=/work2/noaa/global/mterry -+++ config.base[68]export BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[68]BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[71]export BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[71]BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[74]export DO_PREP_SFC=NO -+++ config.base[74]DO_PREP_SFC=NO -+++ config.base[77]export DO_GOES=NO -+++ config.base[77]DO_GOES=NO -+++ config.base[78]export DO_BUFRSND=NO -+++ config.base[78]DO_BUFRSND=NO -+++ config.base[79]export DO_GEMPAK=NO -+++ config.base[79]DO_GEMPAK=NO -+++ config.base[80]export DO_AWIPS=NO -+++ config.base[80]DO_AWIPS=NO -+++ config.base[81]export DO_NPOESS=NO -+++ config.base[81]DO_NPOESS=NO -+++ config.base[82]export DO_TRACKER=YES -+++ config.base[82]DO_TRACKER=YES -+++ config.base[83]export DO_GENESIS=YES -+++ config.base[83]DO_GENESIS=YES -+++ config.base[84]export DO_GENESIS_FSU=NO -+++ config.base[84]DO_GENESIS_FSU=NO -+++ config.base[85]export DO_VERFOZN=YES -+++ config.base[85]DO_VERFOZN=YES -+++ config.base[86]export DO_VERFRAD=YES -+++ config.base[86]DO_VERFRAD=YES -+++ config.base[87]export DO_VMINMON=YES -+++ config.base[87]DO_VMINMON=YES -+++ config.base[88]export DO_ANLSTAT=NO -+++ config.base[88]DO_ANLSTAT=NO -+++ config.base[91]export MODE=forecast-only -+++ config.base[91]MODE=forecast-only -+++ config.base[92]export DO_TEST_MODE=YES -+++ config.base[92]DO_TEST_MODE=YES -+++ config.base[101]export FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[101]FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[102]export HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[102]HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[103]export HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[103]HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[106]export NMV=/bin/mv -+++ config.base[106]NMV=/bin/mv -+++ config.base[107]export 'NLN=/bin/ln -sf' -+++ config.base[107]NLN='/bin/ln -sf' -+++ config.base[108]export VERBOSE=YES -+++ config.base[108]VERBOSE=YES -+++ config.base[109]export KEEPDATA=NO -+++ config.base[109]KEEPDATA=NO -+++ config.base[110]export DEBUG_POSTSCRIPT=NO -+++ config.base[110]DEBUG_POSTSCRIPT=NO -+++ config.base[111]export CHGRP_RSTPROD=YES -+++ config.base[111]CHGRP_RSTPROD=YES -+++ config.base[112]export 'CHGRP_CMD=chgrp rstprod' -+++ config.base[112]CHGRP_CMD='chgrp rstprod' -+++ config.base[113]export NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[113]NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[114]export NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[114]NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[117]export BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[117]BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[120]export SDATE=2021032312 -+++ config.base[120]SDATE=2021032312 -+++ config.base[121]export EDATE=2021032312 -+++ config.base[121]EDATE=2021032312 -+++ config.base[122]export EXP_WARM_START=.false. -+++ config.base[122]EXP_WARM_START=.false. -+++ config.base[123]export assim_freq=6 -+++ config.base[123]assim_freq=6 -+++ config.base[124]export PSLOT=C48_S2SW -+++ config.base[124]PSLOT=C48_S2SW -+++ config.base[125]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[125]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[126]export ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[126]ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[127]export DUMP_SUFFIX= -+++ config.base[127]DUMP_SUFFIX= -+++ config.base[128][[ 2021032312 -ge 2019092100 ]] -+++ config.base[128][[ 2021032312 -le 2019110700 ]] -+++ config.base[131]export ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[131]ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[132]export ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[132]ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[133]export FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[133]FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[136]export envir=prod -+++ config.base[136]envir=prod -+++ config.base[137]export NET=gfs -+++ config.base[137]NET=gfs -+++ config.base[138]export RUN=gfs -+++ config.base[138]RUN=gfs -+++ config.base[141]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.com -++++ config.com[4]echo 'BEGIN: config.com' -BEGIN: config.com -++++ config.com[38][[ emc == \n\c\o ]] -++++ config.com[43]COM_OBSPROC_TMPL='${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos' -++++ config.com[44]COM_RTOFS_TMPL='${DMPDIR}' -++++ config.com[45]COM_TCVITAL_TMPL='${DMPDIR}/${RUN}.${YMD}/${HH}/atmos' -++++ config.com[47]declare -rx 'COM_OBS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/obs' -++++ config.com[48]declare -rx COM_OBSPROC_TMPL COM_RTOFS_TMPL -++++ config.com[50]COM_BASE='${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}' -++++ config.com[52]declare -rx 'COM_TOP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}' -++++ config.com[54]declare -rx 'COM_CONF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf' -++++ config.com[55]declare -rx 'COM_OBS_JEDI=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi' -++++ config.com[57]declare -rx 'COM_ATMOS_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input' -++++ config.com[58]declare -rx 'COM_ATMOS_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart' -++++ config.com[59]declare -rx 'COM_ATMOS_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos' -++++ config.com[60]declare -rx 'COM_SNOW_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow' -++++ config.com[61]declare -rx 'COM_SNOW_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon' -++++ config.com[62]declare -rx 'COM_ATMOS_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history' -++++ config.com[63]declare -rx 'COM_ATMOS_MASTER_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master' -++++ config.com[64]declare -rx 'COM_ATMOS_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2' -++++ config.com[65]declare -rx 'COM_ATMOS_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}' -++++ config.com[66]declare -rx 'COM_ATMOS_BUFR_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr' -++++ config.com[67]declare -rx 'COM_ATMOS_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}' -++++ config.com[68]declare -rx 'COM_ATMOS_GENESIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital' -++++ config.com[69]declare -rx 'COM_ATMOS_TRACK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks' -++++ config.com[70]declare -rx 'COM_ATMOS_GOES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim' -++++ config.com[71]declare -rx 'COM_ATMOS_IMAGERY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery' -++++ config.com[72]declare -rx 'COM_ATMOS_OZNMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon' -++++ config.com[73]declare -rx 'COM_ATMOS_RADMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon' -++++ config.com[74]declare -rx 'COM_ATMOS_MINMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon' -++++ config.com[75]declare -rx 'COM_ATMOS_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon' -++++ config.com[76]declare -rx 'COM_ATMOS_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo' -++++ config.com[78]declare -rx 'COM_WAVE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart' -++++ config.com[79]declare -rx 'COM_WAVE_PREP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep' -++++ config.com[80]declare -rx 'COM_WAVE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history' -++++ config.com[81]declare -rx 'COM_WAVE_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded' -++++ config.com[82]declare -rx 'COM_WAVE_GRID_RES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}' -++++ config.com[83]declare -rx 'COM_WAVE_STATION_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station' -++++ config.com[84]declare -rx 'COM_WAVE_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak' -++++ config.com[85]declare -rx 'COM_WAVE_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo' -++++ config.com[87]declare -rx 'COM_OCEAN_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history' -++++ config.com[88]declare -rx 'COM_OCEAN_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart' -++++ config.com[89]declare -rx 'COM_OCEAN_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input' -++++ config.com[90]declare -rx 'COM_OCEAN_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean' -++++ config.com[91]declare -rx 'COM_OCEAN_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon' -++++ config.com[92]declare -rx 'COM_OCEAN_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf' -++++ config.com[93]declare -rx 'COM_OCEAN_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean' -++++ config.com[94]declare -rx 'COM_OCEAN_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf' -++++ config.com[95]declare -rx 'COM_OCEAN_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2' -++++ config.com[96]declare -rx 'COM_OCEAN_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}' -++++ config.com[98]declare -rx 'COM_ICE_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice' -++++ config.com[99]declare -rx 'COM_ICE_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf' -++++ config.com[100]declare -rx 'COM_ICE_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon' -++++ config.com[101]declare -rx 'COM_ICE_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice' -++++ config.com[102]declare -rx 'COM_ICE_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input' -++++ config.com[103]declare -rx 'COM_ICE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history' -++++ config.com[104]declare -rx 'COM_ICE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart' -++++ config.com[105]declare -rx 'COM_ICE_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf' -++++ config.com[106]declare -rx 'COM_ICE_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2' -++++ config.com[107]declare -rx 'COM_ICE_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}' -++++ config.com[109]declare -rx 'COM_CHEM_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history' -++++ config.com[110]declare -rx 'COM_CHEM_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem' -++++ config.com[111]declare -rx 'COM_CHEM_BMAT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix' -++++ config.com[112]declare -rx 'COM_CHEM_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon' -++++ config.com[114]declare -rx 'COM_MED_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart' -+++ config.base[143]export LOGSCRIPT= -+++ config.base[143]LOGSCRIPT= -+++ config.base[145]export 'REDOUT=1>' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wave -+++ config.wave[6]echo 'BEGIN: config.wave' -BEGIN: config.wave -+++ config.wave[13]export RUNRSTwave=gdas -+++ config.wave[13]RUNRSTwave=gdas -+++ config.wave[16]export MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[16]MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[19]case "${waveGRD}" in -+++ config.wave[64]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[64]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[65]export wavepostGRD= -+++ config.wave[65]wavepostGRD= -+++ config.wave[66]export waveuoutpGRD=uglo_100km -+++ config.wave[66]waveuoutpGRD=uglo_100km -+++ config.wave[75]export WAVEWND_DID= -+++ config.wave[75]WAVEWND_DID= -+++ config.wave[76]export WAVEWND_FID= -+++ config.wave[76]WAVEWND_FID= -+++ config.wave[79][[ gfs == \g\f\s ]] -+++ config.wave[80]export FHMAX_WAV=120 -+++ config.wave[80]FHMAX_WAV=120 -+++ config.wave[82]export WAVHINDH=0 -+++ config.wave[82]WAVHINDH=0 -+++ config.wave[83]export FHMAX_WAV_IBP=180 -+++ config.wave[83]FHMAX_WAV_IBP=180 -+++ config.wave[84](( FHMAX_WAV < FHMAX_WAV_IBP )) -+++ config.wave[84]export FHMAX_WAV_IBP=120 -+++ config.wave[84]FHMAX_WAV_IBP=120 -+++ config.wave[87]export DTFLD_WAV=3600 -+++ config.wave[87]DTFLD_WAV=3600 -+++ config.wave[88]export DTPNT_WAV=3600 -+++ config.wave[88]DTPNT_WAV=3600 -+++ config.wave[89]export FHINCP_WAV=1 -+++ config.wave[89]FHINCP_WAV=1 -+++ config.wave[92]export 'OUTPARS_WAV=WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[92]OUTPARS_WAV='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[95][[ gfs == \g\d\a\s ]] -+++ config.wave[99](( INTERVAL_GFS > 0 )) -+++ config.wave[100]export WAVHCYC=6 -+++ config.wave[100]WAVHCYC=6 -+++ config.wave[101]export FHMAX_WAV_CUR=192 -+++ config.wave[101]FHMAX_WAV_CUR=192 -+++ config.wave[109]export RSTTYPE_WAV=T -+++ config.wave[109]RSTTYPE_WAV=T -+++ config.wave[110][[ gfs != gfs ]] -+++ config.wave[115]rst_dt_gfs=43200 -+++ config.wave[116][[ 43200 -gt 0 ]] -+++ config.wave[117]export DT_1_RST_WAV=0 -+++ config.wave[117]DT_1_RST_WAV=0 -+++ config.wave[120]export DT_2_RST_WAV=43200 -+++ config.wave[120]DT_2_RST_WAV=43200 -+++ config.wave[126]export RSTIOFF_WAV=0 -+++ config.wave[126]RSTIOFF_WAV=0 -+++ config.wave[131]export RUNMEM=-1 -+++ config.wave[131]RUNMEM=-1 -+++ config.wave[134](( RUNMEM == -1 )) -+++ config.wave[136]export waveMEMB= -+++ config.wave[136]waveMEMB= -+++ config.wave[143]export WW3ATMINP=CPL -+++ config.wave[143]WW3ATMINP=CPL -+++ config.wave[144][[ YES == \Y\E\S ]] -+++ config.wave[145]export WW3ICEINP=CPL -+++ config.wave[145]WW3ICEINP=CPL -+++ config.wave[146]export WAVEICE_FID= -+++ config.wave[146]WAVEICE_FID= -+++ config.wave[152][[ YES == \Y\E\S ]] -+++ config.wave[153]export WW3CURINP=CPL -+++ config.wave[153]WW3CURINP=CPL -+++ config.wave[154]export WAVECUR_FID= -+++ config.wave[154]WAVECUR_FID= -+++ config.wave[161]export WW3ATMIENS=F -+++ config.wave[161]WW3ATMIENS=F -+++ config.wave[162]export WW3ICEIENS=F -+++ config.wave[162]WW3ICEIENS=F -+++ config.wave[163]export WW3CURIENS=F -+++ config.wave[163]WW3CURIENS=F -+++ config.wave[165]export GOFILETYPE=1 -+++ config.wave[165]GOFILETYPE=1 -+++ config.wave[166]export POFILETYPE=1 -+++ config.wave[166]POFILETYPE=1 -+++ config.wave[170]export FUNIPNT=T -+++ config.wave[170]FUNIPNT=T -+++ config.wave[172]export IOSRV=1 -+++ config.wave[172]IOSRV=1 -+++ config.wave[174]export FPNTPROC=T -+++ config.wave[174]FPNTPROC=T -+++ config.wave[176]export FGRDPROC=F -+++ config.wave[176]FGRDPROC=F -+++ config.wave[178]export FLAGMASKCOMP=F -+++ config.wave[178]FLAGMASKCOMP=F -+++ config.wave[180]export FLAGMASKOUT=F -+++ config.wave[180]FLAGMASKOUT=F -+++ config.wave[182]echo 'END: config.wave' -END: config.wave -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wavepostsbs -+++ config.wavepostsbs[6]echo 'BEGIN: config.wavepostsbs' -BEGIN: config.wavepostsbs -+++ config.wavepostsbs[9]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources wavepostsbs -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=wavepostsbs -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[156]ntasks=8 -++++ config.resources[157]threads_per_task=1 -++++ config.resources[158]tasks_per_node=80 -++++ config.resources[159]NTASKS=8 -++++ config.resources[160]memory=20GB -++++ config.resources[161]walltime=00:15:00 -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export NTASKS -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export memory -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.wavepostsbs[12]export MAX_TASKS=25 -+++ config.wavepostsbs[12]MAX_TASKS=25 -+++ config.wavepostsbs[15]export WAV_SUBGRBSRC= -+++ config.wavepostsbs[15]WAV_SUBGRBSRC= -+++ config.wavepostsbs[16]export WAV_SUBGRB= -+++ config.wavepostsbs[16]WAV_SUBGRB= -+++ config.wavepostsbs[19]export DOFLD_WAV=YES -+++ config.wavepostsbs[19]DOFLD_WAV=YES -+++ config.wavepostsbs[20]export DOPNT_WAV=YES -+++ config.wavepostsbs[20]DOPNT_WAV=YES -+++ config.wavepostsbs[21]export DOGRB_WAV=YES -+++ config.wavepostsbs[21]DOGRB_WAV=YES -+++ config.wavepostsbs[22][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+++ config.wavepostsbs[23]export DOGRI_WAV=YES -+++ config.wavepostsbs[23]DOGRI_WAV=YES -+++ config.wavepostsbs[27]export DOSPC_WAV=YES -+++ config.wavepostsbs[27]DOSPC_WAV=YES -+++ config.wavepostsbs[28]export DOBLL_WAV=YES -+++ config.wavepostsbs[28]DOBLL_WAV=YES -+++ config.wavepostsbs[30]echo 'END: config.wavepostsbs' -END: config.wavepostsbs -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env wavepostsbs -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=wavepostsbs -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 20971520 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 8 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[34]max_threads_per_task=1 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 1 ]] -+++ HERCULES.env[40][[ 1 -gt 1 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 8' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[63]export USE_CFP=YES -+++ HERCULES.env[63]USE_CFP=YES -+++ HERCULES.env[64][[ wavepostsbs == \w\a\v\e\p\r\e\p ]] -+++ HERCULES.env[65]export 'wavempexec=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[65]wavempexec='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[66]export 'wave_mpmd=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[66]wave_mpmd='--multi-prog --output=mpmd.%j.%t.out' -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[4]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ JGLOBAL_WAVE_POST_SBS[7]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[7]HH=12 -+ JGLOBAL_WAVE_POST_SBS[7]declare_from_tmpl -rx COMIN_WAVE_PREP:COM_WAVE_PREP_TMPL COMIN_WAVE_HISTORY:COM_WAVE_HISTORY_TMPL COMOUT_WAVE_GRID:COM_WAVE_GRID_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_WAVE_PREP=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep -declare_from_tmpl :: COMIN_WAVE_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history -declare_from_tmpl :: COMOUT_WAVE_GRID=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded -+ JGLOBAL_WAVE_POST_SBS[12][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n '' ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_atlocn_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_atlocn_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_atlocn_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_epacif_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_epacif_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_epacif_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_wcoast_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_wcoast_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_wcoast_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p50 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p50:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p50=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p50 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=arctic.9km -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_arctic_9km:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_arctic_9km=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_arctic_9km -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_gsouth_0p25:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_gsouth_0p25=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 ]] -+ JGLOBAL_WAVE_POST_SBS[28]/work2/noaa/global/mterry/global-workflow_forked/scripts/exgfs_wave_post_gridded_sbs.sh -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ exgfs_wave_post_gridded_sbs.sh[24]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ exgfs_wave_post_gridded_sbs.sh[26]DOGRI_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[27]DOGRB_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[29]export waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[29]waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[30]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[30]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[31]export wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[31]wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[33]cat - INFO: Grid information: - INFO: Native wave grids: uglo_100km - INFO: Interpolated grids: at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m - INFO: Post-process grids: -++ exgfs_wave_post_gridded_sbs.sh[40]printf %03i 93 -+ exgfs_wave_post_gridded_sbs.sh[40]fhr3=093 -++ exgfs_wave_post_gridded_sbs.sh[41]date -u -d '20210323 12 + 93 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[41]valid_time=2021032709 -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[49]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f093.bin ./out_grd.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f093.bin ./out_grd.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[52][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[62]cat - INFO: Summary: - INFO: Grid interp: DOGRI_WAV="YES" - INFO: Grib files: DOGRB_WAV="YES" - INFO: Fields to be included in grib files: - INFO: OUTPARS_WAV="WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA" -+ exgfs_wave_post_gridded_sbs.sh[70][[ YES == \N\O ]] -+ exgfs_wave_post_gridded_sbs.sh[76]rm -f 'cmdfile.*' cmdfile -+ exgfs_wave_post_gridded_sbs.sh[77]count=0 -+ exgfs_wave_post_gridded_sbs.sh[80][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[81]dt_int=3600. -+ exgfs_wave_post_gridded_sbs.sh[82]n_int=9999 -++ exgfs_wave_post_gridded_sbs.sh[83]date -u -d '20210327 09 - 0 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[83]ymdh_int=2021032709 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=1 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032709 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032709 93 atlocn 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib2_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib2_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=2 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032709 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032709 93 epacif 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib2_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib2_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=3 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032709 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032709 93 wcoast 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib2_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib2_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=4 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032709 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032709 93 global 0p50 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib2_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib2_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=5 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032709 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032709 93 arctic 9km '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib2_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib2_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=6 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032709 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032709 93 global 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib2_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib2_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=7 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032709 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032709 93 gsouth 0p25 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib2_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib2_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[100][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[113][[ 8 -lt 7 ]] -+ exgfs_wave_post_gridded_sbs.sh[121]echo 'INFO: Running MPMD job with 7 commands' -INFO: Running MPMD job with 7 commands -+ exgfs_wave_post_gridded_sbs.sh[122]/work2/noaa/global/mterry/global-workflow_forked/ush/run_mpmd.sh /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/cmdfile -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ run_mpmd.sh[31]source /work2/noaa/global/mterry/global-workflow_forked/ush/preamble.sh -++ preamble.sh[20]set +x -Begin run_mpmd.sh at Tue Jul 29 03:04:38 UTC 2025 -+ run_mpmd.sh[33]cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/cmdfile -+ run_mpmd.sh[36][[ YES != \Y\E\S ]] -+ run_mpmd.sh[46]export OMP_NUM_THREADS=1 -+ run_mpmd.sh[46]OMP_NUM_THREADS=1 -++ run_mpmd.sh[49]wc -l -+ run_mpmd.sh[49]nprocs=7 -+ run_mpmd.sh[52]mpmd_cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mpmd_cmdfile -+ run_mpmd.sh[53][[ -s /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mpmd_cmdfile ]] -+ run_mpmd.sh[55]cat - INFO: Executing MPMD job, STDOUT redirected for each process separately - INFO: On failure, logs for each job will be available in /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mpmd.proc_num.out - INFO: The proc_num corresponds to the line in '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mpmd_cmdfile' -+ run_mpmd.sh[61][[ srun -l --export=ALL --hint=nomultithread =~ ^srun.* ]] -+ run_mpmd.sh[65]nm=0 -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '0 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/cmdfile.1' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '1 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/cmdfile.2' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '2 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/cmdfile.3' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '3 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/cmdfile.4' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '4 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/cmdfile.5' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '5 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/cmdfile.6' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '6 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/cmdfile.7' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[72]set +e -+ run_mpmd.sh[74]srun -l --export=ALL --hint=nomultithread --multi-prog --output=mpmd.%j.%t.out -n 7 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mpmd_cmdfile -+ run_mpmd.sh[75]err=0 -+ run_mpmd.sh[76]set_strict -+ preamble.sh[35][[ YES == \Y\E\S ]] -+ preamble.sh[37]set -eu -+ run_mpmd.sh[103][[ 0 -eq 0 ]] -+ run_mpmd.sh[104]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mpmd_cmdfile -++ run_mpmd.sh[105]find . -name 'mpmd.*.out' -+ run_mpmd.sh[105]out_files='./mpmd.5951798.4.out -./mpmd.5951798.2.out -./mpmd.5951798.1.out -./mpmd.5951798.3.out -./mpmd.5951798.5.out -./mpmd.5951798.6.out -./mpmd.5951798.0.out' -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951798.4.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951798.4.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951798.2.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951798.2.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951798.1.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951798.1.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951798.3.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951798.3.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951798.5.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951798.5.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951798.6.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951798.6.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951798.0.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951798.0.out -+ run_mpmd.sh[110]cat mpmd.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + cmdfile.5[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032709 3600. 9999 -4: + cmdfile.5[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grid_interp_sbs.sh[25]grdID=aoc_9km -4: + wave_grid_interp_sbs.sh[26]valid_time=2021032709 -4: + wave_grid_interp_sbs.sh[27]dt=3600. -4: + wave_grid_interp_sbs.sh[28]nst=9999 -4: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676 -4: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.uglo_100km ./out_grd.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mod_def.uglo_100km ./mod_def.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mod_def.aoc_9km ./mod_def.aoc_9km -4: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km'\''' -4: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km' -4: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[51]weights_found=1 -4: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 090000' -4: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 090000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/aoc_9km/g ww3_gint.inp.tmpl -4: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -4: $ Input file for interpolation of uglo_100km to aoc_9km -4: $------------------------------------------------ -4: $ Start Time 3600. NSteps -4: 20210327 090000 3600. 9999 -4: $ Total number of grids -4: 2 -4: $ Grid extensions -4: 'uglo_100km' -4: 'aoc_9km' -4: $ -4: 0 -4: $ -4: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[70]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.315371 ']' -4: ++ prep_step[4]echo gfs_ww3_gint.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -4: INFO: Executing 'gfs_ww3_gint.x' -4: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[73]cat grid_interp.aoc_9km.out -4: -4: *** WAVEWATCH III Grid interpolation *** -4: =============================================== -4: -4: Comment character is '$' -4: -4: Time Information : -4: --------------------------------------------- -4: Starting Time : 2021/03/27 09:00:00 UTC -4: Interval (in sec) : 3600.00 -4: Number of requests : 9999 -4: --------------------------------------------- -4: Number of grids (including output grid) = 2 -4: -4: -4: Extension for grid 1 is --> uglo_100km -4: -4: Grid Particulars are : -4: Dimensions = 45166 1 -4: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Extension for grid 2 is --> aoc_9km -4: -4: Grid Particulars are : -4: Dimensions = 1006 1006 -4: Grid Type = 2 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -4: -4: -4: Preparing interpolation weights for output grid -4: Total number of wet points for interpolation 360052 -4: -4: -4: Variable: Grid Interpolation Map Units: 0.100E+01 -4: -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: *** | | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: 880 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 | -4: 754 | 0 0 0 | -4: | 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 628 | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 | -4: 502 | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 376 | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 250 | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 124 | 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 | -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: -4: -4: Interpolating fields .... -4: -4: Output group 1 -4: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -4: Output group 2 -4: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -4: Output group 3 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 4 -4: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -4: Output group 5 -4: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -4: Output group 6 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 7 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 8 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 9 -4: Output variables skipped -4: Output group 10 -4: Output variables skipped -4: ------------------------------------------------ -4: 1Current vel. -4: 1Wind speed -4: 1Ice concentration -4: 2Wave height -4: 2Mean wave period(+2) -4: 2Mean wave period(+1) -4: 2Peak frequency -4: 2Mean wave dir. a1b1 -4: 2Peak direction -4: 4Part. wave height -4: 4Part. peak period -4: 4Part. mean direction -4: 5Charnock parameter -4: ------------------------------------------------ -4: OUTPUT TIME : 2021/03/27 09:00:00 UTC -4: -4: End of file reached -4: -4: -4: *** End of Grid interpolation Routine *** -4: =============================================== -4: -4: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -4: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -4: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_aoc_9km/out_grd.aoc_9km /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.aoc_9km -4: + cmdfile.5[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032709 93 arctic 9km 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + cmdfile.5[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib2_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grib2_sbs.sh[30]grdID=aoc_9km -4: + wave_grib2_sbs.sh[31]GRIDNR=255 -4: + wave_grib2_sbs.sh[32]MODNR=11 -4: + wave_grib2_sbs.sh[33]valid_time=2021032709 -4: + wave_grib2_sbs.sh[34]fhr=93 -4: + wave_grib2_sbs.sh[35]grid_region=arctic -4: + wave_grib2_sbs.sh[36]grid_res=9km -4: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676 -4: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_aoc_9km -4: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_aoc_9km -4: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_aoc_9km -4: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_aoc_9km -4: ++ wave_grib2_sbs.sh[47]printf %03i 93 -4: + wave_grib2_sbs.sh[47]FH3=093 -4: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_arctic_9km -4: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.arctic.9km.f093.grib2 -4: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f093.grib2 ]] -4: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mod_def.aoc_9km ./mod_def.ww3 -4: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.aoc_9km ./out_grd.ww3 -4: + wave_grib2_sbs.sh[73]ngrib=1 -4: + wave_grib2_sbs.sh[74]dtgrib=3600 -4: + wave_grib2_sbs.sh[75]tstart='20210327 090000' -4: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 090000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.aoc_9km.inp.tmpl -4: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -4: $ WAVEWATCH-III gridded output input file -4: $ ---------------------------------------- -4: 20210327 090000 3600 1 -4: N -4: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -4: $ -4: 20210327 090000 7 11 255 0 20 -4: $ -4: 70 0 9.0 9.0 64 -4: $ 60 0 8.64919046313 8.64919046313 64 -4: $ end of input file -4: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[88]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.315371 ']' -4: ++ prep_step[4]echo gfs_ww3_grib.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[90]export err=0 -4: + wave_grib2_sbs.sh[90]err=0 -4: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[95]cat grib2_arctic_093.out -4: -4: *** WAVEWATCH III GRIB output postp. *** -4: ============================================== -4: -4: Comment character is '$' -4: -4: Grid name : Arctic Ocean PolarStereo 9km -4: -4: LINEIN: -4: 20210327 090000 3600 1 -4: -4: 20210327090000 3600 1 -4: GEN_PRO -99999 -4: -4: Output time data : -4: ----------------------------------------------------- -4: First time : 2021/03/27 09:00:00 UTC -4: Interval : 01:00:00 -4: Number of requests : 1 -4: Fields : Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: Requested output fields not yet available: -4: ----------------------------------------------------- -4: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -4: -4: Successfully requested output fields : -4: ----------------------------------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: -4: Additional GRIB parameters : -4: ----------------------------------------------------- -4: Run time : 2021/03/27 09:00:00 UTC -4: GRIB center ID : 7 -4: GRIB gen. proc. ID : 11 -4: GRIB grid ID : 255 -4: GRIB GDS parameter : 0 -4: Fields in file : -4: -------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: CHOSEN GRID TYPE: : POLARSTEREO -4: -4: -4: -4: Generating file -4: ----------------------------------------------------- -4: Data for 2021/03/27 09:00:00 UTC 0H forecast. -4: -4: End of program -4: ========================================= -4: WAVEWATCH III GRIB output -4: -4: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -4: + wave_grib2_sbs.sh[102][[ 93 -gt 0 ]] -4: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '93 hour fcst' -grib gfs.wave.t12z.arctic.9km.f093.grib2 -4: 1:0:d=2021032312:SPC:surface:93 hour fcst: -4: 2:189892:d=2021032312:DIRC:surface:93 hour fcst: -4: 3:663114:d=2021032312:UOGRD:surface:93 hour fcst: -4: 4:851741:d=2021032312:VOGRD:surface:93 hour fcst: -4: 5:1043116:d=2021032312:WIND:surface:93 hour fcst: -4: 6:1343146:d=2021032312:WDIR:surface:93 hour fcst: -4: 7:1773944:d=2021032312:UGRD:surface:93 hour fcst: -4: 8:2071127:d=2021032312:VGRD:surface:93 hour fcst: -4: 9:2370834:d=2021032312:ICEC:surface:93 hour fcst: -4: 10:2540892:d=2021032312:HTSGW:surface:93 hour fcst: -4: 11:2752252:d=2021032312:IMWF:surface:93 hour fcst: -4: 12:2969646:d=2021032312:MWSPER:surface:93 hour fcst: -4: 13:3188428:d=2021032312:PERPW:surface:93 hour fcst: -4: 14:3413829:d=2021032312:WWSDIR:surface:93 hour fcst: -4: 15:3716724:d=2021032312:DIRPW:surface:93 hour fcst: -4: 16:4026119:d=2021032312:WVHGT:surface:93 hour fcst: -4: 17:4231804:d=2021032312:SWELL:1 in sequence:93 hour fcst: -4: 18:4446204:d=2021032312:SWELL:2 in sequence:93 hour fcst: -4: 19:4609378:d=2021032312:SWELL:3 in sequence:93 hour fcst: -4: 20:4742794:d=2021032312:WVPER:surface:93 hour fcst: -4: 21:4958921:d=2021032312:SWPER:1 in sequence:93 hour fcst: -4: 22:5182754:d=2021032312:SWPER:2 in sequence:93 hour fcst: -4: 23:5360090:d=2021032312:SWPER:3 in sequence:93 hour fcst: -4: 24:5496957:d=2021032312:WVDIR:surface:93 hour fcst: -4: 25:5775171:d=2021032312:SWDIR:1 in sequence:93 hour fcst: -4: 26:6094761:d=2021032312:SWDIR:2 in sequence:93 hour fcst: -4: 27:6315672:d=2021032312:SWDIR:3 in sequence:93 hour fcst: -4: + wave_grib2_sbs.sh[104]err=0 -4: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.arctic.9km.f093.grib2 -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f093.grib2 ]] -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f093.grib2.idx ]] -4: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.arctic.9km.f093.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f093.grib2 -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f093.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f093.grib2 = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f093.grib2 ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f093.grib2 -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f093.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f093.grib2.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f093.grib2.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f093.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f093.grib2 -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.arctic.9km.f093.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f093.grib2.idx -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f093.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f093.grib2.idx = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f093.grib2.idx ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f093.grib2.idx -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f093.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f093.grib2.idx.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f093.grib2.idx.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f093.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f093.grib2.idx -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.arctic.9km.f093.grib2 and gfs.wave.t12z.arctic.9km.f093.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_aoc_9km to COM' -4: INFO: Copied gfs.wave.t12z.arctic.9km.f093.grib2 and gfs.wave.t12z.arctic.9km.f093.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_aoc_9km to COM -4: + wave_grib2_sbs.sh[130][[ aoc_9km == '' ]] -4: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -4: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.arctic.9km.f093.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -4: INFO: gfs.wave.t12z.arctic.9km.f093.grib2 is global.0p50 or SENDDBN is NO, no alert sent -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + cmdfile.3[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032709 3600. 9999 -2: + cmdfile.3[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grid_interp_sbs.sh[25]grdID=wc_10m -2: + wave_grid_interp_sbs.sh[26]valid_time=2021032709 -2: + wave_grid_interp_sbs.sh[27]dt=3600. -2: + wave_grid_interp_sbs.sh[28]nst=9999 -2: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676 -2: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.uglo_100km ./out_grd.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mod_def.uglo_100km ./mod_def.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mod_def.wc_10m ./mod_def.wc_10m -2: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ]] -2: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m'\''' -2: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m' -2: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[51]weights_found=1 -2: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 090000' -2: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 090000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/wc_10m/g ww3_gint.inp.tmpl -2: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -2: $ Input file for interpolation of uglo_100km to wc_10m -2: $------------------------------------------------ -2: $ Start Time 3600. NSteps -2: 20210327 090000 3600. 9999 -2: $ Total number of grids -2: 2 -2: $ Grid extensions -2: 'uglo_100km' -2: 'wc_10m' -2: $ -2: 0 -2: $ -2: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[70]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.315371 ']' -2: ++ prep_step[4]echo gfs_ww3_gint.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -2: INFO: Executing 'gfs_ww3_gint.x' -2: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[73]cat grid_interp.wc_10m.out -2: -2: *** WAVEWATCH III Grid interpolation *** -2: =============================================== -2: -2: Comment character is '$' -2: -2: Time Information : -2: --------------------------------------------- -2: Starting Time : 2021/03/27 09:00:00 UTC -2: Interval (in sec) : 3600.00 -2: Number of requests : 9999 -2: --------------------------------------------- -2: Number of grids (including output grid) = 2 -2: -2: -2: Extension for grid 1 is --> uglo_100km -2: -2: Grid Particulars are : -2: Dimensions = 45166 1 -2: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Extension for grid 2 is --> wc_10m -2: -2: Grid Particulars are : -2: Dimensions = 241 151 -2: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -2: -2: -2: Preparing interpolation weights for output grid -2: Total number of wet points for interpolation 11044 -2: -2: -2: Variable: Grid Interpolation Map Units: 0.100E+01 -2: -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: +---------------------------------------------------------------------------------------------------------------+ -2: 151 | 0 0 0 0 | -2: | 0 0 0 0 0 | -2: | 0 0 0 0 0 | -2: 130 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 109 | 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 88 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 | -2: 67 | 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 | -2: 46 | 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: 25 | 0 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 0 | -2: | | -2: 4 | | -2: +---------------------------------------------------------------------------------------------------------------+ -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: -2: -2: Interpolating fields .... -2: -2: Output group 1 -2: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -2: Output group 2 -2: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -2: Output group 3 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 4 -2: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -2: Output group 5 -2: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -2: Output group 6 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 7 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 8 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 9 -2: Output variables skipped -2: Output group 10 -2: Output variables skipped -2: ------------------------------------------------ -2: 1Current vel. -2: 1Wind speed -2: 1Ice concentration -2: 2Wave height -2: 2Mean wave period(+2) -2: 2Mean wave period(+1) -2: 2Peak frequency -2: 2Mean wave dir. a1b1 -2: 2Peak direction -2: 4Part. wave height -2: 4Part. peak period -2: 4Part. mean direction -2: 5Charnock parameter -2: ------------------------------------------------ -2: OUTPUT TIME : 2021/03/27 09:00:00 UTC -2: -2: End of file reached -2: -2: -2: *** End of Grid interpolation Routine *** -2: =============================================== -2: -2: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -2: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -2: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_wc_10m/out_grd.wc_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.wc_10m -2: + cmdfile.3[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032709 93 wcoast 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + cmdfile.3[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib2_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grib2_sbs.sh[30]grdID=wc_10m -2: + wave_grib2_sbs.sh[31]GRIDNR=255 -2: + wave_grib2_sbs.sh[32]MODNR=11 -2: + wave_grib2_sbs.sh[33]valid_time=2021032709 -2: + wave_grib2_sbs.sh[34]fhr=93 -2: + wave_grib2_sbs.sh[35]grid_region=wcoast -2: + wave_grib2_sbs.sh[36]grid_res=0p16 -2: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676 -2: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_wc_10m -2: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_wc_10m -2: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_wc_10m -2: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_wc_10m -2: ++ wave_grib2_sbs.sh[47]printf %03i 93 -2: + wave_grib2_sbs.sh[47]FH3=093 -2: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_wcoast_0p16 -2: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.wcoast.0p16.f093.grib2 -2: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f093.grib2 ]] -2: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mod_def.wc_10m ./mod_def.ww3 -2: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.wc_10m ./out_grd.ww3 -2: + wave_grib2_sbs.sh[73]ngrib=1 -2: + wave_grib2_sbs.sh[74]dtgrib=3600 -2: + wave_grib2_sbs.sh[75]tstart='20210327 090000' -2: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 090000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.wc_10m.inp.tmpl -2: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -2: $ WAVEWATCH-III gridded output input file -2: $ ---------------------------------------- -2: 20210327 090000 3600 1 -2: N -2: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -2: $ -2: 20210327 090000 7 11 255 0 0 -2: $ -2: $ end of input file -2: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[88]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.315371 ']' -2: ++ prep_step[4]echo gfs_ww3_grib.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[90]export err=0 -2: + wave_grib2_sbs.sh[90]err=0 -2: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[95]cat grib2_wcoast_093.out -2: -2: *** WAVEWATCH III GRIB output postp. *** -2: ============================================== -2: -2: Comment character is '$' -2: -2: Grid name : West Coast 10 min wave grid -2: -2: LINEIN: -2: 20210327 090000 3600 1 -2: -2: 20210327090000 3600 1 -2: GEN_PRO -99999 -2: -2: Output time data : -2: ----------------------------------------------------- -2: First time : 2021/03/27 09:00:00 UTC -2: Interval : 01:00:00 -2: Number of requests : 1 -2: Fields : Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: Requested output fields not yet available: -2: ----------------------------------------------------- -2: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -2: -2: Successfully requested output fields : -2: ----------------------------------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: -2: Additional GRIB parameters : -2: ----------------------------------------------------- -2: Run time : 2021/03/27 09:00:00 UTC -2: GRIB center ID : 7 -2: GRIB gen. proc. ID : 11 -2: GRIB grid ID : 255 -2: GRIB GDS parameter : 0 -2: Fields in file : -2: -------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: CHOSEN GRID TYPE: : LLRECTILINEAR -2: -2: -2: -2: Generating file -2: ----------------------------------------------------- -2: Data for 2021/03/27 09:00:00 UTC 0H forecast. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: -2: End of program -2: ========================================= -2: WAVEWATCH III GRIB output -2: -2: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -2: + wave_grib2_sbs.sh[102][[ 93 -gt 0 ]] -2: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '93 hour fcst' -grib gfs.wave.t12z.wcoast.0p16.f093.grib2 -2: 1:0:d=2021032312:SPC:surface:93 hour fcst: -2: 2:7407:d=2021032312:DIRC:surface:93 hour fcst: -2: 3:24450:d=2021032312:UOGRD:surface:93 hour fcst: -2: 4:31394:d=2021032312:VOGRD:surface:93 hour fcst: -2: 5:39079:d=2021032312:WIND:surface:93 hour fcst: -2: 6:50610:d=2021032312:WDIR:surface:93 hour fcst: -2: 7:67240:d=2021032312:UGRD:surface:93 hour fcst: -2: 8:78457:d=2021032312:VGRD:surface:93 hour fcst: -2: 9:90142:d=2021032312:ICEC:surface:93 hour fcst: -2: 10:94872:d=2021032312:HTSGW:surface:93 hour fcst: -2: 11:104391:d=2021032312:IMWF:surface:93 hour fcst: -2: 12:113997:d=2021032312:MWSPER:surface:93 hour fcst: -2: 13:123676:d=2021032312:PERPW:surface:93 hour fcst: -2: 14:133716:d=2021032312:WWSDIR:surface:93 hour fcst: -2: 15:149021:d=2021032312:DIRPW:surface:93 hour fcst: -2: 16:164473:d=2021032312:WVHGT:surface:93 hour fcst: -2: 17:172778:d=2021032312:SWELL:1 in sequence:93 hour fcst: -2: 18:180427:d=2021032312:SWELL:2 in sequence:93 hour fcst: -2: 19:185589:d=2021032312:SWELL:3 in sequence:93 hour fcst: -2: 20:190319:d=2021032312:WVPER:surface:93 hour fcst: -2: 21:198997:d=2021032312:SWPER:1 in sequence:93 hour fcst: -2: 22:206699:d=2021032312:SWPER:2 in sequence:93 hour fcst: -2: 23:212059:d=2021032312:SWPER:3 in sequence:93 hour fcst: -2: 24:216789:d=2021032312:WVDIR:surface:93 hour fcst: -2: 25:229224:d=2021032312:SWDIR:1 in sequence:93 hour fcst: -2: 26:239753:d=2021032312:SWDIR:2 in sequence:93 hour fcst: -2: 27:245399:d=2021032312:SWDIR:3 in sequence:93 hour fcst: -2: + wave_grib2_sbs.sh[104]err=0 -2: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.wcoast.0p16.f093.grib2 -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f093.grib2 ]] -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f093.grib2.idx ]] -2: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.wcoast.0p16.f093.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f093.grib2 -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f093.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f093.grib2 = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f093.grib2 ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f093.grib2 -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f093.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f093.grib2.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f093.grib2.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f093.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f093.grib2 -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.wcoast.0p16.f093.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f093.grib2.idx -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f093.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f093.grib2.idx = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f093.grib2.idx ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f093.grib2.idx -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f093.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f093.grib2.idx.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f093.grib2.idx.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f093.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f093.grib2.idx -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.wcoast.0p16.f093.grib2 and gfs.wave.t12z.wcoast.0p16.f093.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_wc_10m to COM' -2: INFO: Copied gfs.wave.t12z.wcoast.0p16.f093.grib2 and gfs.wave.t12z.wcoast.0p16.f093.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_wc_10m to COM -2: + wave_grib2_sbs.sh[130][[ wc_10m == '' ]] -2: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -2: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.wcoast.0p16.f093.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -2: INFO: gfs.wave.t12z.wcoast.0p16.f093.grib2 is global.0p50 or SENDDBN is NO, no alert sent -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + cmdfile.2[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032709 3600. 9999 -1: + cmdfile.2[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grid_interp_sbs.sh[25]grdID=ep_10m -1: + wave_grid_interp_sbs.sh[26]valid_time=2021032709 -1: + wave_grid_interp_sbs.sh[27]dt=3600. -1: + wave_grid_interp_sbs.sh[28]nst=9999 -1: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676 -1: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.uglo_100km ./out_grd.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mod_def.uglo_100km ./mod_def.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mod_def.ep_10m ./mod_def.ep_10m -1: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ]] -1: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m'\''' -1: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m' -1: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[51]weights_found=1 -1: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 090000' -1: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 090000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/ep_10m/g ww3_gint.inp.tmpl -1: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -1: $ Input file for interpolation of uglo_100km to ep_10m -1: $------------------------------------------------ -1: $ Start Time 3600. NSteps -1: 20210327 090000 3600. 9999 -1: $ Total number of grids -1: 2 -1: $ Grid extensions -1: 'uglo_100km' -1: 'ep_10m' -1: $ -1: 0 -1: $ -1: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[70]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.315371 ']' -1: ++ prep_step[4]echo gfs_ww3_gint.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -1: INFO: Executing 'gfs_ww3_gint.x' -1: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[73]cat grid_interp.ep_10m.out -1: -1: *** WAVEWATCH III Grid interpolation *** -1: =============================================== -1: -1: Comment character is '$' -1: -1: Time Information : -1: --------------------------------------------- -1: Starting Time : 2021/03/27 09:00:00 UTC -1: Interval (in sec) : 3600.00 -1: Number of requests : 9999 -1: --------------------------------------------- -1: Number of grids (including output grid) = 2 -1: -1: -1: Extension for grid 1 is --> uglo_100km -1: -1: Grid Particulars are : -1: Dimensions = 45166 1 -1: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Extension for grid 2 is --> ep_10m -1: -1: Grid Particulars are : -1: Dimensions = 511 301 -1: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -1: -1: -1: Preparing interpolation weights for output grid -1: Total number of wet points for interpolation 7439 -1: -1: -1: Variable: Grid Interpolation Map Units: 0.100E+01 -1: -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 301 | | -1: | | -1: | 0 0 | -1: 262 | 0 0 0 0 | -1: | 0 0 0 0 | -1: | 0 0 0 0 | -1: 223 | 0 0 0 | -1: | 0 | -1: | 0 | -1: 184 | | -1: | 0 0 0 | -1: | 0 0 0 0 | -1: 145 | 0 | -1: | | -1: | | -1: 106 | | -1: | | -1: | | -1: 67 | | -1: | 0 | -1: | | -1: 28 | 0 | -1: | | -1: | | -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: -1: -1: Interpolating fields .... -1: -1: Output group 1 -1: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -1: Output group 2 -1: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -1: Output group 3 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 4 -1: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -1: Output group 5 -1: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -1: Output group 6 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 7 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 8 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 9 -1: Output variables skipped -1: Output group 10 -1: Output variables skipped -1: ------------------------------------------------ -1: 1Current vel. -1: 1Wind speed -1: 1Ice concentration -1: 2Wave height -1: 2Mean wave period(+2) -1: 2Mean wave period(+1) -1: 2Peak frequency -1: 2Mean wave dir. a1b1 -1: 2Peak direction -1: 4Part. wave height -1: 4Part. peak period -1: 4Part. mean direction -1: 5Charnock parameter -1: ------------------------------------------------ -1: OUTPUT TIME : 2021/03/27 09:00:00 UTC -1: -1: End of file reached -1: -1: -1: *** End of Grid interpolation Routine *** -1: =============================================== -1: -1: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -1: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -1: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_ep_10m/out_grd.ep_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.ep_10m -1: + cmdfile.2[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032709 93 epacif 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + cmdfile.2[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib2_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grib2_sbs.sh[30]grdID=ep_10m -1: + wave_grib2_sbs.sh[31]GRIDNR=255 -1: + wave_grib2_sbs.sh[32]MODNR=11 -1: + wave_grib2_sbs.sh[33]valid_time=2021032709 -1: + wave_grib2_sbs.sh[34]fhr=93 -1: + wave_grib2_sbs.sh[35]grid_region=epacif -1: + wave_grib2_sbs.sh[36]grid_res=0p16 -1: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676 -1: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_ep_10m -1: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_ep_10m -1: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_ep_10m -1: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_ep_10m -1: ++ wave_grib2_sbs.sh[47]printf %03i 93 -1: + wave_grib2_sbs.sh[47]FH3=093 -1: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_epacif_0p16 -1: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.epacif.0p16.f093.grib2 -1: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f093.grib2 ]] -1: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mod_def.ep_10m ./mod_def.ww3 -1: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.ep_10m ./out_grd.ww3 -1: + wave_grib2_sbs.sh[73]ngrib=1 -1: + wave_grib2_sbs.sh[74]dtgrib=3600 -1: + wave_grib2_sbs.sh[75]tstart='20210327 090000' -1: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 090000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.ep_10m.inp.tmpl -1: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -1: $ WAVEWATCH-III gridded output input file -1: $ ---------------------------------------- -1: 20210327 090000 3600 1 -1: N -1: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -1: $ -1: 20210327 090000 7 11 255 0 0 -1: $ -1: $ end of input file -1: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[88]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.315371 ']' -1: ++ prep_step[4]echo gfs_ww3_grib.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[90]export err=0 -1: + wave_grib2_sbs.sh[90]err=0 -1: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[95]cat grib2_epacif_093.out -1: -1: *** WAVEWATCH III GRIB output postp. *** -1: ============================================== -1: -1: Comment character is '$' -1: -1: Grid name : East Pacific 10 min wave grid -1: -1: LINEIN: -1: 20210327 090000 3600 1 -1: -1: 20210327090000 3600 1 -1: GEN_PRO -99999 -1: -1: Output time data : -1: ----------------------------------------------------- -1: First time : 2021/03/27 09:00:00 UTC -1: Interval : 01:00:00 -1: Number of requests : 1 -1: Fields : Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: Requested output fields not yet available: -1: ----------------------------------------------------- -1: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -1: -1: Successfully requested output fields : -1: ----------------------------------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: -1: Additional GRIB parameters : -1: ----------------------------------------------------- -1: Run time : 2021/03/27 09:00:00 UTC -1: GRIB center ID : 7 -1: GRIB gen. proc. ID : 11 -1: GRIB grid ID : 255 -1: GRIB GDS parameter : 0 -1: Fields in file : -1: -------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: CHOSEN GRID TYPE: : LLRECTILINEAR -1: -1: -1: -1: Generating file -1: ----------------------------------------------------- -1: Data for 2021/03/27 09:00:00 UTC 0H forecast. -1: Warning: bitmask off everywhere. -1: Pretend one point in jpcpack to avoid crash. -1: Warning: bitmask off everywhere. -1: Pretend one point in jpcpack to avoid crash. -1: Warning: bitmask off everywhere. -1: Pretend one point in jpcpack to avoid crash. -1: -1: End of program -1: ========================================= -1: WAVEWATCH III GRIB output -1: -1: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -1: + wave_grib2_sbs.sh[102][[ 93 -gt 0 ]] -1: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '93 hour fcst' -grib gfs.wave.t12z.epacif.0p16.f093.grib2 -1: 1:0:d=2021032312:SPC:surface:93 hour fcst: -1: 2:21812:d=2021032312:DIRC:surface:93 hour fcst: -1: 3:49441:d=2021032312:UOGRD:surface:93 hour fcst: -1: 4:71200:d=2021032312:VOGRD:surface:93 hour fcst: -1: 5:93153:d=2021032312:WIND:surface:93 hour fcst: -1: 6:117878:d=2021032312:WDIR:surface:93 hour fcst: -1: 7:145083:d=2021032312:UGRD:surface:93 hour fcst: -1: 8:169750:d=2021032312:VGRD:surface:93 hour fcst: -1: 9:194337:d=2021032312:ICEC:surface:93 hour fcst: -1: 10:213745:d=2021032312:HTSGW:surface:93 hour fcst: -1: 11:236122:d=2021032312:IMWF:surface:93 hour fcst: -1: 12:258972:d=2021032312:MWSPER:surface:93 hour fcst: -1: 13:281876:d=2021032312:PERPW:surface:93 hour fcst: -1: 14:304654:d=2021032312:WWSDIR:surface:93 hour fcst: -1: 15:330612:d=2021032312:DIRPW:surface:93 hour fcst: -1: 16:356665:d=2021032312:WVHGT:surface:93 hour fcst: -1: 17:378669:d=2021032312:SWELL:1 in sequence:93 hour fcst: -1: 18:401614:d=2021032312:SWELL:2 in sequence:93 hour fcst: -1: 19:422161:d=2021032312:SWELL:3 in sequence:93 hour fcst: -1: 20:441569:d=2021032312:WVPER:surface:93 hour fcst: -1: 21:463643:d=2021032312:SWPER:1 in sequence:93 hour fcst: -1: 22:486497:d=2021032312:SWPER:2 in sequence:93 hour fcst: -1: 23:507075:d=2021032312:SWPER:3 in sequence:93 hour fcst: -1: 24:526483:d=2021032312:WVDIR:surface:93 hour fcst: -1: 25:550564:d=2021032312:SWDIR:1 in sequence:93 hour fcst: -1: 26:577737:d=2021032312:SWDIR:2 in sequence:93 hour fcst: -1: 27:599352:d=2021032312:SWDIR:3 in sequence:93 hour fcst: -1: + wave_grib2_sbs.sh[104]err=0 -1: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.epacif.0p16.f093.grib2 -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f093.grib2 ]] -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f093.grib2.idx ]] -1: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.epacif.0p16.f093.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f093.grib2 -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f093.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f093.grib2 = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f093.grib2 ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f093.grib2 -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f093.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f093.grib2.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f093.grib2.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f093.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f093.grib2 -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.epacif.0p16.f093.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f093.grib2.idx -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f093.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f093.grib2.idx = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f093.grib2.idx ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f093.grib2.idx -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f093.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f093.grib2.idx.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f093.grib2.idx.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f093.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f093.grib2.idx -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.epacif.0p16.f093.grib2 and gfs.wave.t12z.epacif.0p16.f093.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_ep_10m to COM' -1: INFO: Copied gfs.wave.t12z.epacif.0p16.f093.grib2 and gfs.wave.t12z.epacif.0p16.f093.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_ep_10m to COM -1: + wave_grib2_sbs.sh[130][[ ep_10m == '' ]] -1: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -1: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.epacif.0p16.f093.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -1: INFO: gfs.wave.t12z.epacif.0p16.f093.grib2 is global.0p50 or SENDDBN is NO, no alert sent -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + cmdfile.4[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032709 3600. 9999 -3: + cmdfile.4[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grid_interp_sbs.sh[25]grdID=glo_30m -3: + wave_grid_interp_sbs.sh[26]valid_time=2021032709 -3: + wave_grid_interp_sbs.sh[27]dt=3600. -3: + wave_grid_interp_sbs.sh[28]nst=9999 -3: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676 -3: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.uglo_100km ./out_grd.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mod_def.uglo_100km ./mod_def.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mod_def.glo_30m ./mod_def.glo_30m -3: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ]] -3: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m'\''' -3: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m' -3: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[51]weights_found=1 -3: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 090000' -3: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 090000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/glo_30m/g ww3_gint.inp.tmpl -3: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -3: $ Input file for interpolation of uglo_100km to glo_30m -3: $------------------------------------------------ -3: $ Start Time 3600. NSteps -3: 20210327 090000 3600. 9999 -3: $ Total number of grids -3: 2 -3: $ Grid extensions -3: 'uglo_100km' -3: 'glo_30m' -3: $ -3: 0 -3: $ -3: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[70]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.315371 ']' -3: ++ prep_step[4]echo gfs_ww3_gint.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -3: INFO: Executing 'gfs_ww3_gint.x' -3: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[73]cat grid_interp.glo_30m.out -3: -3: *** WAVEWATCH III Grid interpolation *** -3: =============================================== -3: -3: Comment character is '$' -3: -3: Time Information : -3: --------------------------------------------- -3: Starting Time : 2021/03/27 09:00:00 UTC -3: Interval (in sec) : 3600.00 -3: Number of requests : 9999 -3: --------------------------------------------- -3: Number of grids (including output grid) = 2 -3: -3: -3: Extension for grid 1 is --> uglo_100km -3: -3: Grid Particulars are : -3: Dimensions = 45166 1 -3: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Extension for grid 2 is --> glo_30m -3: -3: Grid Particulars are : -3: Dimensions = 720 336 -3: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -3: -3: -3: Preparing interpolation weights for output grid -3: Total number of wet points for interpolation 167619 -3: -3: -3: Variable: Grid Interpolation Map Units: 0.100E+01 -3: -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 336 | | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 291 | 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 | -3: 246 | 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 201 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 156 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 111 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 66 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 21 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 | -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: -3: -3: Interpolating fields .... -3: -3: Output group 1 -3: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -3: Output group 2 -3: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -3: Output group 3 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 4 -3: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -3: Output group 5 -3: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -3: Output group 6 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 7 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 8 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 9 -3: Output variables skipped -3: Output group 10 -3: Output variables skipped -3: ------------------------------------------------ -3: 1Current vel. -3: 1Wind speed -3: 1Ice concentration -3: 2Wave height -3: 2Mean wave period(+2) -3: 2Mean wave period(+1) -3: 2Peak frequency -3: 2Mean wave dir. a1b1 -3: 2Peak direction -3: 4Part. wave height -3: 4Part. peak period -3: 4Part. mean direction -3: 5Charnock parameter -3: ------------------------------------------------ -3: OUTPUT TIME : 2021/03/27 09:00:00 UTC -3: -3: End of file reached -3: -3: -3: *** End of Grid interpolation Routine *** -3: =============================================== -3: -3: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -3: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -3: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_glo_30m/out_grd.glo_30m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.glo_30m -3: + cmdfile.4[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032709 93 global 0p50 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + cmdfile.4[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib2_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grib2_sbs.sh[30]grdID=glo_30m -3: + wave_grib2_sbs.sh[31]GRIDNR=255 -3: + wave_grib2_sbs.sh[32]MODNR=11 -3: + wave_grib2_sbs.sh[33]valid_time=2021032709 -3: + wave_grib2_sbs.sh[34]fhr=93 -3: + wave_grib2_sbs.sh[35]grid_region=global -3: + wave_grib2_sbs.sh[36]grid_res=0p50 -3: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676 -3: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_glo_30m -3: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_glo_30m -3: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_glo_30m -3: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_glo_30m -3: ++ wave_grib2_sbs.sh[47]printf %03i 93 -3: + wave_grib2_sbs.sh[47]FH3=093 -3: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p50 -3: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p50.f093.grib2 -3: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f093.grib2 ]] -3: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mod_def.glo_30m ./mod_def.ww3 -3: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.glo_30m ./out_grd.ww3 -3: + wave_grib2_sbs.sh[73]ngrib=1 -3: + wave_grib2_sbs.sh[74]dtgrib=3600 -3: + wave_grib2_sbs.sh[75]tstart='20210327 090000' -3: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 090000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.glo_30m.inp.tmpl -3: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -3: $ WAVEWATCH-III gridded output input file -3: $ ---------------------------------------- -3: 20210327 090000 3600 1 -3: N -3: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -3: $ -3: 20210327 090000 7 11 255 0 0 -3: $ -3: $ end of input file -3: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[88]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.315371 ']' -3: ++ prep_step[4]echo gfs_ww3_grib.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[90]export err=0 -3: + wave_grib2_sbs.sh[90]err=0 -3: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[95]cat grib2_global_093.out -3: -3: *** WAVEWATCH III GRIB output postp. *** -3: ============================================== -3: -3: Comment character is '$' -3: -3: Grid name : Global 30 min wave grid -3: -3: LINEIN: -3: 20210327 090000 3600 1 -3: -3: 20210327090000 3600 1 -3: GEN_PRO -99999 -3: -3: Output time data : -3: ----------------------------------------------------- -3: First time : 2021/03/27 09:00:00 UTC -3: Interval : 01:00:00 -3: Number of requests : 1 -3: Fields : Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: Requested output fields not yet available: -3: ----------------------------------------------------- -3: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -3: -3: Successfully requested output fields : -3: ----------------------------------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: -3: Additional GRIB parameters : -3: ----------------------------------------------------- -3: Run time : 2021/03/27 09:00:00 UTC -3: GRIB center ID : 7 -3: GRIB gen. proc. ID : 11 -3: GRIB grid ID : 255 -3: GRIB GDS parameter : 0 -3: Fields in file : -3: -------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: CHOSEN GRID TYPE: : LLRECTILINEAR -3: -3: -3: -3: Generating file -3: ----------------------------------------------------- -3: Data for 2021/03/27 09:00:00 UTC 0H forecast. -3: -3: End of program -3: ========================================= -3: WAVEWATCH III GRIB output -3: -3: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -3: + wave_grib2_sbs.sh[102][[ 93 -gt 0 ]] -3: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '93 hour fcst' -grib gfs.wave.t12z.global.0p50.f093.grib2 -3: 1:0:d=2021032312:SPC:surface:93 hour fcst: -3: 2:77842:d=2021032312:DIRC:surface:93 hour fcst: -3: 3:321251:d=2021032312:UOGRD:surface:93 hour fcst: -3: 4:396768:d=2021032312:VOGRD:surface:93 hour fcst: -3: 5:476500:d=2021032312:WIND:surface:93 hour fcst: -3: 6:634139:d=2021032312:WDIR:surface:93 hour fcst: -3: 7:869810:d=2021032312:UGRD:surface:93 hour fcst: -3: 8:1024176:d=2021032312:VGRD:surface:93 hour fcst: -3: 9:1180420:d=2021032312:ICEC:surface:93 hour fcst: -3: 10:1220572:d=2021032312:HTSGW:surface:93 hour fcst: -3: 11:1319272:d=2021032312:IMWF:surface:93 hour fcst: -3: 12:1428345:d=2021032312:MWSPER:surface:93 hour fcst: -3: 13:1538400:d=2021032312:PERPW:surface:93 hour fcst: -3: 14:1655151:d=2021032312:WWSDIR:surface:93 hour fcst: -3: 15:1855489:d=2021032312:DIRPW:surface:93 hour fcst: -3: 16:2062352:d=2021032312:WVHGT:surface:93 hour fcst: -3: 17:2157592:d=2021032312:SWELL:1 in sequence:93 hour fcst: -3: 18:2261294:d=2021032312:SWELL:2 in sequence:93 hour fcst: -3: 19:2332004:d=2021032312:SWELL:3 in sequence:93 hour fcst: -3: 20:2374907:d=2021032312:WVPER:surface:93 hour fcst: -3: 21:2481500:d=2021032312:SWPER:1 in sequence:93 hour fcst: -3: 22:2598638:d=2021032312:SWPER:2 in sequence:93 hour fcst: -3: 23:2687479:d=2021032312:SWPER:3 in sequence:93 hour fcst: -3: 24:2738570:d=2021032312:WVDIR:surface:93 hour fcst: -3: 25:2900683:d=2021032312:SWDIR:1 in sequence:93 hour fcst: -3: 26:3101550:d=2021032312:SWDIR:2 in sequence:93 hour fcst: -3: 27:3241685:d=2021032312:SWDIR:3 in sequence:93 hour fcst: -3: + wave_grib2_sbs.sh[104]err=0 -3: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p50.f093.grib2 -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f093.grib2 ]] -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f093.grib2.idx ]] -3: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p50.f093.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f093.grib2 -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f093.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f093.grib2 = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f093.grib2 ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f093.grib2 -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f093.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f093.grib2.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f093.grib2.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f093.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f093.grib2 -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p50.f093.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f093.grib2.idx -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f093.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f093.grib2.idx = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f093.grib2.idx ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f093.grib2.idx -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f093.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f093.grib2.idx.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f093.grib2.idx.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f093.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f093.grib2.idx -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p50.f093.grib2 and gfs.wave.t12z.global.0p50.f093.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_glo_30m to COM' -3: INFO: Copied gfs.wave.t12z.global.0p50.f093.grib2 and gfs.wave.t12z.global.0p50.f093.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_glo_30m to COM -3: + wave_grib2_sbs.sh[130][[ glo_30m == '' ]] -3: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -3: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p50.f093.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -3: INFO: gfs.wave.t12z.global.0p50.f093.grib2 is global.0p50 or SENDDBN is NO, no alert sent -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + cmdfile.6[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032709 3600. 9999 -5: + cmdfile.6[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grid_interp_sbs.sh[25]grdID=gnh_10m -5: + wave_grid_interp_sbs.sh[26]valid_time=2021032709 -5: + wave_grid_interp_sbs.sh[27]dt=3600. -5: + wave_grid_interp_sbs.sh[28]nst=9999 -5: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676 -5: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.uglo_100km ./out_grd.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mod_def.uglo_100km ./mod_def.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mod_def.gnh_10m ./mod_def.gnh_10m -5: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m'\''' -5: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m' -5: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[51]weights_found=1 -5: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 090000' -5: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 090000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gnh_10m/g ww3_gint.inp.tmpl -5: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -5: $ Input file for interpolation of uglo_100km to gnh_10m -5: $------------------------------------------------ -5: $ Start Time 3600. NSteps -5: 20210327 090000 3600. 9999 -5: $ Total number of grids -5: 2 -5: $ Grid extensions -5: 'uglo_100km' -5: 'gnh_10m' -5: $ -5: 0 -5: $ -5: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[70]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.315371 ']' -5: ++ prep_step[4]echo gfs_ww3_gint.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -5: INFO: Executing 'gfs_ww3_gint.x' -5: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[73]cat grid_interp.gnh_10m.out -5: -5: *** WAVEWATCH III Grid interpolation *** -5: =============================================== -5: -5: Comment character is '$' -5: -5: Time Information : -5: --------------------------------------------- -5: Starting Time : 2021/03/27 09:00:00 UTC -5: Interval (in sec) : 3600.00 -5: Number of requests : 9999 -5: --------------------------------------------- -5: Number of grids (including output grid) = 2 -5: -5: -5: Extension for grid 1 is --> uglo_100km -5: -5: Grid Particulars are : -5: Dimensions = 45166 1 -5: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Extension for grid 2 is --> gnh_10m -5: -5: Grid Particulars are : -5: Dimensions = 2160 406 -5: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -5: -5: -5: Preparing interpolation weights for output grid -5: Total number of wet points for interpolation 571209 -5: -5: -5: Variable: Grid Interpolation Map Units: 0.100E+01 -5: -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 406 | | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: 355 | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 304 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 253 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 202 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 151 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 100 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 49 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: -5: -5: Interpolating fields .... -5: -5: Output group 1 -5: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -5: Output group 2 -5: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -5: Output group 3 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 4 -5: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -5: Output group 5 -5: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -5: Output group 6 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 7 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 8 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 9 -5: Output variables skipped -5: Output group 10 -5: Output variables skipped -5: ------------------------------------------------ -5: 1Current vel. -5: 1Wind speed -5: 1Ice concentration -5: 2Wave height -5: 2Mean wave period(+2) -5: 2Mean wave period(+1) -5: 2Peak frequency -5: 2Mean wave dir. a1b1 -5: 2Peak direction -5: 4Part. wave height -5: 4Part. peak period -5: 4Part. mean direction -5: 5Charnock parameter -5: ------------------------------------------------ -5: OUTPUT TIME : 2021/03/27 09:00:00 UTC -5: -5: End of file reached -5: -5: -5: *** End of Grid interpolation Routine *** -5: =============================================== -5: -5: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -5: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -5: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_gnh_10m/out_grd.gnh_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.gnh_10m -5: + cmdfile.6[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032709 93 global 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + cmdfile.6[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib2_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grib2_sbs.sh[30]grdID=gnh_10m -5: + wave_grib2_sbs.sh[31]GRIDNR=255 -5: + wave_grib2_sbs.sh[32]MODNR=11 -5: + wave_grib2_sbs.sh[33]valid_time=2021032709 -5: + wave_grib2_sbs.sh[34]fhr=93 -5: + wave_grib2_sbs.sh[35]grid_region=global -5: + wave_grib2_sbs.sh[36]grid_res=0p16 -5: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676 -5: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_gnh_10m -5: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_gnh_10m -5: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_gnh_10m -5: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_gnh_10m -5: ++ wave_grib2_sbs.sh[47]printf %03i 93 -5: + wave_grib2_sbs.sh[47]FH3=093 -5: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p16 -5: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p16.f093.grib2 -5: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f093.grib2 ]] -5: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mod_def.gnh_10m ./mod_def.ww3 -5: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.gnh_10m ./out_grd.ww3 -5: + wave_grib2_sbs.sh[73]ngrib=1 -5: + wave_grib2_sbs.sh[74]dtgrib=3600 -5: + wave_grib2_sbs.sh[75]tstart='20210327 090000' -5: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 090000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gnh_10m.inp.tmpl -5: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -5: $ WAVEWATCH-III gridded output input file -5: $ ---------------------------------------- -5: 20210327 090000 3600 1 -5: N -5: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -5: $ -5: 20210327 090000 7 11 255 0 0 -5: $ -5: $ end of input file -5: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[88]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.315371 ']' -5: ++ prep_step[4]echo gfs_ww3_grib.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[90]export err=0 -5: + wave_grib2_sbs.sh[90]err=0 -5: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[95]cat grib2_global_093.out -5: -5: *** WAVEWATCH III GRIB output postp. *** -5: ============================================== -5: -5: Comment character is '$' -5: -5: Grid name : GFSv16-wave N Hemisphere 1/6 d -5: -5: LINEIN: -5: 20210327 090000 3600 1 -5: -5: 20210327090000 3600 1 -5: GEN_PRO -99999 -5: -5: Output time data : -5: ----------------------------------------------------- -5: First time : 2021/03/27 09:00:00 UTC -5: Interval : 01:00:00 -5: Number of requests : 1 -5: Fields : Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: Requested output fields not yet available: -5: ----------------------------------------------------- -5: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -5: -5: Successfully requested output fields : -5: ----------------------------------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: -5: Additional GRIB parameters : -5: ----------------------------------------------------- -5: Run time : 2021/03/27 09:00:00 UTC -5: GRIB center ID : 7 -5: GRIB gen. proc. ID : 11 -5: GRIB grid ID : 255 -5: GRIB GDS parameter : 0 -5: Fields in file : -5: -------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: CHOSEN GRID TYPE: : LLRECTILINEAR -5: -5: -5: -5: Generating file -5: ----------------------------------------------------- -5: Data for 2021/03/27 09:00:00 UTC 0H forecast. -5: -5: End of program -5: ========================================= -5: WAVEWATCH III GRIB output -5: -5: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -5: + wave_grib2_sbs.sh[102][[ 93 -gt 0 ]] -5: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '93 hour fcst' -grib gfs.wave.t12z.global.0p16.f093.grib2 -5: 1:0:d=2021032312:SPC:surface:93 hour fcst: -5: 2:238719:d=2021032312:DIRC:surface:93 hour fcst: -5: 3:868138:d=2021032312:UOGRD:surface:93 hour fcst: -5: 4:1102626:d=2021032312:VOGRD:surface:93 hour fcst: -5: 5:1341604:d=2021032312:WIND:surface:93 hour fcst: -5: 6:1747295:d=2021032312:WDIR:surface:93 hour fcst: -5: 7:2381322:d=2021032312:UGRD:surface:93 hour fcst: -5: 8:2780648:d=2021032312:VGRD:surface:93 hour fcst: -5: 9:3180499:d=2021032312:ICEC:surface:93 hour fcst: -5: 10:3295029:d=2021032312:HTSGW:surface:93 hour fcst: -5: 11:3583949:d=2021032312:IMWF:surface:93 hour fcst: -5: 12:3895991:d=2021032312:MWSPER:surface:93 hour fcst: -5: 13:4210025:d=2021032312:PERPW:surface:93 hour fcst: -5: 14:4544868:d=2021032312:WWSDIR:surface:93 hour fcst: -5: 15:5125438:d=2021032312:DIRPW:surface:93 hour fcst: -5: 16:5735287:d=2021032312:WVHGT:surface:93 hour fcst: -5: 17:6012943:d=2021032312:SWELL:1 in sequence:93 hour fcst: -5: 18:6300791:d=2021032312:SWELL:2 in sequence:93 hour fcst: -5: 19:6493471:d=2021032312:SWELL:3 in sequence:93 hour fcst: -5: 20:6618834:d=2021032312:WVPER:surface:93 hour fcst: -5: 21:6930154:d=2021032312:SWPER:1 in sequence:93 hour fcst: -5: 22:7257021:d=2021032312:SWPER:2 in sequence:93 hour fcst: -5: 23:7493168:d=2021032312:SWPER:3 in sequence:93 hour fcst: -5: 24:7632745:d=2021032312:WVDIR:surface:93 hour fcst: -5: 25:8130785:d=2021032312:SWDIR:1 in sequence:93 hour fcst: -5: 26:8718154:d=2021032312:SWDIR:2 in sequence:93 hour fcst: -5: 27:9089920:d=2021032312:SWDIR:3 in sequence:93 hour fcst: -5: + wave_grib2_sbs.sh[104]err=0 -5: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p16.f093.grib2 -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f093.grib2 ]] -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f093.grib2.idx ]] -5: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p16.f093.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f093.grib2 -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f093.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f093.grib2 = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f093.grib2 ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f093.grib2 -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f093.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f093.grib2.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f093.grib2.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f093.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f093.grib2 -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p16.f093.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f093.grib2.idx -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f093.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f093.grib2.idx = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f093.grib2.idx ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f093.grib2.idx -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f093.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f093.grib2.idx.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f093.grib2.idx.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f093.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f093.grib2.idx -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p16.f093.grib2 and gfs.wave.t12z.global.0p16.f093.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_gnh_10m to COM' -5: INFO: Copied gfs.wave.t12z.global.0p16.f093.grib2 and gfs.wave.t12z.global.0p16.f093.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_gnh_10m to COM -5: + wave_grib2_sbs.sh[130][[ gnh_10m == '' ]] -5: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -5: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p16.f093.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -5: INFO: gfs.wave.t12z.global.0p16.f093.grib2 is global.0p50 or SENDDBN is NO, no alert sent -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + cmdfile.7[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032709 3600. 9999 -6: + cmdfile.7[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grid_interp_sbs.sh[25]grdID=gsh_15m -6: + wave_grid_interp_sbs.sh[26]valid_time=2021032709 -6: + wave_grid_interp_sbs.sh[27]dt=3600. -6: + wave_grid_interp_sbs.sh[28]nst=9999 -6: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676 -6: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.uglo_100km ./out_grd.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mod_def.uglo_100km ./mod_def.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mod_def.gsh_15m ./mod_def.gsh_15m -6: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m'\''' -6: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m' -6: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[51]weights_found=1 -6: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 090000' -6: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 090000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gsh_15m/g ww3_gint.inp.tmpl -6: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -6: $ Input file for interpolation of uglo_100km to gsh_15m -6: $------------------------------------------------ -6: $ Start Time 3600. NSteps -6: 20210327 090000 3600. 9999 -6: $ Total number of grids -6: 2 -6: $ Grid extensions -6: 'uglo_100km' -6: 'gsh_15m' -6: $ -6: 0 -6: $ -6: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[70]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.315371 ']' -6: ++ prep_step[4]echo gfs_ww3_gint.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -6: INFO: Executing 'gfs_ww3_gint.x' -6: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[73]cat grid_interp.gsh_15m.out -6: -6: *** WAVEWATCH III Grid interpolation *** -6: =============================================== -6: -6: Comment character is '$' -6: -6: Time Information : -6: --------------------------------------------- -6: Starting Time : 2021/03/27 09:00:00 UTC -6: Interval (in sec) : 3600.00 -6: Number of requests : 9999 -6: --------------------------------------------- -6: Number of grids (including output grid) = 2 -6: -6: -6: Extension for grid 1 is --> uglo_100km -6: -6: Grid Particulars are : -6: Dimensions = 45166 1 -6: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Extension for grid 2 is --> gsh_15m -6: -6: Grid Particulars are : -6: Dimensions = 1440 277 -6: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -6: -6: -6: Preparing interpolation weights for output grid -6: Total number of wet points for interpolation 317192 -6: -6: -6: Variable: Grid Interpolation Map Units: 0.100E+01 -6: -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 277 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 241 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 205 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 169 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 133 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 97 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 61 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 | -6: 25 | 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 | -6: | | -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: -6: -6: Interpolating fields .... -6: -6: Output group 1 -6: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -6: Output group 2 -6: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -6: Output group 3 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 4 -6: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -6: Output group 5 -6: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -6: Output group 6 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 7 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 8 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 9 -6: Output variables skipped -6: Output group 10 -6: Output variables skipped -6: ------------------------------------------------ -6: 1Current vel. -6: 1Wind speed -6: 1Ice concentration -6: 2Wave height -6: 2Mean wave period(+2) -6: 2Mean wave period(+1) -6: 2Peak frequency -6: 2Mean wave dir. a1b1 -6: 2Peak direction -6: 4Part. wave height -6: 4Part. peak period -6: 4Part. mean direction -6: 5Charnock parameter -6: ------------------------------------------------ -6: OUTPUT TIME : 2021/03/27 09:00:00 UTC -6: -6: End of file reached -6: -6: -6: *** End of Grid interpolation Routine *** -6: =============================================== -6: -6: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -6: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -6: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_gsh_15m/out_grd.gsh_15m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.gsh_15m -6: + cmdfile.7[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032709 93 gsouth 0p25 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + cmdfile.7[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib2_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grib2_sbs.sh[30]grdID=gsh_15m -6: + wave_grib2_sbs.sh[31]GRIDNR=255 -6: + wave_grib2_sbs.sh[32]MODNR=11 -6: + wave_grib2_sbs.sh[33]valid_time=2021032709 -6: + wave_grib2_sbs.sh[34]fhr=93 -6: + wave_grib2_sbs.sh[35]grid_region=gsouth -6: + wave_grib2_sbs.sh[36]grid_res=0p25 -6: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676 -6: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_gsh_15m -6: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_gsh_15m -6: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_gsh_15m -6: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_gsh_15m -6: ++ wave_grib2_sbs.sh[47]printf %03i 93 -6: + wave_grib2_sbs.sh[47]FH3=093 -6: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -6: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.gsouth.0p25.f093.grib2 -6: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f093.grib2 ]] -6: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mod_def.gsh_15m ./mod_def.ww3 -6: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.gsh_15m ./out_grd.ww3 -6: + wave_grib2_sbs.sh[73]ngrib=1 -6: + wave_grib2_sbs.sh[74]dtgrib=3600 -6: + wave_grib2_sbs.sh[75]tstart='20210327 090000' -6: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 090000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gsh_15m.inp.tmpl -6: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -6: $ WAVEWATCH-III gridded output input file -6: $ ---------------------------------------- -6: 20210327 090000 3600 1 -6: N -6: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -6: $ -6: 20210327 090000 7 11 255 0 0 -6: $ -6: $ end of input file -6: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[88]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.315371 ']' -6: ++ prep_step[4]echo gfs_ww3_grib.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[90]export err=0 -6: + wave_grib2_sbs.sh[90]err=0 -6: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[95]cat grib2_gsouth_093.out -6: -6: *** WAVEWATCH III GRIB output postp. *** -6: ============================================== -6: -6: Comment character is '$' -6: -6: Grid name : GFSv16-wave S Hemisphere 1/4 d -6: -6: LINEIN: -6: 20210327 090000 3600 1 -6: -6: 20210327090000 3600 1 -6: GEN_PRO -99999 -6: -6: Output time data : -6: ----------------------------------------------------- -6: First time : 2021/03/27 09:00:00 UTC -6: Interval : 01:00:00 -6: Number of requests : 1 -6: Fields : Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: Requested output fields not yet available: -6: ----------------------------------------------------- -6: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -6: -6: Successfully requested output fields : -6: ----------------------------------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: -6: Additional GRIB parameters : -6: ----------------------------------------------------- -6: Run time : 2021/03/27 09:00:00 UTC -6: GRIB center ID : 7 -6: GRIB gen. proc. ID : 11 -6: GRIB grid ID : 255 -6: GRIB GDS parameter : 0 -6: Fields in file : -6: -------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: CHOSEN GRID TYPE: : LLRECTILINEAR -6: -6: -6: -6: Generating file -6: ----------------------------------------------------- -6: Data for 2021/03/27 09:00:00 UTC 0H forecast. -6: -6: End of program -6: ========================================= -6: WAVEWATCH III GRIB output -6: -6: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -6: + wave_grib2_sbs.sh[102][[ 93 -gt 0 ]] -6: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '93 hour fcst' -grib gfs.wave.t12z.gsouth.0p25.f093.grib2 -6: 1:0:d=2021032312:SPC:surface:93 hour fcst: -6: 2:118237:d=2021032312:DIRC:surface:93 hour fcst: -6: 3:504204:d=2021032312:UOGRD:surface:93 hour fcst: -6: 4:617782:d=2021032312:VOGRD:surface:93 hour fcst: -6: 5:740197:d=2021032312:WIND:surface:93 hour fcst: -6: 6:983100:d=2021032312:WDIR:surface:93 hour fcst: -6: 7:1364079:d=2021032312:UGRD:surface:93 hour fcst: -6: 8:1600580:d=2021032312:VGRD:surface:93 hour fcst: -6: 9:1840043:d=2021032312:ICEC:surface:93 hour fcst: -6: 10:1901187:d=2021032312:HTSGW:surface:93 hour fcst: -6: 11:2056798:d=2021032312:IMWF:surface:93 hour fcst: -6: 12:2227190:d=2021032312:MWSPER:surface:93 hour fcst: -6: 13:2397700:d=2021032312:PERPW:surface:93 hour fcst: -6: 14:2581245:d=2021032312:WWSDIR:surface:93 hour fcst: -6: 15:2912856:d=2021032312:DIRPW:surface:93 hour fcst: -6: 16:3252945:d=2021032312:WVHGT:surface:93 hour fcst: -6: 17:3408718:d=2021032312:SWELL:1 in sequence:93 hour fcst: -6: 18:3581677:d=2021032312:SWELL:2 in sequence:93 hour fcst: -6: 19:3709561:d=2021032312:SWELL:3 in sequence:93 hour fcst: -6: 20:3788782:d=2021032312:WVPER:surface:93 hour fcst: -6: 21:3962592:d=2021032312:SWPER:1 in sequence:93 hour fcst: -6: 22:4157593:d=2021032312:SWPER:2 in sequence:93 hour fcst: -6: 23:4320241:d=2021032312:SWPER:3 in sequence:93 hour fcst: -6: 24:4418794:d=2021032312:WVDIR:surface:93 hour fcst: -6: 25:4690452:d=2021032312:SWDIR:1 in sequence:93 hour fcst: -6: 26:5038343:d=2021032312:SWDIR:2 in sequence:93 hour fcst: -6: 27:5312826:d=2021032312:SWDIR:3 in sequence:93 hour fcst: -6: + wave_grib2_sbs.sh[104]err=0 -6: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.gsouth.0p25.f093.grib2 -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f093.grib2 ]] -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f093.grib2.idx ]] -6: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.gsouth.0p25.f093.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f093.grib2 -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f093.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f093.grib2 = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f093.grib2 ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f093.grib2 -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f093.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f093.grib2.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f093.grib2.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f093.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f093.grib2 -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.gsouth.0p25.f093.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f093.grib2.idx -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f093.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f093.grib2.idx = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f093.grib2.idx ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f093.grib2.idx -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f093.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f093.grib2.idx.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f093.grib2.idx.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f093.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f093.grib2.idx -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.gsouth.0p25.f093.grib2 and gfs.wave.t12z.gsouth.0p25.f093.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_gsh_15m to COM' -6: INFO: Copied gfs.wave.t12z.gsouth.0p25.f093.grib2 and gfs.wave.t12z.gsouth.0p25.f093.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_gsh_15m to COM -6: + wave_grib2_sbs.sh[130][[ gsh_15m == '' ]] -6: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -6: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.gsouth.0p25.f093.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -6: INFO: gfs.wave.t12z.gsouth.0p25.f093.grib2 is global.0p50 or SENDDBN is NO, no alert sent -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + cmdfile.1[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032709 3600. 9999 -0: + cmdfile.1[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grid_interp_sbs.sh[25]grdID=at_10m -0: + wave_grid_interp_sbs.sh[26]valid_time=2021032709 -0: + wave_grid_interp_sbs.sh[27]dt=3600. -0: + wave_grid_interp_sbs.sh[28]nst=9999 -0: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676 -0: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.uglo_100km ./out_grd.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mod_def.uglo_100km ./mod_def.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mod_def.at_10m ./mod_def.at_10m -0: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ]] -0: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m'\''' -0: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m' -0: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[51]weights_found=1 -0: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 090000' -0: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 090000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/at_10m/g ww3_gint.inp.tmpl -0: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -0: $ Input file for interpolation of uglo_100km to at_10m -0: $------------------------------------------------ -0: $ Start Time 3600. NSteps -0: 20210327 090000 3600. 9999 -0: $ Total number of grids -0: 2 -0: $ Grid extensions -0: 'uglo_100km' -0: 'at_10m' -0: $ -0: 0 -0: $ -0: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[70]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.315371 ']' -0: ++ prep_step[4]echo gfs_ww3_gint.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -0: INFO: Executing 'gfs_ww3_gint.x' -0: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[73]cat grid_interp.at_10m.out -0: -0: *** WAVEWATCH III Grid interpolation *** -0: =============================================== -0: -0: Comment character is '$' -0: -0: Time Information : -0: --------------------------------------------- -0: Starting Time : 2021/03/27 09:00:00 UTC -0: Interval (in sec) : 3600.00 -0: Number of requests : 9999 -0: --------------------------------------------- -0: Number of grids (including output grid) = 2 -0: -0: -0: Extension for grid 1 is --> uglo_100km -0: -0: Grid Particulars are : -0: Dimensions = 45166 1 -0: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Extension for grid 2 is --> at_10m -0: -0: Grid Particulars are : -0: Dimensions = 301 331 -0: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -0: -0: -0: Preparing interpolation weights for output grid -0: Total number of wet points for interpolation 29591 -0: -0: -0: Variable: Grid Interpolation Map Units: 0.100E+01 -0: -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 331 | | -0: | | -0: | | -0: 289 | | -0: | | -0: | 0 0 0 | -0: 247 | 0 0 0 0 | -0: | 0 0 0 0 0 | -0: | 0 0 0 0 | -0: 205 | 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 163 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 121 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 79 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 | -0: | 0 0 | -0: 37 | 0 | -0: | | -0: | | -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: -0: -0: Interpolating fields .... -0: -0: Output group 1 -0: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -0: Output group 2 -0: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -0: Output group 3 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 4 -0: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -0: Output group 5 -0: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -0: Output group 6 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 7 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 8 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 9 -0: Output variables skipped -0: Output group 10 -0: Output variables skipped -0: ------------------------------------------------ -0: 1Current vel. -0: 1Wind speed -0: 1Ice concentration -0: 2Wave height -0: 2Mean wave period(+2) -0: 2Mean wave period(+1) -0: 2Peak frequency -0: 2Mean wave dir. a1b1 -0: 2Peak direction -0: 4Part. wave height -0: 4Part. peak period -0: 4Part. mean direction -0: 5Charnock parameter -0: ------------------------------------------------ -0: OUTPUT TIME : 2021/03/27 09:00:00 UTC -0: -0: End of file reached -0: -0: -0: *** End of Grid interpolation Routine *** -0: =============================================== -0: -0: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -0: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -0: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grid_interp_at_10m/out_grd.at_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.at_10m -0: + cmdfile.1[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032709 93 atlocn 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + cmdfile.1[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib2_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grib2_sbs.sh[30]grdID=at_10m -0: + wave_grib2_sbs.sh[31]GRIDNR=255 -0: + wave_grib2_sbs.sh[32]MODNR=11 -0: + wave_grib2_sbs.sh[33]valid_time=2021032709 -0: + wave_grib2_sbs.sh[34]fhr=93 -0: + wave_grib2_sbs.sh[35]grid_region=atlocn -0: + wave_grib2_sbs.sh[36]grid_res=0p16 -0: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676 -0: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_at_10m -0: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_at_10m -0: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_at_10m -0: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_at_10m -0: ++ wave_grib2_sbs.sh[47]printf %03i 93 -0: + wave_grib2_sbs.sh[47]FH3=093 -0: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_atlocn_0p16 -0: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.atlocn.0p16.f093.grib2 -0: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f093.grib2 ]] -0: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/mod_def.at_10m ./mod_def.ww3 -0: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/out_grd.at_10m ./out_grd.ww3 -0: + wave_grib2_sbs.sh[73]ngrib=1 -0: + wave_grib2_sbs.sh[74]dtgrib=3600 -0: + wave_grib2_sbs.sh[75]tstart='20210327 090000' -0: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 090000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.at_10m.inp.tmpl -0: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -0: $ WAVEWATCH-III gridded output input file -0: $ ---------------------------------------- -0: 20210327 090000 3600 1 -0: N -0: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -0: $ -0: 20210327 090000 7 11 255 0 0 -0: $ -0: $ end of input file -0: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[88]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.315371 ']' -0: ++ prep_step[4]echo gfs_ww3_grib.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[90]export err=0 -0: + wave_grib2_sbs.sh[90]err=0 -0: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[95]cat grib2_atlocn_093.out -0: -0: *** WAVEWATCH III GRIB output postp. *** -0: ============================================== -0: -0: Comment character is '$' -0: -0: Grid name : NW Atlantic 10 min wave grid -0: -0: LINEIN: -0: 20210327 090000 3600 1 -0: -0: 20210327090000 3600 1 -0: GEN_PRO -99999 -0: -0: Output time data : -0: ----------------------------------------------------- -0: First time : 2021/03/27 09:00:00 UTC -0: Interval : 01:00:00 -0: Number of requests : 1 -0: Fields : Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: Requested output fields not yet available: -0: ----------------------------------------------------- -0: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -0: -0: Successfully requested output fields : -0: ----------------------------------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: -0: Additional GRIB parameters : -0: ----------------------------------------------------- -0: Run time : 2021/03/27 09:00:00 UTC -0: GRIB center ID : 7 -0: GRIB gen. proc. ID : 11 -0: GRIB grid ID : 255 -0: GRIB GDS parameter : 0 -0: Fields in file : -0: -------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: CHOSEN GRID TYPE: : LLRECTILINEAR -0: -0: -0: -0: Generating file -0: ----------------------------------------------------- -0: Data for 2021/03/27 09:00:00 UTC 0H forecast. -0: -0: End of program -0: ========================================= -0: WAVEWATCH III GRIB output -0: -0: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -0: + wave_grib2_sbs.sh[102][[ 93 -gt 0 ]] -0: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '93 hour fcst' -grib gfs.wave.t12z.atlocn.0p16.f093.grib2 -0: 1:0:d=2021032312:SPC:surface:93 hour fcst: -0: 2:20203:d=2021032312:DIRC:surface:93 hour fcst: -0: 3:60659:d=2021032312:UOGRD:surface:93 hour fcst: -0: 4:80533:d=2021032312:VOGRD:surface:93 hour fcst: -0: 5:100354:d=2021032312:WIND:surface:93 hour fcst: -0: 6:131399:d=2021032312:WDIR:surface:93 hour fcst: -0: 7:171513:d=2021032312:UGRD:surface:93 hour fcst: -0: 8:201933:d=2021032312:VGRD:surface:93 hour fcst: -0: 9:232038:d=2021032312:ICEC:surface:93 hour fcst: -0: 10:244673:d=2021032312:HTSGW:surface:93 hour fcst: -0: 11:270173:d=2021032312:IMWF:surface:93 hour fcst: -0: 12:296637:d=2021032312:MWSPER:surface:93 hour fcst: -0: 13:323347:d=2021032312:PERPW:surface:93 hour fcst: -0: 14:351319:d=2021032312:WWSDIR:surface:93 hour fcst: -0: 15:391177:d=2021032312:DIRPW:surface:93 hour fcst: -0: 16:432223:d=2021032312:WVHGT:surface:93 hour fcst: -0: 17:457160:d=2021032312:SWELL:1 in sequence:93 hour fcst: -0: 18:478183:d=2021032312:SWELL:2 in sequence:93 hour fcst: -0: 19:493849:d=2021032312:SWELL:3 in sequence:93 hour fcst: -0: 20:507119:d=2021032312:WVPER:surface:93 hour fcst: -0: 21:534112:d=2021032312:SWPER:1 in sequence:93 hour fcst: -0: 22:556919:d=2021032312:SWPER:2 in sequence:93 hour fcst: -0: 23:574328:d=2021032312:SWPER:3 in sequence:93 hour fcst: -0: 24:588129:d=2021032312:WVDIR:surface:93 hour fcst: -0: 25:625862:d=2021032312:SWDIR:1 in sequence:93 hour fcst: -0: 26:658331:d=2021032312:SWDIR:2 in sequence:93 hour fcst: -0: 27:679814:d=2021032312:SWDIR:3 in sequence:93 hour fcst: -0: + wave_grib2_sbs.sh[104]err=0 -0: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.atlocn.0p16.f093.grib2 -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f093.grib2 ]] -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f093.grib2.idx ]] -0: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.atlocn.0p16.f093.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f093.grib2 -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f093.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f093.grib2 = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f093.grib2 ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f093.grib2 -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f093.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f093.grib2.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f093.grib2.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f093.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f093.grib2 -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.atlocn.0p16.f093.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f093.grib2.idx -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f093.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f093.grib2.idx = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f093.grib2.idx ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f093.grib2.idx -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f093.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f093.grib2.idx.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f093.grib2.idx.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f093.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f093.grib2.idx -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.atlocn.0p16.f093.grib2 and gfs.wave.t12z.atlocn.0p16.f093.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_at_10m to COM' -0: INFO: Copied gfs.wave.t12z.atlocn.0p16.f093.grib2 and gfs.wave.t12z.atlocn.0p16.f093.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676/grib_at_10m to COM -0: + wave_grib2_sbs.sh[130][[ at_10m == '' ]] -0: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -0: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.atlocn.0p16.f093.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -0: INFO: gfs.wave.t12z.atlocn.0p16.f093.grib2 is global.0p50 or SENDDBN is NO, no alert sent -+ run_mpmd.sh[113]exit 0 -+ run_mpmd.sh[1]postamble run_mpmd.sh 1753758278 0 -+ preamble.sh[62]set +x -End run_mpmd.sh at 03:04:49 with error code 0 (time elapsed: 00:00:11) -+ exgfs_wave_post_gridded_sbs.sh[122]true -+ exgfs_wave_post_gridded_sbs.sh[123]export err=0 -+ exgfs_wave_post_gridded_sbs.sh[123]err=0 -+ exgfs_wave_post_gridded_sbs.sh[124][[ 0 -ne 0 ]] -+ exgfs_wave_post_gridded_sbs.sh[130]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ exgfs_wave_post_gridded_sbs.sh[131]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ exgfs_wave_post_gridded_sbs.sh[132]gribchk=gfs.wave.t12z.gsouth.0p25.f093.grib2 -+ exgfs_wave_post_gridded_sbs.sh[133][[ ! -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f093.grib2 ]] -+ exgfs_wave_post_gridded_sbs.sh[138]exit 0 -+ JGLOBAL_WAVE_POST_SBS[28]true -+ JGLOBAL_WAVE_POST_SBS[29]export err=0 -+ JGLOBAL_WAVE_POST_SBS[29]err=0 -+ JGLOBAL_WAVE_POST_SBS[30][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[37]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312 -+ JGLOBAL_WAVE_POST_SBS[38][[ NO != \Y\E\S ]] -+ JGLOBAL_WAVE_POST_SBS[39]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f093.314676 -+ JGLOBAL_WAVE_POST_SBS[42]exit 0 -+ JGLOBAL_WAVE_POST_SBS[1]postamble /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS 1753758266 0 -+ preamble.sh[62]set +x -End /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at 03:04:49 with error code 0 (time elapsed: 00:00:23) -Begin /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at Tue Jul 29 03:04:49 UTC 2025 -++ jjob_header.sh[46]OPTIND=1 -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[50]env_job=wavepostsbs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[49]read -ra configs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[61]shift 4 -++ jjob_header.sh[63][[ -z wavepostsbs ]] -++ jjob_header.sh[71]export DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676 -++ jjob_header.sh[71]DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676 -++ jjob_header.sh[72][[ YES == \Y\E\S ]] -++ jjob_header.sh[73]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676 -++ jjob_header.sh[75]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676 -++ jjob_header.sh[76]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676 -++ jjob_header.sh[85]export pid=318362 -++ jjob_header.sh[85]pid=318362 -++ jjob_header.sh[86]export pgmout=OUTPUT.318362 -++ jjob_header.sh[86]pgmout=OUTPUT.318362 -++ jjob_header.sh[87]export pgmerr=errfile -++ jjob_header.sh[87]pgmerr=errfile -++ jjob_header.sh[90]export pgm= -++ jjob_header.sh[90]pgm= -++ jjob_header.sh[96]export cycle=t12z -++ jjob_header.sh[96]cycle=t12z -++ jjob_header.sh[97]setpdy.sh -+ setpdy.sh[20]'[' /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676 == /home/mterry ']' -+ setpdy.sh[25][[ ! t12z =~ t??z ]] -+ setpdy.sh[30]case $# in -+ setpdy.sh[31]dates_before_PDY=7 -+ setpdy.sh[32]dates_after_PDY=7 -+ setpdy.sh[50]COMDATEROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+ setpdy.sh[53]'[' -z 20210323 ']' -+ setpdy.sh[57]sed 's/[0-9]\{8\}/20210323/' /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z -sed: can't read /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z: No such file or directory -++ jjob_header.sh[97]true -++ jjob_header.sh[98]source ./PDY -/work2/noaa/global/mterry/global-workflow_forked/ush/jjob_header.sh: line 98: ./PDY: No such file or directory -++ jjob_header.sh[98]true -++ jjob_header.sh[104]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[104]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.base -+++ config.base[6]echo 'BEGIN: config.base' -BEGIN: config.base -+++ config.base[9]export machine=HERCULES -+++ config.base[9]machine=HERCULES -+++ config.base[12]export RUN_ENVIR=emc -+++ config.base[12]RUN_ENVIR=emc -+++ config.base[15]export ACCOUNT=fv3-cpu -+++ config.base[15]ACCOUNT=fv3-cpu -+++ config.base[16]export QUEUE=batch -+++ config.base[16]QUEUE=batch -+++ config.base[17]export QUEUE_SERVICE=batch -+++ config.base[17]QUEUE_SERVICE=batch -+++ config.base[18]export QUEUE_DTN=batch -+++ config.base[18]QUEUE_DTN=batch -+++ config.base[19]export PARTITION_BATCH=hercules -+++ config.base[19]PARTITION_BATCH=hercules -+++ config.base[20]export PARTITION_SERVICE=service -+++ config.base[20]PARTITION_SERVICE=service -+++ config.base[21]export PARTITION_DTN= -+++ config.base[21]PARTITION_DTN= -+++ config.base[22]export RESERVATION= -+++ config.base[22]RESERVATION= -+++ config.base[23]export CLUSTERS= -+++ config.base[23]CLUSTERS= -+++ config.base[24]export CLUSTERS_SERVICE= -+++ config.base[24]CLUSTERS_SERVICE= -+++ config.base[25]export CLUSTERS_DTN= -+++ config.base[25]CLUSTERS_DTN= -+++ config.base[28]export HPSS_PROJECT=emc-global -+++ config.base[28]HPSS_PROJECT=emc-global -+++ config.base[31]export HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[31]HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[32]export EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[32]EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[33]export FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[33]FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[34]export PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[34]PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[35]export SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[35]SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[36]export USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[36]USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[38]export FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[38]FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[39]export FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[39]FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[40]export FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[40]FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[41]export FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[41]FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[42]export FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[42]FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[43]export FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[43]FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[44]export FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[44]FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[45]export FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[45]FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[50]export PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[50]PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[51]export COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[51]COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[52]export COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[52]COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[53]export DMPDIR=/work/noaa/rstprod/dump -+++ config.base[53]DMPDIR=/work/noaa/rstprod/dump -+++ config.base[57]export COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[57]COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[58]export COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[58]COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[59]export COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[59]COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[62]export HOMEDIR=/work2/noaa/global/mterry -+++ config.base[62]HOMEDIR=/work2/noaa/global/mterry -+++ config.base[63]export STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[63]STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]export PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[65]export NOSCRUB=/work2/noaa/global/mterry -+++ config.base[65]NOSCRUB=/work2/noaa/global/mterry -+++ config.base[68]export BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[68]BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[71]export BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[71]BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[74]export DO_PREP_SFC=NO -+++ config.base[74]DO_PREP_SFC=NO -+++ config.base[77]export DO_GOES=NO -+++ config.base[77]DO_GOES=NO -+++ config.base[78]export DO_BUFRSND=NO -+++ config.base[78]DO_BUFRSND=NO -+++ config.base[79]export DO_GEMPAK=NO -+++ config.base[79]DO_GEMPAK=NO -+++ config.base[80]export DO_AWIPS=NO -+++ config.base[80]DO_AWIPS=NO -+++ config.base[81]export DO_NPOESS=NO -+++ config.base[81]DO_NPOESS=NO -+++ config.base[82]export DO_TRACKER=YES -+++ config.base[82]DO_TRACKER=YES -+++ config.base[83]export DO_GENESIS=YES -+++ config.base[83]DO_GENESIS=YES -+++ config.base[84]export DO_GENESIS_FSU=NO -+++ config.base[84]DO_GENESIS_FSU=NO -+++ config.base[85]export DO_VERFOZN=YES -+++ config.base[85]DO_VERFOZN=YES -+++ config.base[86]export DO_VERFRAD=YES -+++ config.base[86]DO_VERFRAD=YES -+++ config.base[87]export DO_VMINMON=YES -+++ config.base[87]DO_VMINMON=YES -+++ config.base[88]export DO_ANLSTAT=NO -+++ config.base[88]DO_ANLSTAT=NO -+++ config.base[91]export MODE=forecast-only -+++ config.base[91]MODE=forecast-only -+++ config.base[92]export DO_TEST_MODE=YES -+++ config.base[92]DO_TEST_MODE=YES -+++ config.base[101]export FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[101]FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[102]export HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[102]HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[103]export HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[103]HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[106]export NMV=/bin/mv -+++ config.base[106]NMV=/bin/mv -+++ config.base[107]export 'NLN=/bin/ln -sf' -+++ config.base[107]NLN='/bin/ln -sf' -+++ config.base[108]export VERBOSE=YES -+++ config.base[108]VERBOSE=YES -+++ config.base[109]export KEEPDATA=NO -+++ config.base[109]KEEPDATA=NO -+++ config.base[110]export DEBUG_POSTSCRIPT=NO -+++ config.base[110]DEBUG_POSTSCRIPT=NO -+++ config.base[111]export CHGRP_RSTPROD=YES -+++ config.base[111]CHGRP_RSTPROD=YES -+++ config.base[112]export 'CHGRP_CMD=chgrp rstprod' -+++ config.base[112]CHGRP_CMD='chgrp rstprod' -+++ config.base[113]export NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[113]NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[114]export NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[114]NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[117]export BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[117]BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[120]export SDATE=2021032312 -+++ config.base[120]SDATE=2021032312 -+++ config.base[121]export EDATE=2021032312 -+++ config.base[121]EDATE=2021032312 -+++ config.base[122]export EXP_WARM_START=.false. -+++ config.base[122]EXP_WARM_START=.false. -+++ config.base[123]export assim_freq=6 -+++ config.base[123]assim_freq=6 -+++ config.base[124]export PSLOT=C48_S2SW -+++ config.base[124]PSLOT=C48_S2SW -+++ config.base[125]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[125]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[126]export ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[126]ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[127]export DUMP_SUFFIX= -+++ config.base[127]DUMP_SUFFIX= -+++ config.base[128][[ 2021032312 -ge 2019092100 ]] -+++ config.base[128][[ 2021032312 -le 2019110700 ]] -+++ config.base[131]export ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[131]ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[132]export ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[132]ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[133]export FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[133]FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[136]export envir=prod -+++ config.base[136]envir=prod -+++ config.base[137]export NET=gfs -+++ config.base[137]NET=gfs -+++ config.base[138]export RUN=gfs -+++ config.base[138]RUN=gfs -+++ config.base[141]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.com -++++ config.com[4]echo 'BEGIN: config.com' -BEGIN: config.com -++++ config.com[38][[ emc == \n\c\o ]] -++++ config.com[43]COM_OBSPROC_TMPL='${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos' -++++ config.com[44]COM_RTOFS_TMPL='${DMPDIR}' -++++ config.com[45]COM_TCVITAL_TMPL='${DMPDIR}/${RUN}.${YMD}/${HH}/atmos' -++++ config.com[47]declare -rx 'COM_OBS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/obs' -++++ config.com[48]declare -rx COM_OBSPROC_TMPL COM_RTOFS_TMPL -++++ config.com[50]COM_BASE='${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}' -++++ config.com[52]declare -rx 'COM_TOP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}' -++++ config.com[54]declare -rx 'COM_CONF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf' -++++ config.com[55]declare -rx 'COM_OBS_JEDI=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi' -++++ config.com[57]declare -rx 'COM_ATMOS_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input' -++++ config.com[58]declare -rx 'COM_ATMOS_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart' -++++ config.com[59]declare -rx 'COM_ATMOS_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos' -++++ config.com[60]declare -rx 'COM_SNOW_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow' -++++ config.com[61]declare -rx 'COM_SNOW_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon' -++++ config.com[62]declare -rx 'COM_ATMOS_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history' -++++ config.com[63]declare -rx 'COM_ATMOS_MASTER_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master' -++++ config.com[64]declare -rx 'COM_ATMOS_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2' -++++ config.com[65]declare -rx 'COM_ATMOS_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}' -++++ config.com[66]declare -rx 'COM_ATMOS_BUFR_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr' -++++ config.com[67]declare -rx 'COM_ATMOS_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}' -++++ config.com[68]declare -rx 'COM_ATMOS_GENESIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital' -++++ config.com[69]declare -rx 'COM_ATMOS_TRACK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks' -++++ config.com[70]declare -rx 'COM_ATMOS_GOES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim' -++++ config.com[71]declare -rx 'COM_ATMOS_IMAGERY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery' -++++ config.com[72]declare -rx 'COM_ATMOS_OZNMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon' -++++ config.com[73]declare -rx 'COM_ATMOS_RADMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon' -++++ config.com[74]declare -rx 'COM_ATMOS_MINMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon' -++++ config.com[75]declare -rx 'COM_ATMOS_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon' -++++ config.com[76]declare -rx 'COM_ATMOS_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo' -++++ config.com[78]declare -rx 'COM_WAVE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart' -++++ config.com[79]declare -rx 'COM_WAVE_PREP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep' -++++ config.com[80]declare -rx 'COM_WAVE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history' -++++ config.com[81]declare -rx 'COM_WAVE_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded' -++++ config.com[82]declare -rx 'COM_WAVE_GRID_RES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}' -++++ config.com[83]declare -rx 'COM_WAVE_STATION_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station' -++++ config.com[84]declare -rx 'COM_WAVE_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak' -++++ config.com[85]declare -rx 'COM_WAVE_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo' -++++ config.com[87]declare -rx 'COM_OCEAN_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history' -++++ config.com[88]declare -rx 'COM_OCEAN_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart' -++++ config.com[89]declare -rx 'COM_OCEAN_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input' -++++ config.com[90]declare -rx 'COM_OCEAN_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean' -++++ config.com[91]declare -rx 'COM_OCEAN_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon' -++++ config.com[92]declare -rx 'COM_OCEAN_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf' -++++ config.com[93]declare -rx 'COM_OCEAN_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean' -++++ config.com[94]declare -rx 'COM_OCEAN_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf' -++++ config.com[95]declare -rx 'COM_OCEAN_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2' -++++ config.com[96]declare -rx 'COM_OCEAN_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}' -++++ config.com[98]declare -rx 'COM_ICE_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice' -++++ config.com[99]declare -rx 'COM_ICE_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf' -++++ config.com[100]declare -rx 'COM_ICE_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon' -++++ config.com[101]declare -rx 'COM_ICE_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice' -++++ config.com[102]declare -rx 'COM_ICE_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input' -++++ config.com[103]declare -rx 'COM_ICE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history' -++++ config.com[104]declare -rx 'COM_ICE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart' -++++ config.com[105]declare -rx 'COM_ICE_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf' -++++ config.com[106]declare -rx 'COM_ICE_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2' -++++ config.com[107]declare -rx 'COM_ICE_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}' -++++ config.com[109]declare -rx 'COM_CHEM_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history' -++++ config.com[110]declare -rx 'COM_CHEM_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem' -++++ config.com[111]declare -rx 'COM_CHEM_BMAT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix' -++++ config.com[112]declare -rx 'COM_CHEM_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon' -++++ config.com[114]declare -rx 'COM_MED_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart' -+++ config.base[143]export LOGSCRIPT= -+++ config.base[143]LOGSCRIPT= -+++ config.base[145]export 'REDOUT=1>' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wave -+++ config.wave[6]echo 'BEGIN: config.wave' -BEGIN: config.wave -+++ config.wave[13]export RUNRSTwave=gdas -+++ config.wave[13]RUNRSTwave=gdas -+++ config.wave[16]export MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[16]MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[19]case "${waveGRD}" in -+++ config.wave[64]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[64]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[65]export wavepostGRD= -+++ config.wave[65]wavepostGRD= -+++ config.wave[66]export waveuoutpGRD=uglo_100km -+++ config.wave[66]waveuoutpGRD=uglo_100km -+++ config.wave[75]export WAVEWND_DID= -+++ config.wave[75]WAVEWND_DID= -+++ config.wave[76]export WAVEWND_FID= -+++ config.wave[76]WAVEWND_FID= -+++ config.wave[79][[ gfs == \g\f\s ]] -+++ config.wave[80]export FHMAX_WAV=120 -+++ config.wave[80]FHMAX_WAV=120 -+++ config.wave[82]export WAVHINDH=0 -+++ config.wave[82]WAVHINDH=0 -+++ config.wave[83]export FHMAX_WAV_IBP=180 -+++ config.wave[83]FHMAX_WAV_IBP=180 -+++ config.wave[84](( FHMAX_WAV < FHMAX_WAV_IBP )) -+++ config.wave[84]export FHMAX_WAV_IBP=120 -+++ config.wave[84]FHMAX_WAV_IBP=120 -+++ config.wave[87]export DTFLD_WAV=3600 -+++ config.wave[87]DTFLD_WAV=3600 -+++ config.wave[88]export DTPNT_WAV=3600 -+++ config.wave[88]DTPNT_WAV=3600 -+++ config.wave[89]export FHINCP_WAV=1 -+++ config.wave[89]FHINCP_WAV=1 -+++ config.wave[92]export 'OUTPARS_WAV=WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[92]OUTPARS_WAV='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[95][[ gfs == \g\d\a\s ]] -+++ config.wave[99](( INTERVAL_GFS > 0 )) -+++ config.wave[100]export WAVHCYC=6 -+++ config.wave[100]WAVHCYC=6 -+++ config.wave[101]export FHMAX_WAV_CUR=192 -+++ config.wave[101]FHMAX_WAV_CUR=192 -+++ config.wave[109]export RSTTYPE_WAV=T -+++ config.wave[109]RSTTYPE_WAV=T -+++ config.wave[110][[ gfs != gfs ]] -+++ config.wave[115]rst_dt_gfs=43200 -+++ config.wave[116][[ 43200 -gt 0 ]] -+++ config.wave[117]export DT_1_RST_WAV=0 -+++ config.wave[117]DT_1_RST_WAV=0 -+++ config.wave[120]export DT_2_RST_WAV=43200 -+++ config.wave[120]DT_2_RST_WAV=43200 -+++ config.wave[126]export RSTIOFF_WAV=0 -+++ config.wave[126]RSTIOFF_WAV=0 -+++ config.wave[131]export RUNMEM=-1 -+++ config.wave[131]RUNMEM=-1 -+++ config.wave[134](( RUNMEM == -1 )) -+++ config.wave[136]export waveMEMB= -+++ config.wave[136]waveMEMB= -+++ config.wave[143]export WW3ATMINP=CPL -+++ config.wave[143]WW3ATMINP=CPL -+++ config.wave[144][[ YES == \Y\E\S ]] -+++ config.wave[145]export WW3ICEINP=CPL -+++ config.wave[145]WW3ICEINP=CPL -+++ config.wave[146]export WAVEICE_FID= -+++ config.wave[146]WAVEICE_FID= -+++ config.wave[152][[ YES == \Y\E\S ]] -+++ config.wave[153]export WW3CURINP=CPL -+++ config.wave[153]WW3CURINP=CPL -+++ config.wave[154]export WAVECUR_FID= -+++ config.wave[154]WAVECUR_FID= -+++ config.wave[161]export WW3ATMIENS=F -+++ config.wave[161]WW3ATMIENS=F -+++ config.wave[162]export WW3ICEIENS=F -+++ config.wave[162]WW3ICEIENS=F -+++ config.wave[163]export WW3CURIENS=F -+++ config.wave[163]WW3CURIENS=F -+++ config.wave[165]export GOFILETYPE=1 -+++ config.wave[165]GOFILETYPE=1 -+++ config.wave[166]export POFILETYPE=1 -+++ config.wave[166]POFILETYPE=1 -+++ config.wave[170]export FUNIPNT=T -+++ config.wave[170]FUNIPNT=T -+++ config.wave[172]export IOSRV=1 -+++ config.wave[172]IOSRV=1 -+++ config.wave[174]export FPNTPROC=T -+++ config.wave[174]FPNTPROC=T -+++ config.wave[176]export FGRDPROC=F -+++ config.wave[176]FGRDPROC=F -+++ config.wave[178]export FLAGMASKCOMP=F -+++ config.wave[178]FLAGMASKCOMP=F -+++ config.wave[180]export FLAGMASKOUT=F -+++ config.wave[180]FLAGMASKOUT=F -+++ config.wave[182]echo 'END: config.wave' -END: config.wave -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wavepostsbs -+++ config.wavepostsbs[6]echo 'BEGIN: config.wavepostsbs' -BEGIN: config.wavepostsbs -+++ config.wavepostsbs[9]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources wavepostsbs -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=wavepostsbs -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[156]ntasks=8 -++++ config.resources[157]threads_per_task=1 -++++ config.resources[158]tasks_per_node=80 -++++ config.resources[159]NTASKS=8 -++++ config.resources[160]memory=20GB -++++ config.resources[161]walltime=00:15:00 -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export NTASKS -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export memory -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.wavepostsbs[12]export MAX_TASKS=25 -+++ config.wavepostsbs[12]MAX_TASKS=25 -+++ config.wavepostsbs[15]export WAV_SUBGRBSRC= -+++ config.wavepostsbs[15]WAV_SUBGRBSRC= -+++ config.wavepostsbs[16]export WAV_SUBGRB= -+++ config.wavepostsbs[16]WAV_SUBGRB= -+++ config.wavepostsbs[19]export DOFLD_WAV=YES -+++ config.wavepostsbs[19]DOFLD_WAV=YES -+++ config.wavepostsbs[20]export DOPNT_WAV=YES -+++ config.wavepostsbs[20]DOPNT_WAV=YES -+++ config.wavepostsbs[21]export DOGRB_WAV=YES -+++ config.wavepostsbs[21]DOGRB_WAV=YES -+++ config.wavepostsbs[22][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+++ config.wavepostsbs[23]export DOGRI_WAV=YES -+++ config.wavepostsbs[23]DOGRI_WAV=YES -+++ config.wavepostsbs[27]export DOSPC_WAV=YES -+++ config.wavepostsbs[27]DOSPC_WAV=YES -+++ config.wavepostsbs[28]export DOBLL_WAV=YES -+++ config.wavepostsbs[28]DOBLL_WAV=YES -+++ config.wavepostsbs[30]echo 'END: config.wavepostsbs' -END: config.wavepostsbs -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env wavepostsbs -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=wavepostsbs -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 20971520 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 8 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[34]max_threads_per_task=1 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 1 ]] -+++ HERCULES.env[40][[ 1 -gt 1 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 8' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[63]export USE_CFP=YES -+++ HERCULES.env[63]USE_CFP=YES -+++ HERCULES.env[64][[ wavepostsbs == \w\a\v\e\p\r\e\p ]] -+++ HERCULES.env[65]export 'wavempexec=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[65]wavempexec='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[66]export 'wave_mpmd=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[66]wave_mpmd='--multi-prog --output=mpmd.%j.%t.out' -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[4]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ JGLOBAL_WAVE_POST_SBS[7]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[7]HH=12 -+ JGLOBAL_WAVE_POST_SBS[7]declare_from_tmpl -rx COMIN_WAVE_PREP:COM_WAVE_PREP_TMPL COMIN_WAVE_HISTORY:COM_WAVE_HISTORY_TMPL COMOUT_WAVE_GRID:COM_WAVE_GRID_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_WAVE_PREP=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep -declare_from_tmpl :: COMIN_WAVE_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history -declare_from_tmpl :: COMOUT_WAVE_GRID=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded -+ JGLOBAL_WAVE_POST_SBS[12][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n '' ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_atlocn_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_atlocn_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_atlocn_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_epacif_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_epacif_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_epacif_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_wcoast_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_wcoast_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_wcoast_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p50 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p50:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p50=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p50 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=arctic.9km -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_arctic_9km:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_arctic_9km=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_arctic_9km -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_gsouth_0p25:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_gsouth_0p25=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 ]] -+ JGLOBAL_WAVE_POST_SBS[28]/work2/noaa/global/mterry/global-workflow_forked/scripts/exgfs_wave_post_gridded_sbs.sh -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ exgfs_wave_post_gridded_sbs.sh[24]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ exgfs_wave_post_gridded_sbs.sh[26]DOGRI_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[27]DOGRB_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[29]export waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[29]waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[30]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[30]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[31]export wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[31]wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[33]cat - INFO: Grid information: - INFO: Native wave grids: uglo_100km - INFO: Interpolated grids: at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m - INFO: Post-process grids: -++ exgfs_wave_post_gridded_sbs.sh[40]printf %03i 96 -+ exgfs_wave_post_gridded_sbs.sh[40]fhr3=096 -++ exgfs_wave_post_gridded_sbs.sh[41]date -u -d '20210323 12 + 96 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[41]valid_time=2021032712 -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[49]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f096.bin ./out_grd.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f096.bin ./out_grd.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[52][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[62]cat - INFO: Summary: - INFO: Grid interp: DOGRI_WAV="YES" - INFO: Grib files: DOGRB_WAV="YES" - INFO: Fields to be included in grib files: - INFO: OUTPARS_WAV="WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA" -+ exgfs_wave_post_gridded_sbs.sh[70][[ YES == \N\O ]] -+ exgfs_wave_post_gridded_sbs.sh[76]rm -f 'cmdfile.*' cmdfile -+ exgfs_wave_post_gridded_sbs.sh[77]count=0 -+ exgfs_wave_post_gridded_sbs.sh[80][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[81]dt_int=3600. -+ exgfs_wave_post_gridded_sbs.sh[82]n_int=9999 -++ exgfs_wave_post_gridded_sbs.sh[83]date -u -d '20210327 12 - 0 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[83]ymdh_int=2021032712 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=1 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032712 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032712 96 atlocn 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib2_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib2_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=2 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032712 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032712 96 epacif 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib2_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib2_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=3 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032712 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032712 96 wcoast 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib2_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib2_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=4 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032712 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032712 96 global 0p50 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib2_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib2_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=5 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032712 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032712 96 arctic 9km '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib2_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib2_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=6 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032712 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032712 96 global 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib2_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib2_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=7 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032712 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032712 96 gsouth 0p25 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib2_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib2_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[100][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[113][[ 8 -lt 7 ]] -+ exgfs_wave_post_gridded_sbs.sh[121]echo 'INFO: Running MPMD job with 7 commands' -INFO: Running MPMD job with 7 commands -+ exgfs_wave_post_gridded_sbs.sh[122]/work2/noaa/global/mterry/global-workflow_forked/ush/run_mpmd.sh /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/cmdfile -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ run_mpmd.sh[31]source /work2/noaa/global/mterry/global-workflow_forked/ush/preamble.sh -++ preamble.sh[20]set +x -Begin run_mpmd.sh at Tue Jul 29 03:05:01 UTC 2025 -+ run_mpmd.sh[33]cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/cmdfile -+ run_mpmd.sh[36][[ YES != \Y\E\S ]] -+ run_mpmd.sh[46]export OMP_NUM_THREADS=1 -+ run_mpmd.sh[46]OMP_NUM_THREADS=1 -++ run_mpmd.sh[49]wc -l -+ run_mpmd.sh[49]nprocs=7 -+ run_mpmd.sh[52]mpmd_cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mpmd_cmdfile -+ run_mpmd.sh[53][[ -s /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mpmd_cmdfile ]] -+ run_mpmd.sh[55]cat - INFO: Executing MPMD job, STDOUT redirected for each process separately - INFO: On failure, logs for each job will be available in /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mpmd.proc_num.out - INFO: The proc_num corresponds to the line in '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mpmd_cmdfile' -+ run_mpmd.sh[61][[ srun -l --export=ALL --hint=nomultithread =~ ^srun.* ]] -+ run_mpmd.sh[65]nm=0 -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '0 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/cmdfile.1' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '1 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/cmdfile.2' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '2 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/cmdfile.3' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '3 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/cmdfile.4' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '4 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/cmdfile.5' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '5 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/cmdfile.6' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '6 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/cmdfile.7' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[72]set +e -+ run_mpmd.sh[74]srun -l --export=ALL --hint=nomultithread --multi-prog --output=mpmd.%j.%t.out -n 7 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mpmd_cmdfile -+ run_mpmd.sh[75]err=0 -+ run_mpmd.sh[76]set_strict -+ preamble.sh[35][[ YES == \Y\E\S ]] -+ preamble.sh[37]set -eu -+ run_mpmd.sh[103][[ 0 -eq 0 ]] -+ run_mpmd.sh[104]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mpmd_cmdfile -++ run_mpmd.sh[105]find . -name 'mpmd.*.out' -+ run_mpmd.sh[105]out_files='./mpmd.5951798.4.out -./mpmd.5951798.2.out -./mpmd.5951798.1.out -./mpmd.5951798.3.out -./mpmd.5951798.5.out -./mpmd.5951798.6.out -./mpmd.5951798.0.out' -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951798.4.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951798.4.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951798.2.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951798.2.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951798.1.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951798.1.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951798.3.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951798.3.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951798.5.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951798.5.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951798.6.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951798.6.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951798.0.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951798.0.out -+ run_mpmd.sh[110]cat mpmd.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + cmdfile.5[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032712 3600. 9999 -4: + cmdfile.5[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grid_interp_sbs.sh[25]grdID=aoc_9km -4: + wave_grid_interp_sbs.sh[26]valid_time=2021032712 -4: + wave_grid_interp_sbs.sh[27]dt=3600. -4: + wave_grid_interp_sbs.sh[28]nst=9999 -4: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676 -4: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.uglo_100km ./out_grd.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mod_def.uglo_100km ./mod_def.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mod_def.aoc_9km ./mod_def.aoc_9km -4: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km'\''' -4: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km' -4: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[51]weights_found=1 -4: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 120000' -4: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 120000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/aoc_9km/g ww3_gint.inp.tmpl -4: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -4: $ Input file for interpolation of uglo_100km to aoc_9km -4: $------------------------------------------------ -4: $ Start Time 3600. NSteps -4: 20210327 120000 3600. 9999 -4: $ Total number of grids -4: 2 -4: $ Grid extensions -4: 'uglo_100km' -4: 'aoc_9km' -4: $ -4: 0 -4: $ -4: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[70]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.318362 ']' -4: ++ prep_step[4]echo gfs_ww3_gint.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -4: INFO: Executing 'gfs_ww3_gint.x' -4: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[73]cat grid_interp.aoc_9km.out -4: -4: *** WAVEWATCH III Grid interpolation *** -4: =============================================== -4: -4: Comment character is '$' -4: -4: Time Information : -4: --------------------------------------------- -4: Starting Time : 2021/03/27 12:00:00 UTC -4: Interval (in sec) : 3600.00 -4: Number of requests : 9999 -4: --------------------------------------------- -4: Number of grids (including output grid) = 2 -4: -4: -4: Extension for grid 1 is --> uglo_100km -4: -4: Grid Particulars are : -4: Dimensions = 45166 1 -4: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Extension for grid 2 is --> aoc_9km -4: -4: Grid Particulars are : -4: Dimensions = 1006 1006 -4: Grid Type = 2 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -4: -4: -4: Preparing interpolation weights for output grid -4: Total number of wet points for interpolation 360052 -4: -4: -4: Variable: Grid Interpolation Map Units: 0.100E+01 -4: -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: *** | | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: 880 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 | -4: 754 | 0 0 0 | -4: | 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 628 | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 | -4: 502 | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 376 | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 250 | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 124 | 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 | -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: -4: -4: Interpolating fields .... -4: -4: Output group 1 -4: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -4: Output group 2 -4: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -4: Output group 3 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 4 -4: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -4: Output group 5 -4: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -4: Output group 6 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 7 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 8 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 9 -4: Output variables skipped -4: Output group 10 -4: Output variables skipped -4: ------------------------------------------------ -4: 1Current vel. -4: 1Wind speed -4: 1Ice concentration -4: 2Wave height -4: 2Mean wave period(+2) -4: 2Mean wave period(+1) -4: 2Peak frequency -4: 2Mean wave dir. a1b1 -4: 2Peak direction -4: 4Part. wave height -4: 4Part. peak period -4: 4Part. mean direction -4: 5Charnock parameter -4: ------------------------------------------------ -4: OUTPUT TIME : 2021/03/27 12:00:00 UTC -4: -4: End of file reached -4: -4: -4: *** End of Grid interpolation Routine *** -4: =============================================== -4: -4: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -4: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -4: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_aoc_9km/out_grd.aoc_9km /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.aoc_9km -4: + cmdfile.5[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032712 96 arctic 9km 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + cmdfile.5[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib2_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grib2_sbs.sh[30]grdID=aoc_9km -4: + wave_grib2_sbs.sh[31]GRIDNR=255 -4: + wave_grib2_sbs.sh[32]MODNR=11 -4: + wave_grib2_sbs.sh[33]valid_time=2021032712 -4: + wave_grib2_sbs.sh[34]fhr=96 -4: + wave_grib2_sbs.sh[35]grid_region=arctic -4: + wave_grib2_sbs.sh[36]grid_res=9km -4: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676 -4: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_aoc_9km -4: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_aoc_9km -4: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_aoc_9km -4: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_aoc_9km -4: ++ wave_grib2_sbs.sh[47]printf %03i 96 -4: + wave_grib2_sbs.sh[47]FH3=096 -4: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_arctic_9km -4: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.arctic.9km.f096.grib2 -4: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f096.grib2 ]] -4: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mod_def.aoc_9km ./mod_def.ww3 -4: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.aoc_9km ./out_grd.ww3 -4: + wave_grib2_sbs.sh[73]ngrib=1 -4: + wave_grib2_sbs.sh[74]dtgrib=3600 -4: + wave_grib2_sbs.sh[75]tstart='20210327 120000' -4: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 120000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.aoc_9km.inp.tmpl -4: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -4: $ WAVEWATCH-III gridded output input file -4: $ ---------------------------------------- -4: 20210327 120000 3600 1 -4: N -4: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -4: $ -4: 20210327 120000 7 11 255 0 20 -4: $ -4: 70 0 9.0 9.0 64 -4: $ 60 0 8.64919046313 8.64919046313 64 -4: $ end of input file -4: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[88]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.318362 ']' -4: ++ prep_step[4]echo gfs_ww3_grib.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[90]export err=0 -4: + wave_grib2_sbs.sh[90]err=0 -4: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[95]cat grib2_arctic_096.out -4: -4: *** WAVEWATCH III GRIB output postp. *** -4: ============================================== -4: -4: Comment character is '$' -4: -4: Grid name : Arctic Ocean PolarStereo 9km -4: -4: LINEIN: -4: 20210327 120000 3600 1 -4: -4: 20210327120000 3600 1 -4: GEN_PRO -99999 -4: -4: Output time data : -4: ----------------------------------------------------- -4: First time : 2021/03/27 12:00:00 UTC -4: Interval : 01:00:00 -4: Number of requests : 1 -4: Fields : Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: Requested output fields not yet available: -4: ----------------------------------------------------- -4: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -4: -4: Successfully requested output fields : -4: ----------------------------------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: -4: Additional GRIB parameters : -4: ----------------------------------------------------- -4: Run time : 2021/03/27 12:00:00 UTC -4: GRIB center ID : 7 -4: GRIB gen. proc. ID : 11 -4: GRIB grid ID : 255 -4: GRIB GDS parameter : 0 -4: Fields in file : -4: -------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: CHOSEN GRID TYPE: : POLARSTEREO -4: -4: -4: -4: Generating file -4: ----------------------------------------------------- -4: Data for 2021/03/27 12:00:00 UTC 0H forecast. -4: -4: End of program -4: ========================================= -4: WAVEWATCH III GRIB output -4: -4: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -4: + wave_grib2_sbs.sh[102][[ 96 -gt 0 ]] -4: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '96 hour fcst' -grib gfs.wave.t12z.arctic.9km.f096.grib2 -4: 1:0:d=2021032312:SPC:surface:96 hour fcst: -4: 2:190319:d=2021032312:DIRC:surface:96 hour fcst: -4: 3:661631:d=2021032312:UOGRD:surface:96 hour fcst: -4: 4:849679:d=2021032312:VOGRD:surface:96 hour fcst: -4: 5:1040632:d=2021032312:WIND:surface:96 hour fcst: -4: 6:1341941:d=2021032312:WDIR:surface:96 hour fcst: -4: 7:1773739:d=2021032312:UGRD:surface:96 hour fcst: -4: 8:2071951:d=2021032312:VGRD:surface:96 hour fcst: -4: 9:2372455:d=2021032312:ICEC:surface:96 hour fcst: -4: 10:2542337:d=2021032312:HTSGW:surface:96 hour fcst: -4: 11:2754634:d=2021032312:IMWF:surface:96 hour fcst: -4: 12:2972311:d=2021032312:MWSPER:surface:96 hour fcst: -4: 13:3191709:d=2021032312:PERPW:surface:96 hour fcst: -4: 14:3419479:d=2021032312:WWSDIR:surface:96 hour fcst: -4: 15:3724435:d=2021032312:DIRPW:surface:96 hour fcst: -4: 16:4036594:d=2021032312:WVHGT:surface:96 hour fcst: -4: 17:4242798:d=2021032312:SWELL:1 in sequence:96 hour fcst: -4: 18:4456152:d=2021032312:SWELL:2 in sequence:96 hour fcst: -4: 19:4620973:d=2021032312:SWELL:3 in sequence:96 hour fcst: -4: 20:4755288:d=2021032312:WVPER:surface:96 hour fcst: -4: 21:4972071:d=2021032312:SWPER:1 in sequence:96 hour fcst: -4: 22:5193890:d=2021032312:SWPER:2 in sequence:96 hour fcst: -4: 23:5371907:d=2021032312:SWPER:3 in sequence:96 hour fcst: -4: 24:5509882:d=2021032312:WVDIR:surface:96 hour fcst: -4: 25:5787787:d=2021032312:SWDIR:1 in sequence:96 hour fcst: -4: 26:6103766:d=2021032312:SWDIR:2 in sequence:96 hour fcst: -4: 27:6325571:d=2021032312:SWDIR:3 in sequence:96 hour fcst: -4: + wave_grib2_sbs.sh[104]err=0 -4: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.arctic.9km.f096.grib2 -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f096.grib2 ]] -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f096.grib2.idx ]] -4: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.arctic.9km.f096.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f096.grib2 -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f096.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f096.grib2 = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f096.grib2 ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f096.grib2 -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f096.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f096.grib2.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f096.grib2.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f096.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f096.grib2 -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.arctic.9km.f096.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f096.grib2.idx -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f096.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f096.grib2.idx = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f096.grib2.idx ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f096.grib2.idx -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f096.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f096.grib2.idx.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f096.grib2.idx.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f096.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f096.grib2.idx -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.arctic.9km.f096.grib2 and gfs.wave.t12z.arctic.9km.f096.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_aoc_9km to COM' -4: INFO: Copied gfs.wave.t12z.arctic.9km.f096.grib2 and gfs.wave.t12z.arctic.9km.f096.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_aoc_9km to COM -4: + wave_grib2_sbs.sh[130][[ aoc_9km == '' ]] -4: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -4: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.arctic.9km.f096.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -4: INFO: gfs.wave.t12z.arctic.9km.f096.grib2 is global.0p50 or SENDDBN is NO, no alert sent -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + cmdfile.3[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032712 3600. 9999 -2: + cmdfile.3[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grid_interp_sbs.sh[25]grdID=wc_10m -2: + wave_grid_interp_sbs.sh[26]valid_time=2021032712 -2: + wave_grid_interp_sbs.sh[27]dt=3600. -2: + wave_grid_interp_sbs.sh[28]nst=9999 -2: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676 -2: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.uglo_100km ./out_grd.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mod_def.uglo_100km ./mod_def.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mod_def.wc_10m ./mod_def.wc_10m -2: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ]] -2: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m'\''' -2: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m' -2: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[51]weights_found=1 -2: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 120000' -2: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 120000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/wc_10m/g ww3_gint.inp.tmpl -2: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -2: $ Input file for interpolation of uglo_100km to wc_10m -2: $------------------------------------------------ -2: $ Start Time 3600. NSteps -2: 20210327 120000 3600. 9999 -2: $ Total number of grids -2: 2 -2: $ Grid extensions -2: 'uglo_100km' -2: 'wc_10m' -2: $ -2: 0 -2: $ -2: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[70]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.318362 ']' -2: ++ prep_step[4]echo gfs_ww3_gint.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -2: INFO: Executing 'gfs_ww3_gint.x' -2: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[73]cat grid_interp.wc_10m.out -2: -2: *** WAVEWATCH III Grid interpolation *** -2: =============================================== -2: -2: Comment character is '$' -2: -2: Time Information : -2: --------------------------------------------- -2: Starting Time : 2021/03/27 12:00:00 UTC -2: Interval (in sec) : 3600.00 -2: Number of requests : 9999 -2: --------------------------------------------- -2: Number of grids (including output grid) = 2 -2: -2: -2: Extension for grid 1 is --> uglo_100km -2: -2: Grid Particulars are : -2: Dimensions = 45166 1 -2: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Extension for grid 2 is --> wc_10m -2: -2: Grid Particulars are : -2: Dimensions = 241 151 -2: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -2: -2: -2: Preparing interpolation weights for output grid -2: Total number of wet points for interpolation 11044 -2: -2: -2: Variable: Grid Interpolation Map Units: 0.100E+01 -2: -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: +---------------------------------------------------------------------------------------------------------------+ -2: 151 | 0 0 0 0 | -2: | 0 0 0 0 0 | -2: | 0 0 0 0 0 | -2: 130 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 109 | 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 88 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 | -2: 67 | 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 | -2: 46 | 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: 25 | 0 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 0 | -2: | | -2: 4 | | -2: +---------------------------------------------------------------------------------------------------------------+ -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: -2: -2: Interpolating fields .... -2: -2: Output group 1 -2: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -2: Output group 2 -2: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -2: Output group 3 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 4 -2: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -2: Output group 5 -2: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -2: Output group 6 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 7 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 8 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 9 -2: Output variables skipped -2: Output group 10 -2: Output variables skipped -2: ------------------------------------------------ -2: 1Current vel. -2: 1Wind speed -2: 1Ice concentration -2: 2Wave height -2: 2Mean wave period(+2) -2: 2Mean wave period(+1) -2: 2Peak frequency -2: 2Mean wave dir. a1b1 -2: 2Peak direction -2: 4Part. wave height -2: 4Part. peak period -2: 4Part. mean direction -2: 5Charnock parameter -2: ------------------------------------------------ -2: OUTPUT TIME : 2021/03/27 12:00:00 UTC -2: -2: End of file reached -2: -2: -2: *** End of Grid interpolation Routine *** -2: =============================================== -2: -2: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -2: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -2: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_wc_10m/out_grd.wc_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.wc_10m -2: + cmdfile.3[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032712 96 wcoast 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + cmdfile.3[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib2_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grib2_sbs.sh[30]grdID=wc_10m -2: + wave_grib2_sbs.sh[31]GRIDNR=255 -2: + wave_grib2_sbs.sh[32]MODNR=11 -2: + wave_grib2_sbs.sh[33]valid_time=2021032712 -2: + wave_grib2_sbs.sh[34]fhr=96 -2: + wave_grib2_sbs.sh[35]grid_region=wcoast -2: + wave_grib2_sbs.sh[36]grid_res=0p16 -2: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676 -2: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_wc_10m -2: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_wc_10m -2: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_wc_10m -2: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_wc_10m -2: ++ wave_grib2_sbs.sh[47]printf %03i 96 -2: + wave_grib2_sbs.sh[47]FH3=096 -2: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_wcoast_0p16 -2: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.wcoast.0p16.f096.grib2 -2: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f096.grib2 ]] -2: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mod_def.wc_10m ./mod_def.ww3 -2: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.wc_10m ./out_grd.ww3 -2: + wave_grib2_sbs.sh[73]ngrib=1 -2: + wave_grib2_sbs.sh[74]dtgrib=3600 -2: + wave_grib2_sbs.sh[75]tstart='20210327 120000' -2: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 120000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.wc_10m.inp.tmpl -2: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -2: $ WAVEWATCH-III gridded output input file -2: $ ---------------------------------------- -2: 20210327 120000 3600 1 -2: N -2: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -2: $ -2: 20210327 120000 7 11 255 0 0 -2: $ -2: $ end of input file -2: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[88]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.318362 ']' -2: ++ prep_step[4]echo gfs_ww3_grib.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[90]export err=0 -2: + wave_grib2_sbs.sh[90]err=0 -2: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[95]cat grib2_wcoast_096.out -2: -2: *** WAVEWATCH III GRIB output postp. *** -2: ============================================== -2: -2: Comment character is '$' -2: -2: Grid name : West Coast 10 min wave grid -2: -2: LINEIN: -2: 20210327 120000 3600 1 -2: -2: 20210327120000 3600 1 -2: GEN_PRO -99999 -2: -2: Output time data : -2: ----------------------------------------------------- -2: First time : 2021/03/27 12:00:00 UTC -2: Interval : 01:00:00 -2: Number of requests : 1 -2: Fields : Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: Requested output fields not yet available: -2: ----------------------------------------------------- -2: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -2: -2: Successfully requested output fields : -2: ----------------------------------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: -2: Additional GRIB parameters : -2: ----------------------------------------------------- -2: Run time : 2021/03/27 12:00:00 UTC -2: GRIB center ID : 7 -2: GRIB gen. proc. ID : 11 -2: GRIB grid ID : 255 -2: GRIB GDS parameter : 0 -2: Fields in file : -2: -------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: CHOSEN GRID TYPE: : LLRECTILINEAR -2: -2: -2: -2: Generating file -2: ----------------------------------------------------- -2: Data for 2021/03/27 12:00:00 UTC 0H forecast. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: -2: End of program -2: ========================================= -2: WAVEWATCH III GRIB output -2: -2: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -2: + wave_grib2_sbs.sh[102][[ 96 -gt 0 ]] -2: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '96 hour fcst' -grib gfs.wave.t12z.wcoast.0p16.f096.grib2 -2: 1:0:d=2021032312:SPC:surface:96 hour fcst: -2: 2:7474:d=2021032312:DIRC:surface:96 hour fcst: -2: 3:25190:d=2021032312:UOGRD:surface:96 hour fcst: -2: 4:32224:d=2021032312:VOGRD:surface:96 hour fcst: -2: 5:39873:d=2021032312:WIND:surface:96 hour fcst: -2: 6:51553:d=2021032312:WDIR:surface:96 hour fcst: -2: 7:68300:d=2021032312:UGRD:surface:96 hour fcst: -2: 8:79410:d=2021032312:VGRD:surface:96 hour fcst: -2: 9:91208:d=2021032312:ICEC:surface:96 hour fcst: -2: 10:95938:d=2021032312:HTSGW:surface:96 hour fcst: -2: 11:105451:d=2021032312:IMWF:surface:96 hour fcst: -2: 12:115106:d=2021032312:MWSPER:surface:96 hour fcst: -2: 13:124821:d=2021032312:PERPW:surface:96 hour fcst: -2: 14:134826:d=2021032312:WWSDIR:surface:96 hour fcst: -2: 15:150239:d=2021032312:DIRPW:surface:96 hour fcst: -2: 16:165992:d=2021032312:WVHGT:surface:96 hour fcst: -2: 17:173955:d=2021032312:SWELL:1 in sequence:96 hour fcst: -2: 18:182016:d=2021032312:SWELL:2 in sequence:96 hour fcst: -2: 19:187245:d=2021032312:SWELL:3 in sequence:96 hour fcst: -2: 20:191975:d=2021032312:WVPER:surface:96 hour fcst: -2: 21:200405:d=2021032312:SWPER:1 in sequence:96 hour fcst: -2: 22:208512:d=2021032312:SWPER:2 in sequence:96 hour fcst: -2: 23:213901:d=2021032312:SWPER:3 in sequence:96 hour fcst: -2: 24:218631:d=2021032312:WVDIR:surface:96 hour fcst: -2: 25:230285:d=2021032312:SWDIR:1 in sequence:96 hour fcst: -2: 26:241843:d=2021032312:SWDIR:2 in sequence:96 hour fcst: -2: 27:247625:d=2021032312:SWDIR:3 in sequence:96 hour fcst: -2: + wave_grib2_sbs.sh[104]err=0 -2: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.wcoast.0p16.f096.grib2 -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f096.grib2 ]] -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f096.grib2.idx ]] -2: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.wcoast.0p16.f096.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f096.grib2 -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f096.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f096.grib2 = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f096.grib2 ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f096.grib2 -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f096.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f096.grib2.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f096.grib2.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f096.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f096.grib2 -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.wcoast.0p16.f096.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f096.grib2.idx -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f096.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f096.grib2.idx = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f096.grib2.idx ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f096.grib2.idx -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f096.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f096.grib2.idx.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f096.grib2.idx.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f096.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f096.grib2.idx -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.wcoast.0p16.f096.grib2 and gfs.wave.t12z.wcoast.0p16.f096.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_wc_10m to COM' -2: INFO: Copied gfs.wave.t12z.wcoast.0p16.f096.grib2 and gfs.wave.t12z.wcoast.0p16.f096.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_wc_10m to COM -2: + wave_grib2_sbs.sh[130][[ wc_10m == '' ]] -2: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -2: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.wcoast.0p16.f096.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -2: INFO: gfs.wave.t12z.wcoast.0p16.f096.grib2 is global.0p50 or SENDDBN is NO, no alert sent -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + cmdfile.2[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032712 3600. 9999 -1: + cmdfile.2[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grid_interp_sbs.sh[25]grdID=ep_10m -1: + wave_grid_interp_sbs.sh[26]valid_time=2021032712 -1: + wave_grid_interp_sbs.sh[27]dt=3600. -1: + wave_grid_interp_sbs.sh[28]nst=9999 -1: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676 -1: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.uglo_100km ./out_grd.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mod_def.uglo_100km ./mod_def.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mod_def.ep_10m ./mod_def.ep_10m -1: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ]] -1: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m'\''' -1: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m' -1: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[51]weights_found=1 -1: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 120000' -1: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 120000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/ep_10m/g ww3_gint.inp.tmpl -1: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -1: $ Input file for interpolation of uglo_100km to ep_10m -1: $------------------------------------------------ -1: $ Start Time 3600. NSteps -1: 20210327 120000 3600. 9999 -1: $ Total number of grids -1: 2 -1: $ Grid extensions -1: 'uglo_100km' -1: 'ep_10m' -1: $ -1: 0 -1: $ -1: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[70]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.318362 ']' -1: ++ prep_step[4]echo gfs_ww3_gint.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]env -1: +++ prep_step[12]awk -F= '{print $1}' -1: ++ prep_step[12]unset FORT01 -1: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -1: INFO: Executing 'gfs_ww3_gint.x' -1: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[73]cat grid_interp.ep_10m.out -1: -1: *** WAVEWATCH III Grid interpolation *** -1: =============================================== -1: -1: Comment character is '$' -1: -1: Time Information : -1: --------------------------------------------- -1: Starting Time : 2021/03/27 12:00:00 UTC -1: Interval (in sec) : 3600.00 -1: Number of requests : 9999 -1: --------------------------------------------- -1: Number of grids (including output grid) = 2 -1: -1: -1: Extension for grid 1 is --> uglo_100km -1: -1: Grid Particulars are : -1: Dimensions = 45166 1 -1: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Extension for grid 2 is --> ep_10m -1: -1: Grid Particulars are : -1: Dimensions = 511 301 -1: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -1: -1: -1: Preparing interpolation weights for output grid -1: Total number of wet points for interpolation 7439 -1: -1: -1: Variable: Grid Interpolation Map Units: 0.100E+01 -1: -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 301 | | -1: | | -1: | 0 0 | -1: 262 | 0 0 0 0 | -1: | 0 0 0 0 | -1: | 0 0 0 0 | -1: 223 | 0 0 0 | -1: | 0 | -1: | 0 | -1: 184 | | -1: | 0 0 0 | -1: | 0 0 0 0 | -1: 145 | 0 | -1: | | -1: | | -1: 106 | | -1: | | -1: | | -1: 67 | | -1: | 0 | -1: | | -1: 28 | 0 | -1: | | -1: | | -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: -1: -1: Interpolating fields .... -1: -1: Output group 1 -1: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -1: Output group 2 -1: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -1: Output group 3 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 4 -1: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -1: Output group 5 -1: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -1: Output group 6 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 7 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 8 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 9 -1: Output variables skipped -1: Output group 10 -1: Output variables skipped -1: ------------------------------------------------ -1: 1Current vel. -1: 1Wind speed -1: 1Ice concentration -1: 2Wave height -1: 2Mean wave period(+2) -1: 2Mean wave period(+1) -1: 2Peak frequency -1: 2Mean wave dir. a1b1 -1: 2Peak direction -1: 4Part. wave height -1: 4Part. peak period -1: 4Part. mean direction -1: 5Charnock parameter -1: ------------------------------------------------ -1: OUTPUT TIME : 2021/03/27 12:00:00 UTC -1: -1: End of file reached -1: -1: -1: *** End of Grid interpolation Routine *** -1: =============================================== -1: -1: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -1: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -1: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_ep_10m/out_grd.ep_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.ep_10m -1: + cmdfile.2[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032712 96 epacif 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + cmdfile.2[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib2_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grib2_sbs.sh[30]grdID=ep_10m -1: + wave_grib2_sbs.sh[31]GRIDNR=255 -1: + wave_grib2_sbs.sh[32]MODNR=11 -1: + wave_grib2_sbs.sh[33]valid_time=2021032712 -1: + wave_grib2_sbs.sh[34]fhr=96 -1: + wave_grib2_sbs.sh[35]grid_region=epacif -1: + wave_grib2_sbs.sh[36]grid_res=0p16 -1: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676 -1: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_ep_10m -1: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_ep_10m -1: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_ep_10m -1: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_ep_10m -1: ++ wave_grib2_sbs.sh[47]printf %03i 96 -1: + wave_grib2_sbs.sh[47]FH3=096 -1: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_epacif_0p16 -1: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.epacif.0p16.f096.grib2 -1: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f096.grib2 ]] -1: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mod_def.ep_10m ./mod_def.ww3 -1: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.ep_10m ./out_grd.ww3 -1: + wave_grib2_sbs.sh[73]ngrib=1 -1: + wave_grib2_sbs.sh[74]dtgrib=3600 -1: + wave_grib2_sbs.sh[75]tstart='20210327 120000' -1: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 120000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.ep_10m.inp.tmpl -1: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -1: $ WAVEWATCH-III gridded output input file -1: $ ---------------------------------------- -1: 20210327 120000 3600 1 -1: N -1: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -1: $ -1: 20210327 120000 7 11 255 0 0 -1: $ -1: $ end of input file -1: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[88]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.318362 ']' -1: ++ prep_step[4]echo gfs_ww3_grib.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[90]export err=0 -1: + wave_grib2_sbs.sh[90]err=0 -1: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[95]cat grib2_epacif_096.out -1: -1: *** WAVEWATCH III GRIB output postp. *** -1: ============================================== -1: -1: Comment character is '$' -1: -1: Grid name : East Pacific 10 min wave grid -1: -1: LINEIN: -1: 20210327 120000 3600 1 -1: -1: 20210327120000 3600 1 -1: GEN_PRO -99999 -1: -1: Output time data : -1: ----------------------------------------------------- -1: First time : 2021/03/27 12:00:00 UTC -1: Interval : 01:00:00 -1: Number of requests : 1 -1: Fields : Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: Requested output fields not yet available: -1: ----------------------------------------------------- -1: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -1: -1: Successfully requested output fields : -1: ----------------------------------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: -1: Additional GRIB parameters : -1: ----------------------------------------------------- -1: Run time : 2021/03/27 12:00:00 UTC -1: GRIB center ID : 7 -1: GRIB gen. proc. ID : 11 -1: GRIB grid ID : 255 -1: GRIB GDS parameter : 0 -1: Fields in file : -1: -------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: CHOSEN GRID TYPE: : LLRECTILINEAR -1: -1: -1: -1: Generating file -1: ----------------------------------------------------- -1: Data for 2021/03/27 12:00:00 UTC 0H forecast. -1: Warning: bitmask off everywhere. -1: Pretend one point in jpcpack to avoid crash. -1: Warning: bitmask off everywhere. -1: Pretend one point in jpcpack to avoid crash. -1: Warning: bitmask off everywhere. -1: Pretend one point in jpcpack to avoid crash. -1: -1: End of program -1: ========================================= -1: WAVEWATCH III GRIB output -1: -1: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -1: + wave_grib2_sbs.sh[102][[ 96 -gt 0 ]] -1: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '96 hour fcst' -grib gfs.wave.t12z.epacif.0p16.f096.grib2 -1: 1:0:d=2021032312:SPC:surface:96 hour fcst: -1: 2:21808:d=2021032312:DIRC:surface:96 hour fcst: -1: 3:49903:d=2021032312:UOGRD:surface:96 hour fcst: -1: 4:71742:d=2021032312:VOGRD:surface:96 hour fcst: -1: 5:93674:d=2021032312:WIND:surface:96 hour fcst: -1: 6:118274:d=2021032312:WDIR:surface:96 hour fcst: -1: 7:145439:d=2021032312:UGRD:surface:96 hour fcst: -1: 8:170024:d=2021032312:VGRD:surface:96 hour fcst: -1: 9:194594:d=2021032312:ICEC:surface:96 hour fcst: -1: 10:214002:d=2021032312:HTSGW:surface:96 hour fcst: -1: 11:236441:d=2021032312:IMWF:surface:96 hour fcst: -1: 12:259354:d=2021032312:MWSPER:surface:96 hour fcst: -1: 13:282258:d=2021032312:PERPW:surface:96 hour fcst: -1: 14:305093:d=2021032312:WWSDIR:surface:96 hour fcst: -1: 15:331010:d=2021032312:DIRPW:surface:96 hour fcst: -1: 16:357060:d=2021032312:WVHGT:surface:96 hour fcst: -1: 17:378994:d=2021032312:SWELL:1 in sequence:96 hour fcst: -1: 18:401821:d=2021032312:SWELL:2 in sequence:96 hour fcst: -1: 19:422214:d=2021032312:SWELL:3 in sequence:96 hour fcst: -1: 20:441622:d=2021032312:WVPER:surface:96 hour fcst: -1: 21:463539:d=2021032312:SWPER:1 in sequence:96 hour fcst: -1: 22:486365:d=2021032312:SWPER:2 in sequence:96 hour fcst: -1: 23:506742:d=2021032312:SWPER:3 in sequence:96 hour fcst: -1: 24:526150:d=2021032312:WVDIR:surface:96 hour fcst: -1: 25:550221:d=2021032312:SWDIR:1 in sequence:96 hour fcst: -1: 26:577204:d=2021032312:SWDIR:2 in sequence:96 hour fcst: -1: 27:598484:d=2021032312:SWDIR:3 in sequence:96 hour fcst: -1: + wave_grib2_sbs.sh[104]err=0 -1: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.epacif.0p16.f096.grib2 -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f096.grib2 ]] -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f096.grib2.idx ]] -1: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.epacif.0p16.f096.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f096.grib2 -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f096.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f096.grib2 = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f096.grib2 ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f096.grib2 -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f096.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f096.grib2.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f096.grib2.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f096.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f096.grib2 -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.epacif.0p16.f096.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f096.grib2.idx -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f096.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f096.grib2.idx = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f096.grib2.idx ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f096.grib2.idx -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f096.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f096.grib2.idx.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f096.grib2.idx.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f096.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f096.grib2.idx -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.epacif.0p16.f096.grib2 and gfs.wave.t12z.epacif.0p16.f096.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_ep_10m to COM' -1: INFO: Copied gfs.wave.t12z.epacif.0p16.f096.grib2 and gfs.wave.t12z.epacif.0p16.f096.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_ep_10m to COM -1: + wave_grib2_sbs.sh[130][[ ep_10m == '' ]] -1: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -1: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.epacif.0p16.f096.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -1: INFO: gfs.wave.t12z.epacif.0p16.f096.grib2 is global.0p50 or SENDDBN is NO, no alert sent -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + cmdfile.4[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032712 3600. 9999 -3: + cmdfile.4[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grid_interp_sbs.sh[25]grdID=glo_30m -3: + wave_grid_interp_sbs.sh[26]valid_time=2021032712 -3: + wave_grid_interp_sbs.sh[27]dt=3600. -3: + wave_grid_interp_sbs.sh[28]nst=9999 -3: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676 -3: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.uglo_100km ./out_grd.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mod_def.uglo_100km ./mod_def.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mod_def.glo_30m ./mod_def.glo_30m -3: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ]] -3: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m'\''' -3: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m' -3: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[51]weights_found=1 -3: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 120000' -3: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 120000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/glo_30m/g ww3_gint.inp.tmpl -3: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -3: $ Input file for interpolation of uglo_100km to glo_30m -3: $------------------------------------------------ -3: $ Start Time 3600. NSteps -3: 20210327 120000 3600. 9999 -3: $ Total number of grids -3: 2 -3: $ Grid extensions -3: 'uglo_100km' -3: 'glo_30m' -3: $ -3: 0 -3: $ -3: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[70]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.318362 ']' -3: ++ prep_step[4]echo gfs_ww3_gint.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -3: INFO: Executing 'gfs_ww3_gint.x' -3: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[73]cat grid_interp.glo_30m.out -3: -3: *** WAVEWATCH III Grid interpolation *** -3: =============================================== -3: -3: Comment character is '$' -3: -3: Time Information : -3: --------------------------------------------- -3: Starting Time : 2021/03/27 12:00:00 UTC -3: Interval (in sec) : 3600.00 -3: Number of requests : 9999 -3: --------------------------------------------- -3: Number of grids (including output grid) = 2 -3: -3: -3: Extension for grid 1 is --> uglo_100km -3: -3: Grid Particulars are : -3: Dimensions = 45166 1 -3: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Extension for grid 2 is --> glo_30m -3: -3: Grid Particulars are : -3: Dimensions = 720 336 -3: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -3: -3: -3: Preparing interpolation weights for output grid -3: Total number of wet points for interpolation 167619 -3: -3: -3: Variable: Grid Interpolation Map Units: 0.100E+01 -3: -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 336 | | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 291 | 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 | -3: 246 | 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 201 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 156 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 111 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 66 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 21 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 | -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: -3: -3: Interpolating fields .... -3: -3: Output group 1 -3: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -3: Output group 2 -3: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -3: Output group 3 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 4 -3: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -3: Output group 5 -3: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -3: Output group 6 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 7 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 8 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 9 -3: Output variables skipped -3: Output group 10 -3: Output variables skipped -3: ------------------------------------------------ -3: 1Current vel. -3: 1Wind speed -3: 1Ice concentration -3: 2Wave height -3: 2Mean wave period(+2) -3: 2Mean wave period(+1) -3: 2Peak frequency -3: 2Mean wave dir. a1b1 -3: 2Peak direction -3: 4Part. wave height -3: 4Part. peak period -3: 4Part. mean direction -3: 5Charnock parameter -3: ------------------------------------------------ -3: OUTPUT TIME : 2021/03/27 12:00:00 UTC -3: -3: End of file reached -3: -3: -3: *** End of Grid interpolation Routine *** -3: =============================================== -3: -3: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -3: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -3: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_glo_30m/out_grd.glo_30m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.glo_30m -3: + cmdfile.4[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032712 96 global 0p50 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + cmdfile.4[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib2_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grib2_sbs.sh[30]grdID=glo_30m -3: + wave_grib2_sbs.sh[31]GRIDNR=255 -3: + wave_grib2_sbs.sh[32]MODNR=11 -3: + wave_grib2_sbs.sh[33]valid_time=2021032712 -3: + wave_grib2_sbs.sh[34]fhr=96 -3: + wave_grib2_sbs.sh[35]grid_region=global -3: + wave_grib2_sbs.sh[36]grid_res=0p50 -3: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676 -3: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_glo_30m -3: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_glo_30m -3: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_glo_30m -3: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_glo_30m -3: ++ wave_grib2_sbs.sh[47]printf %03i 96 -3: + wave_grib2_sbs.sh[47]FH3=096 -3: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p50 -3: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p50.f096.grib2 -3: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f096.grib2 ]] -3: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mod_def.glo_30m ./mod_def.ww3 -3: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.glo_30m ./out_grd.ww3 -3: + wave_grib2_sbs.sh[73]ngrib=1 -3: + wave_grib2_sbs.sh[74]dtgrib=3600 -3: + wave_grib2_sbs.sh[75]tstart='20210327 120000' -3: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 120000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.glo_30m.inp.tmpl -3: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -3: $ WAVEWATCH-III gridded output input file -3: $ ---------------------------------------- -3: 20210327 120000 3600 1 -3: N -3: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -3: $ -3: 20210327 120000 7 11 255 0 0 -3: $ -3: $ end of input file -3: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[88]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.318362 ']' -3: ++ prep_step[4]echo gfs_ww3_grib.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[90]export err=0 -3: + wave_grib2_sbs.sh[90]err=0 -3: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[95]cat grib2_global_096.out -3: -3: *** WAVEWATCH III GRIB output postp. *** -3: ============================================== -3: -3: Comment character is '$' -3: -3: Grid name : Global 30 min wave grid -3: -3: LINEIN: -3: 20210327 120000 3600 1 -3: -3: 20210327120000 3600 1 -3: GEN_PRO -99999 -3: -3: Output time data : -3: ----------------------------------------------------- -3: First time : 2021/03/27 12:00:00 UTC -3: Interval : 01:00:00 -3: Number of requests : 1 -3: Fields : Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: Requested output fields not yet available: -3: ----------------------------------------------------- -3: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -3: -3: Successfully requested output fields : -3: ----------------------------------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: -3: Additional GRIB parameters : -3: ----------------------------------------------------- -3: Run time : 2021/03/27 12:00:00 UTC -3: GRIB center ID : 7 -3: GRIB gen. proc. ID : 11 -3: GRIB grid ID : 255 -3: GRIB GDS parameter : 0 -3: Fields in file : -3: -------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: CHOSEN GRID TYPE: : LLRECTILINEAR -3: -3: -3: -3: Generating file -3: ----------------------------------------------------- -3: Data for 2021/03/27 12:00:00 UTC 0H forecast. -3: -3: End of program -3: ========================================= -3: WAVEWATCH III GRIB output -3: -3: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -3: + wave_grib2_sbs.sh[102][[ 96 -gt 0 ]] -3: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '96 hour fcst' -grib gfs.wave.t12z.global.0p50.f096.grib2 -3: 1:0:d=2021032312:SPC:surface:96 hour fcst: -3: 2:77706:d=2021032312:DIRC:surface:96 hour fcst: -3: 3:322228:d=2021032312:UOGRD:surface:96 hour fcst: -3: 4:397651:d=2021032312:VOGRD:surface:96 hour fcst: -3: 5:477047:d=2021032312:WIND:surface:96 hour fcst: -3: 6:635289:d=2021032312:WDIR:surface:96 hour fcst: -3: 7:871592:d=2021032312:UGRD:surface:96 hour fcst: -3: 8:1026061:d=2021032312:VGRD:surface:96 hour fcst: -3: 9:1182585:d=2021032312:ICEC:surface:96 hour fcst: -3: 10:1222780:d=2021032312:HTSGW:surface:96 hour fcst: -3: 11:1321569:d=2021032312:IMWF:surface:96 hour fcst: -3: 12:1431042:d=2021032312:MWSPER:surface:96 hour fcst: -3: 13:1541402:d=2021032312:PERPW:surface:96 hour fcst: -3: 14:1658354:d=2021032312:WWSDIR:surface:96 hour fcst: -3: 15:1858707:d=2021032312:DIRPW:surface:96 hour fcst: -3: 16:2065073:d=2021032312:WVHGT:surface:96 hour fcst: -3: 17:2159971:d=2021032312:SWELL:1 in sequence:96 hour fcst: -3: 18:2264219:d=2021032312:SWELL:2 in sequence:96 hour fcst: -3: 19:2336125:d=2021032312:SWELL:3 in sequence:96 hour fcst: -3: 20:2379544:d=2021032312:WVPER:surface:96 hour fcst: -3: 21:2485741:d=2021032312:SWPER:1 in sequence:96 hour fcst: -3: 22:2603585:d=2021032312:SWPER:2 in sequence:96 hour fcst: -3: 23:2694007:d=2021032312:SWPER:3 in sequence:96 hour fcst: -3: 24:2745613:d=2021032312:WVDIR:surface:96 hour fcst: -3: 25:2906712:d=2021032312:SWDIR:1 in sequence:96 hour fcst: -3: 26:3108606:d=2021032312:SWDIR:2 in sequence:96 hour fcst: -3: 27:3251202:d=2021032312:SWDIR:3 in sequence:96 hour fcst: -3: + wave_grib2_sbs.sh[104]err=0 -3: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p50.f096.grib2 -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f096.grib2 ]] -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f096.grib2.idx ]] -3: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p50.f096.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f096.grib2 -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f096.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f096.grib2 = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f096.grib2 ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f096.grib2 -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f096.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f096.grib2.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f096.grib2.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f096.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f096.grib2 -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p50.f096.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f096.grib2.idx -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f096.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f096.grib2.idx = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f096.grib2.idx ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f096.grib2.idx -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f096.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f096.grib2.idx.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f096.grib2.idx.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f096.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f096.grib2.idx -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p50.f096.grib2 and gfs.wave.t12z.global.0p50.f096.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_glo_30m to COM' -3: INFO: Copied gfs.wave.t12z.global.0p50.f096.grib2 and gfs.wave.t12z.global.0p50.f096.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_glo_30m to COM -3: + wave_grib2_sbs.sh[130][[ glo_30m == '' ]] -3: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -3: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p50.f096.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -3: INFO: gfs.wave.t12z.global.0p50.f096.grib2 is global.0p50 or SENDDBN is NO, no alert sent -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + cmdfile.6[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032712 3600. 9999 -5: + cmdfile.6[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grid_interp_sbs.sh[25]grdID=gnh_10m -5: + wave_grid_interp_sbs.sh[26]valid_time=2021032712 -5: + wave_grid_interp_sbs.sh[27]dt=3600. -5: + wave_grid_interp_sbs.sh[28]nst=9999 -5: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676 -5: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.uglo_100km ./out_grd.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mod_def.uglo_100km ./mod_def.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mod_def.gnh_10m ./mod_def.gnh_10m -5: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m'\''' -5: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m' -5: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[51]weights_found=1 -5: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 120000' -5: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 120000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gnh_10m/g ww3_gint.inp.tmpl -5: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -5: $ Input file for interpolation of uglo_100km to gnh_10m -5: $------------------------------------------------ -5: $ Start Time 3600. NSteps -5: 20210327 120000 3600. 9999 -5: $ Total number of grids -5: 2 -5: $ Grid extensions -5: 'uglo_100km' -5: 'gnh_10m' -5: $ -5: 0 -5: $ -5: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[70]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.318362 ']' -5: ++ prep_step[4]echo gfs_ww3_gint.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -5: INFO: Executing 'gfs_ww3_gint.x' -5: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[73]cat grid_interp.gnh_10m.out -5: -5: *** WAVEWATCH III Grid interpolation *** -5: =============================================== -5: -5: Comment character is '$' -5: -5: Time Information : -5: --------------------------------------------- -5: Starting Time : 2021/03/27 12:00:00 UTC -5: Interval (in sec) : 3600.00 -5: Number of requests : 9999 -5: --------------------------------------------- -5: Number of grids (including output grid) = 2 -5: -5: -5: Extension for grid 1 is --> uglo_100km -5: -5: Grid Particulars are : -5: Dimensions = 45166 1 -5: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Extension for grid 2 is --> gnh_10m -5: -5: Grid Particulars are : -5: Dimensions = 2160 406 -5: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -5: -5: -5: Preparing interpolation weights for output grid -5: Total number of wet points for interpolation 571209 -5: -5: -5: Variable: Grid Interpolation Map Units: 0.100E+01 -5: -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 406 | | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: 355 | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 304 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 253 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 202 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 151 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 100 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 49 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: -5: -5: Interpolating fields .... -5: -5: Output group 1 -5: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -5: Output group 2 -5: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -5: Output group 3 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 4 -5: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -5: Output group 5 -5: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -5: Output group 6 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 7 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 8 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 9 -5: Output variables skipped -5: Output group 10 -5: Output variables skipped -5: ------------------------------------------------ -5: 1Current vel. -5: 1Wind speed -5: 1Ice concentration -5: 2Wave height -5: 2Mean wave period(+2) -5: 2Mean wave period(+1) -5: 2Peak frequency -5: 2Mean wave dir. a1b1 -5: 2Peak direction -5: 4Part. wave height -5: 4Part. peak period -5: 4Part. mean direction -5: 5Charnock parameter -5: ------------------------------------------------ -5: OUTPUT TIME : 2021/03/27 12:00:00 UTC -5: -5: End of file reached -5: -5: -5: *** End of Grid interpolation Routine *** -5: =============================================== -5: -5: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -5: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -5: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_gnh_10m/out_grd.gnh_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.gnh_10m -5: + cmdfile.6[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032712 96 global 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + cmdfile.6[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib2_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grib2_sbs.sh[30]grdID=gnh_10m -5: + wave_grib2_sbs.sh[31]GRIDNR=255 -5: + wave_grib2_sbs.sh[32]MODNR=11 -5: + wave_grib2_sbs.sh[33]valid_time=2021032712 -5: + wave_grib2_sbs.sh[34]fhr=96 -5: + wave_grib2_sbs.sh[35]grid_region=global -5: + wave_grib2_sbs.sh[36]grid_res=0p16 -5: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676 -5: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_gnh_10m -5: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_gnh_10m -5: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_gnh_10m -5: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_gnh_10m -5: ++ wave_grib2_sbs.sh[47]printf %03i 96 -5: + wave_grib2_sbs.sh[47]FH3=096 -5: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p16 -5: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p16.f096.grib2 -5: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f096.grib2 ]] -5: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mod_def.gnh_10m ./mod_def.ww3 -5: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.gnh_10m ./out_grd.ww3 -5: + wave_grib2_sbs.sh[73]ngrib=1 -5: + wave_grib2_sbs.sh[74]dtgrib=3600 -5: + wave_grib2_sbs.sh[75]tstart='20210327 120000' -5: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 120000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gnh_10m.inp.tmpl -5: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -5: $ WAVEWATCH-III gridded output input file -5: $ ---------------------------------------- -5: 20210327 120000 3600 1 -5: N -5: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -5: $ -5: 20210327 120000 7 11 255 0 0 -5: $ -5: $ end of input file -5: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[88]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.318362 ']' -5: ++ prep_step[4]echo gfs_ww3_grib.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[90]export err=0 -5: + wave_grib2_sbs.sh[90]err=0 -5: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[95]cat grib2_global_096.out -5: -5: *** WAVEWATCH III GRIB output postp. *** -5: ============================================== -5: -5: Comment character is '$' -5: -5: Grid name : GFSv16-wave N Hemisphere 1/6 d -5: -5: LINEIN: -5: 20210327 120000 3600 1 -5: -5: 20210327120000 3600 1 -5: GEN_PRO -99999 -5: -5: Output time data : -5: ----------------------------------------------------- -5: First time : 2021/03/27 12:00:00 UTC -5: Interval : 01:00:00 -5: Number of requests : 1 -5: Fields : Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: Requested output fields not yet available: -5: ----------------------------------------------------- -5: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -5: -5: Successfully requested output fields : -5: ----------------------------------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: -5: Additional GRIB parameters : -5: ----------------------------------------------------- -5: Run time : 2021/03/27 12:00:00 UTC -5: GRIB center ID : 7 -5: GRIB gen. proc. ID : 11 -5: GRIB grid ID : 255 -5: GRIB GDS parameter : 0 -5: Fields in file : -5: -------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: CHOSEN GRID TYPE: : LLRECTILINEAR -5: -5: -5: -5: Generating file -5: ----------------------------------------------------- -5: Data for 2021/03/27 12:00:00 UTC 0H forecast. -5: -5: End of program -5: ========================================= -5: WAVEWATCH III GRIB output -5: -5: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -5: + wave_grib2_sbs.sh[102][[ 96 -gt 0 ]] -5: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '96 hour fcst' -grib gfs.wave.t12z.global.0p16.f096.grib2 -5: 1:0:d=2021032312:SPC:surface:96 hour fcst: -5: 2:237558:d=2021032312:DIRC:surface:96 hour fcst: -5: 3:870203:d=2021032312:UOGRD:surface:96 hour fcst: -5: 4:1104159:d=2021032312:VOGRD:surface:96 hour fcst: -5: 5:1341742:d=2021032312:WIND:surface:96 hour fcst: -5: 6:1749379:d=2021032312:WDIR:surface:96 hour fcst: -5: 7:2385653:d=2021032312:UGRD:surface:96 hour fcst: -5: 8:2786220:d=2021032312:VGRD:surface:96 hour fcst: -5: 9:3187158:d=2021032312:ICEC:surface:96 hour fcst: -5: 10:3301697:d=2021032312:HTSGW:surface:96 hour fcst: -5: 11:3590488:d=2021032312:IMWF:surface:96 hour fcst: -5: 12:3902668:d=2021032312:MWSPER:surface:96 hour fcst: -5: 13:4216878:d=2021032312:PERPW:surface:96 hour fcst: -5: 14:4551526:d=2021032312:WWSDIR:surface:96 hour fcst: -5: 15:5131620:d=2021032312:DIRPW:surface:96 hour fcst: -5: 16:5737712:d=2021032312:WVHGT:surface:96 hour fcst: -5: 17:6013351:d=2021032312:SWELL:1 in sequence:96 hour fcst: -5: 18:6302767:d=2021032312:SWELL:2 in sequence:96 hour fcst: -5: 19:6496456:d=2021032312:SWELL:3 in sequence:96 hour fcst: -5: 20:6622470:d=2021032312:WVPER:surface:96 hour fcst: -5: 21:6931581:d=2021032312:SWPER:1 in sequence:96 hour fcst: -5: 22:7259993:d=2021032312:SWPER:2 in sequence:96 hour fcst: -5: 23:7497419:d=2021032312:SWPER:3 in sequence:96 hour fcst: -5: 24:7637898:d=2021032312:WVDIR:surface:96 hour fcst: -5: 25:8130217:d=2021032312:SWDIR:1 in sequence:96 hour fcst: -5: 26:8719377:d=2021032312:SWDIR:2 in sequence:96 hour fcst: -5: 27:9092617:d=2021032312:SWDIR:3 in sequence:96 hour fcst: -5: + wave_grib2_sbs.sh[104]err=0 -5: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p16.f096.grib2 -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f096.grib2 ]] -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f096.grib2.idx ]] -5: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p16.f096.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f096.grib2 -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f096.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f096.grib2 = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f096.grib2 ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f096.grib2 -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f096.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f096.grib2.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f096.grib2.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f096.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f096.grib2 -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p16.f096.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f096.grib2.idx -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f096.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f096.grib2.idx = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f096.grib2.idx ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f096.grib2.idx -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f096.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f096.grib2.idx.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f096.grib2.idx.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f096.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f096.grib2.idx -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p16.f096.grib2 and gfs.wave.t12z.global.0p16.f096.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_gnh_10m to COM' -5: INFO: Copied gfs.wave.t12z.global.0p16.f096.grib2 and gfs.wave.t12z.global.0p16.f096.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_gnh_10m to COM -5: + wave_grib2_sbs.sh[130][[ gnh_10m == '' ]] -5: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -5: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p16.f096.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -5: INFO: gfs.wave.t12z.global.0p16.f096.grib2 is global.0p50 or SENDDBN is NO, no alert sent -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + cmdfile.7[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032712 3600. 9999 -6: + cmdfile.7[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grid_interp_sbs.sh[25]grdID=gsh_15m -6: + wave_grid_interp_sbs.sh[26]valid_time=2021032712 -6: + wave_grid_interp_sbs.sh[27]dt=3600. -6: + wave_grid_interp_sbs.sh[28]nst=9999 -6: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676 -6: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.uglo_100km ./out_grd.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mod_def.uglo_100km ./mod_def.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mod_def.gsh_15m ./mod_def.gsh_15m -6: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m'\''' -6: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m' -6: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[51]weights_found=1 -6: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 120000' -6: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 120000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gsh_15m/g ww3_gint.inp.tmpl -6: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -6: $ Input file for interpolation of uglo_100km to gsh_15m -6: $------------------------------------------------ -6: $ Start Time 3600. NSteps -6: 20210327 120000 3600. 9999 -6: $ Total number of grids -6: 2 -6: $ Grid extensions -6: 'uglo_100km' -6: 'gsh_15m' -6: $ -6: 0 -6: $ -6: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[70]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.318362 ']' -6: ++ prep_step[4]echo gfs_ww3_gint.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -6: INFO: Executing 'gfs_ww3_gint.x' -6: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[73]cat grid_interp.gsh_15m.out -6: -6: *** WAVEWATCH III Grid interpolation *** -6: =============================================== -6: -6: Comment character is '$' -6: -6: Time Information : -6: --------------------------------------------- -6: Starting Time : 2021/03/27 12:00:00 UTC -6: Interval (in sec) : 3600.00 -6: Number of requests : 9999 -6: --------------------------------------------- -6: Number of grids (including output grid) = 2 -6: -6: -6: Extension for grid 1 is --> uglo_100km -6: -6: Grid Particulars are : -6: Dimensions = 45166 1 -6: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Extension for grid 2 is --> gsh_15m -6: -6: Grid Particulars are : -6: Dimensions = 1440 277 -6: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -6: -6: -6: Preparing interpolation weights for output grid -6: Total number of wet points for interpolation 317192 -6: -6: -6: Variable: Grid Interpolation Map Units: 0.100E+01 -6: -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 277 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 241 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 205 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 169 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 133 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 97 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 61 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 | -6: 25 | 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 | -6: | | -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: -6: -6: Interpolating fields .... -6: -6: Output group 1 -6: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -6: Output group 2 -6: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -6: Output group 3 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 4 -6: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -6: Output group 5 -6: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -6: Output group 6 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 7 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 8 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 9 -6: Output variables skipped -6: Output group 10 -6: Output variables skipped -6: ------------------------------------------------ -6: 1Current vel. -6: 1Wind speed -6: 1Ice concentration -6: 2Wave height -6: 2Mean wave period(+2) -6: 2Mean wave period(+1) -6: 2Peak frequency -6: 2Mean wave dir. a1b1 -6: 2Peak direction -6: 4Part. wave height -6: 4Part. peak period -6: 4Part. mean direction -6: 5Charnock parameter -6: ------------------------------------------------ -6: OUTPUT TIME : 2021/03/27 12:00:00 UTC -6: -6: End of file reached -6: -6: -6: *** End of Grid interpolation Routine *** -6: =============================================== -6: -6: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -6: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -6: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_gsh_15m/out_grd.gsh_15m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.gsh_15m -6: + cmdfile.7[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032712 96 gsouth 0p25 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + cmdfile.7[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib2_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grib2_sbs.sh[30]grdID=gsh_15m -6: + wave_grib2_sbs.sh[31]GRIDNR=255 -6: + wave_grib2_sbs.sh[32]MODNR=11 -6: + wave_grib2_sbs.sh[33]valid_time=2021032712 -6: + wave_grib2_sbs.sh[34]fhr=96 -6: + wave_grib2_sbs.sh[35]grid_region=gsouth -6: + wave_grib2_sbs.sh[36]grid_res=0p25 -6: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676 -6: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_gsh_15m -6: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_gsh_15m -6: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_gsh_15m -6: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_gsh_15m -6: ++ wave_grib2_sbs.sh[47]printf %03i 96 -6: + wave_grib2_sbs.sh[47]FH3=096 -6: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -6: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.gsouth.0p25.f096.grib2 -6: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f096.grib2 ]] -6: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mod_def.gsh_15m ./mod_def.ww3 -6: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.gsh_15m ./out_grd.ww3 -6: + wave_grib2_sbs.sh[73]ngrib=1 -6: + wave_grib2_sbs.sh[74]dtgrib=3600 -6: + wave_grib2_sbs.sh[75]tstart='20210327 120000' -6: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 120000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gsh_15m.inp.tmpl -6: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -6: $ WAVEWATCH-III gridded output input file -6: $ ---------------------------------------- -6: 20210327 120000 3600 1 -6: N -6: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -6: $ -6: 20210327 120000 7 11 255 0 0 -6: $ -6: $ end of input file -6: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[88]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.318362 ']' -6: ++ prep_step[4]echo gfs_ww3_grib.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[90]export err=0 -6: + wave_grib2_sbs.sh[90]err=0 -6: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[95]cat grib2_gsouth_096.out -6: -6: *** WAVEWATCH III GRIB output postp. *** -6: ============================================== -6: -6: Comment character is '$' -6: -6: Grid name : GFSv16-wave S Hemisphere 1/4 d -6: -6: LINEIN: -6: 20210327 120000 3600 1 -6: -6: 20210327120000 3600 1 -6: GEN_PRO -99999 -6: -6: Output time data : -6: ----------------------------------------------------- -6: First time : 2021/03/27 12:00:00 UTC -6: Interval : 01:00:00 -6: Number of requests : 1 -6: Fields : Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: Requested output fields not yet available: -6: ----------------------------------------------------- -6: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -6: -6: Successfully requested output fields : -6: ----------------------------------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: -6: Additional GRIB parameters : -6: ----------------------------------------------------- -6: Run time : 2021/03/27 12:00:00 UTC -6: GRIB center ID : 7 -6: GRIB gen. proc. ID : 11 -6: GRIB grid ID : 255 -6: GRIB GDS parameter : 0 -6: Fields in file : -6: -------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: CHOSEN GRID TYPE: : LLRECTILINEAR -6: -6: -6: -6: Generating file -6: ----------------------------------------------------- -6: Data for 2021/03/27 12:00:00 UTC 0H forecast. -6: -6: End of program -6: ========================================= -6: WAVEWATCH III GRIB output -6: -6: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -6: + wave_grib2_sbs.sh[102][[ 96 -gt 0 ]] -6: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '96 hour fcst' -grib gfs.wave.t12z.gsouth.0p25.f096.grib2 -6: 1:0:d=2021032312:SPC:surface:96 hour fcst: -6: 2:118473:d=2021032312:DIRC:surface:96 hour fcst: -6: 3:506614:d=2021032312:UOGRD:surface:96 hour fcst: -6: 4:620228:d=2021032312:VOGRD:surface:96 hour fcst: -6: 5:742443:d=2021032312:WIND:surface:96 hour fcst: -6: 6:985959:d=2021032312:WDIR:surface:96 hour fcst: -6: 7:1367599:d=2021032312:UGRD:surface:96 hour fcst: -6: 8:1604033:d=2021032312:VGRD:surface:96 hour fcst: -6: 9:1843374:d=2021032312:ICEC:surface:96 hour fcst: -6: 10:1904510:d=2021032312:HTSGW:surface:96 hour fcst: -6: 11:2060008:d=2021032312:IMWF:surface:96 hour fcst: -6: 12:2230933:d=2021032312:MWSPER:surface:96 hour fcst: -6: 13:2402151:d=2021032312:PERPW:surface:96 hour fcst: -6: 14:2585504:d=2021032312:WWSDIR:surface:96 hour fcst: -6: 15:2917138:d=2021032312:DIRPW:surface:96 hour fcst: -6: 16:3256299:d=2021032312:WVHGT:surface:96 hour fcst: -6: 17:3412173:d=2021032312:SWELL:1 in sequence:96 hour fcst: -6: 18:3586000:d=2021032312:SWELL:2 in sequence:96 hour fcst: -6: 19:3716132:d=2021032312:SWELL:3 in sequence:96 hour fcst: -6: 20:3796595:d=2021032312:WVPER:surface:96 hour fcst: -6: 21:3971096:d=2021032312:SWPER:1 in sequence:96 hour fcst: -6: 22:4167149:d=2021032312:SWPER:2 in sequence:96 hour fcst: -6: 23:4333281:d=2021032312:SWPER:3 in sequence:96 hour fcst: -6: 24:4433168:d=2021032312:WVDIR:surface:96 hour fcst: -6: 25:4704597:d=2021032312:SWDIR:1 in sequence:96 hour fcst: -6: 26:5054632:d=2021032312:SWDIR:2 in sequence:96 hour fcst: -6: 27:5336162:d=2021032312:SWDIR:3 in sequence:96 hour fcst: -6: + wave_grib2_sbs.sh[104]err=0 -6: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.gsouth.0p25.f096.grib2 -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f096.grib2 ]] -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f096.grib2.idx ]] -6: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.gsouth.0p25.f096.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f096.grib2 -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f096.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f096.grib2 = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f096.grib2 ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f096.grib2 -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f096.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f096.grib2.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f096.grib2.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f096.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f096.grib2 -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.gsouth.0p25.f096.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f096.grib2.idx -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f096.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f096.grib2.idx = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f096.grib2.idx ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f096.grib2.idx -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f096.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f096.grib2.idx.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f096.grib2.idx.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f096.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f096.grib2.idx -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.gsouth.0p25.f096.grib2 and gfs.wave.t12z.gsouth.0p25.f096.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_gsh_15m to COM' -6: INFO: Copied gfs.wave.t12z.gsouth.0p25.f096.grib2 and gfs.wave.t12z.gsouth.0p25.f096.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_gsh_15m to COM -6: + wave_grib2_sbs.sh[130][[ gsh_15m == '' ]] -6: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -6: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.gsouth.0p25.f096.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -6: INFO: gfs.wave.t12z.gsouth.0p25.f096.grib2 is global.0p50 or SENDDBN is NO, no alert sent -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + cmdfile.1[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032712 3600. 9999 -0: + cmdfile.1[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grid_interp_sbs.sh[25]grdID=at_10m -0: + wave_grid_interp_sbs.sh[26]valid_time=2021032712 -0: + wave_grid_interp_sbs.sh[27]dt=3600. -0: + wave_grid_interp_sbs.sh[28]nst=9999 -0: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676 -0: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.uglo_100km ./out_grd.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mod_def.uglo_100km ./mod_def.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mod_def.at_10m ./mod_def.at_10m -0: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ]] -0: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m'\''' -0: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m' -0: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[51]weights_found=1 -0: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 120000' -0: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 120000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/at_10m/g ww3_gint.inp.tmpl -0: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -0: $ Input file for interpolation of uglo_100km to at_10m -0: $------------------------------------------------ -0: $ Start Time 3600. NSteps -0: 20210327 120000 3600. 9999 -0: $ Total number of grids -0: 2 -0: $ Grid extensions -0: 'uglo_100km' -0: 'at_10m' -0: $ -0: 0 -0: $ -0: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[70]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.318362 ']' -0: ++ prep_step[4]echo gfs_ww3_gint.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -0: INFO: Executing 'gfs_ww3_gint.x' -0: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[73]cat grid_interp.at_10m.out -0: -0: *** WAVEWATCH III Grid interpolation *** -0: =============================================== -0: -0: Comment character is '$' -0: -0: Time Information : -0: --------------------------------------------- -0: Starting Time : 2021/03/27 12:00:00 UTC -0: Interval (in sec) : 3600.00 -0: Number of requests : 9999 -0: --------------------------------------------- -0: Number of grids (including output grid) = 2 -0: -0: -0: Extension for grid 1 is --> uglo_100km -0: -0: Grid Particulars are : -0: Dimensions = 45166 1 -0: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Extension for grid 2 is --> at_10m -0: -0: Grid Particulars are : -0: Dimensions = 301 331 -0: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -0: -0: -0: Preparing interpolation weights for output grid -0: Total number of wet points for interpolation 29591 -0: -0: -0: Variable: Grid Interpolation Map Units: 0.100E+01 -0: -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 331 | | -0: | | -0: | | -0: 289 | | -0: | | -0: | 0 0 0 | -0: 247 | 0 0 0 0 | -0: | 0 0 0 0 0 | -0: | 0 0 0 0 | -0: 205 | 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 163 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 121 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 79 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 | -0: | 0 0 | -0: 37 | 0 | -0: | | -0: | | -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: -0: -0: Interpolating fields .... -0: -0: Output group 1 -0: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -0: Output group 2 -0: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -0: Output group 3 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 4 -0: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -0: Output group 5 -0: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -0: Output group 6 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 7 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 8 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 9 -0: Output variables skipped -0: Output group 10 -0: Output variables skipped -0: ------------------------------------------------ -0: 1Current vel. -0: 1Wind speed -0: 1Ice concentration -0: 2Wave height -0: 2Mean wave period(+2) -0: 2Mean wave period(+1) -0: 2Peak frequency -0: 2Mean wave dir. a1b1 -0: 2Peak direction -0: 4Part. wave height -0: 4Part. peak period -0: 4Part. mean direction -0: 5Charnock parameter -0: ------------------------------------------------ -0: OUTPUT TIME : 2021/03/27 12:00:00 UTC -0: -0: End of file reached -0: -0: -0: *** End of Grid interpolation Routine *** -0: =============================================== -0: -0: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -0: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -0: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grid_interp_at_10m/out_grd.at_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.at_10m -0: + cmdfile.1[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032712 96 atlocn 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + cmdfile.1[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib2_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grib2_sbs.sh[30]grdID=at_10m -0: + wave_grib2_sbs.sh[31]GRIDNR=255 -0: + wave_grib2_sbs.sh[32]MODNR=11 -0: + wave_grib2_sbs.sh[33]valid_time=2021032712 -0: + wave_grib2_sbs.sh[34]fhr=96 -0: + wave_grib2_sbs.sh[35]grid_region=atlocn -0: + wave_grib2_sbs.sh[36]grid_res=0p16 -0: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676 -0: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_at_10m -0: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_at_10m -0: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_at_10m -0: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_at_10m -0: ++ wave_grib2_sbs.sh[47]printf %03i 96 -0: + wave_grib2_sbs.sh[47]FH3=096 -0: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_atlocn_0p16 -0: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.atlocn.0p16.f096.grib2 -0: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f096.grib2 ]] -0: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/mod_def.at_10m ./mod_def.ww3 -0: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/out_grd.at_10m ./out_grd.ww3 -0: + wave_grib2_sbs.sh[73]ngrib=1 -0: + wave_grib2_sbs.sh[74]dtgrib=3600 -0: + wave_grib2_sbs.sh[75]tstart='20210327 120000' -0: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 120000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.at_10m.inp.tmpl -0: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -0: $ WAVEWATCH-III gridded output input file -0: $ ---------------------------------------- -0: 20210327 120000 3600 1 -0: N -0: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -0: $ -0: 20210327 120000 7 11 255 0 0 -0: $ -0: $ end of input file -0: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[88]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.318362 ']' -0: ++ prep_step[4]echo gfs_ww3_grib.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[90]export err=0 -0: + wave_grib2_sbs.sh[90]err=0 -0: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[95]cat grib2_atlocn_096.out -0: -0: *** WAVEWATCH III GRIB output postp. *** -0: ============================================== -0: -0: Comment character is '$' -0: -0: Grid name : NW Atlantic 10 min wave grid -0: -0: LINEIN: -0: 20210327 120000 3600 1 -0: -0: 20210327120000 3600 1 -0: GEN_PRO -99999 -0: -0: Output time data : -0: ----------------------------------------------------- -0: First time : 2021/03/27 12:00:00 UTC -0: Interval : 01:00:00 -0: Number of requests : 1 -0: Fields : Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: Requested output fields not yet available: -0: ----------------------------------------------------- -0: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -0: -0: Successfully requested output fields : -0: ----------------------------------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: -0: Additional GRIB parameters : -0: ----------------------------------------------------- -0: Run time : 2021/03/27 12:00:00 UTC -0: GRIB center ID : 7 -0: GRIB gen. proc. ID : 11 -0: GRIB grid ID : 255 -0: GRIB GDS parameter : 0 -0: Fields in file : -0: -------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: CHOSEN GRID TYPE: : LLRECTILINEAR -0: -0: -0: -0: Generating file -0: ----------------------------------------------------- -0: Data for 2021/03/27 12:00:00 UTC 0H forecast. -0: -0: End of program -0: ========================================= -0: WAVEWATCH III GRIB output -0: -0: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -0: + wave_grib2_sbs.sh[102][[ 96 -gt 0 ]] -0: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '96 hour fcst' -grib gfs.wave.t12z.atlocn.0p16.f096.grib2 -0: 1:0:d=2021032312:SPC:surface:96 hour fcst: -0: 2:20186:d=2021032312:DIRC:surface:96 hour fcst: -0: 3:60932:d=2021032312:UOGRD:surface:96 hour fcst: -0: 4:80525:d=2021032312:VOGRD:surface:96 hour fcst: -0: 5:100489:d=2021032312:WIND:surface:96 hour fcst: -0: 6:131350:d=2021032312:WDIR:surface:96 hour fcst: -0: 7:171574:d=2021032312:UGRD:surface:96 hour fcst: -0: 8:201843:d=2021032312:VGRD:surface:96 hour fcst: -0: 9:231931:d=2021032312:ICEC:surface:96 hour fcst: -0: 10:244566:d=2021032312:HTSGW:surface:96 hour fcst: -0: 11:270150:d=2021032312:IMWF:surface:96 hour fcst: -0: 12:296663:d=2021032312:MWSPER:surface:96 hour fcst: -0: 13:323429:d=2021032312:PERPW:surface:96 hour fcst: -0: 14:351186:d=2021032312:WWSDIR:surface:96 hour fcst: -0: 15:391111:d=2021032312:DIRPW:surface:96 hour fcst: -0: 16:431891:d=2021032312:WVHGT:surface:96 hour fcst: -0: 17:456672:d=2021032312:SWELL:1 in sequence:96 hour fcst: -0: 18:477906:d=2021032312:SWELL:2 in sequence:96 hour fcst: -0: 19:493836:d=2021032312:SWELL:3 in sequence:96 hour fcst: -0: 20:507192:d=2021032312:WVPER:surface:96 hour fcst: -0: 21:533952:d=2021032312:SWPER:1 in sequence:96 hour fcst: -0: 22:556886:d=2021032312:SWPER:2 in sequence:96 hour fcst: -0: 23:574562:d=2021032312:SWPER:3 in sequence:96 hour fcst: -0: 24:588555:d=2021032312:WVDIR:surface:96 hour fcst: -0: 25:626130:d=2021032312:SWDIR:1 in sequence:96 hour fcst: -0: 26:658771:d=2021032312:SWDIR:2 in sequence:96 hour fcst: -0: 27:680716:d=2021032312:SWDIR:3 in sequence:96 hour fcst: -0: + wave_grib2_sbs.sh[104]err=0 -0: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.atlocn.0p16.f096.grib2 -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f096.grib2 ]] -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f096.grib2.idx ]] -0: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.atlocn.0p16.f096.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f096.grib2 -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f096.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f096.grib2 = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f096.grib2 ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f096.grib2 -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f096.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f096.grib2.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f096.grib2.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f096.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f096.grib2 -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.atlocn.0p16.f096.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f096.grib2.idx -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f096.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f096.grib2.idx = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f096.grib2.idx ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f096.grib2.idx -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f096.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f096.grib2.idx.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f096.grib2.idx.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f096.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f096.grib2.idx -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.atlocn.0p16.f096.grib2 and gfs.wave.t12z.atlocn.0p16.f096.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_at_10m to COM' -0: INFO: Copied gfs.wave.t12z.atlocn.0p16.f096.grib2 and gfs.wave.t12z.atlocn.0p16.f096.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676/grib_at_10m to COM -0: + wave_grib2_sbs.sh[130][[ at_10m == '' ]] -0: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -0: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.atlocn.0p16.f096.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -0: INFO: gfs.wave.t12z.atlocn.0p16.f096.grib2 is global.0p50 or SENDDBN is NO, no alert sent -+ run_mpmd.sh[113]exit 0 -+ run_mpmd.sh[1]postamble run_mpmd.sh 1753758301 0 -+ preamble.sh[62]set +x -End run_mpmd.sh at 03:05:12 with error code 0 (time elapsed: 00:00:11) -+ exgfs_wave_post_gridded_sbs.sh[122]true -+ exgfs_wave_post_gridded_sbs.sh[123]export err=0 -+ exgfs_wave_post_gridded_sbs.sh[123]err=0 -+ exgfs_wave_post_gridded_sbs.sh[124][[ 0 -ne 0 ]] -+ exgfs_wave_post_gridded_sbs.sh[130]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ exgfs_wave_post_gridded_sbs.sh[131]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ exgfs_wave_post_gridded_sbs.sh[132]gribchk=gfs.wave.t12z.gsouth.0p25.f096.grib2 -+ exgfs_wave_post_gridded_sbs.sh[133][[ ! -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f096.grib2 ]] -+ exgfs_wave_post_gridded_sbs.sh[138]exit 0 -+ JGLOBAL_WAVE_POST_SBS[28]true -+ JGLOBAL_WAVE_POST_SBS[29]export err=0 -+ JGLOBAL_WAVE_POST_SBS[29]err=0 -+ JGLOBAL_WAVE_POST_SBS[30][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[37]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312 -+ JGLOBAL_WAVE_POST_SBS[38][[ NO != \Y\E\S ]] -+ JGLOBAL_WAVE_POST_SBS[39]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f096.314676 -+ JGLOBAL_WAVE_POST_SBS[42]exit 0 -+ JGLOBAL_WAVE_POST_SBS[1]postamble /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS 1753758289 0 -+ preamble.sh[62]set +x -End /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at 03:05:12 with error code 0 (time elapsed: 00:00:23) -Begin /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at Tue Jul 29 03:05:12 UTC 2025 -++ jjob_header.sh[46]OPTIND=1 -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[50]env_job=wavepostsbs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[49]read -ra configs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[61]shift 4 -++ jjob_header.sh[63][[ -z wavepostsbs ]] -++ jjob_header.sh[71]export DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676 -++ jjob_header.sh[71]DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676 -++ jjob_header.sh[72][[ YES == \Y\E\S ]] -++ jjob_header.sh[73]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676 -++ jjob_header.sh[75]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676 -++ jjob_header.sh[76]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676 -++ jjob_header.sh[85]export pid=321351 -++ jjob_header.sh[85]pid=321351 -++ jjob_header.sh[86]export pgmout=OUTPUT.321351 -++ jjob_header.sh[86]pgmout=OUTPUT.321351 -++ jjob_header.sh[87]export pgmerr=errfile -++ jjob_header.sh[87]pgmerr=errfile -++ jjob_header.sh[90]export pgm= -++ jjob_header.sh[90]pgm= -++ jjob_header.sh[96]export cycle=t12z -++ jjob_header.sh[96]cycle=t12z -++ jjob_header.sh[97]setpdy.sh -+ setpdy.sh[20]'[' /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676 == /home/mterry ']' -+ setpdy.sh[25][[ ! t12z =~ t??z ]] -+ setpdy.sh[30]case $# in -+ setpdy.sh[31]dates_before_PDY=7 -+ setpdy.sh[32]dates_after_PDY=7 -+ setpdy.sh[50]COMDATEROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+ setpdy.sh[53]'[' -z 20210323 ']' -+ setpdy.sh[57]sed 's/[0-9]\{8\}/20210323/' /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z -sed: can't read /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z: No such file or directory -++ jjob_header.sh[97]true -++ jjob_header.sh[98]source ./PDY -/work2/noaa/global/mterry/global-workflow_forked/ush/jjob_header.sh: line 98: ./PDY: No such file or directory -++ jjob_header.sh[98]true -++ jjob_header.sh[104]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[104]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.base -+++ config.base[6]echo 'BEGIN: config.base' -BEGIN: config.base -+++ config.base[9]export machine=HERCULES -+++ config.base[9]machine=HERCULES -+++ config.base[12]export RUN_ENVIR=emc -+++ config.base[12]RUN_ENVIR=emc -+++ config.base[15]export ACCOUNT=fv3-cpu -+++ config.base[15]ACCOUNT=fv3-cpu -+++ config.base[16]export QUEUE=batch -+++ config.base[16]QUEUE=batch -+++ config.base[17]export QUEUE_SERVICE=batch -+++ config.base[17]QUEUE_SERVICE=batch -+++ config.base[18]export QUEUE_DTN=batch -+++ config.base[18]QUEUE_DTN=batch -+++ config.base[19]export PARTITION_BATCH=hercules -+++ config.base[19]PARTITION_BATCH=hercules -+++ config.base[20]export PARTITION_SERVICE=service -+++ config.base[20]PARTITION_SERVICE=service -+++ config.base[21]export PARTITION_DTN= -+++ config.base[21]PARTITION_DTN= -+++ config.base[22]export RESERVATION= -+++ config.base[22]RESERVATION= -+++ config.base[23]export CLUSTERS= -+++ config.base[23]CLUSTERS= -+++ config.base[24]export CLUSTERS_SERVICE= -+++ config.base[24]CLUSTERS_SERVICE= -+++ config.base[25]export CLUSTERS_DTN= -+++ config.base[25]CLUSTERS_DTN= -+++ config.base[28]export HPSS_PROJECT=emc-global -+++ config.base[28]HPSS_PROJECT=emc-global -+++ config.base[31]export HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[31]HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[32]export EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[32]EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[33]export FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[33]FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[34]export PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[34]PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[35]export SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[35]SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[36]export USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[36]USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[38]export FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[38]FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[39]export FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[39]FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[40]export FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[40]FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[41]export FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[41]FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[42]export FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[42]FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[43]export FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[43]FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[44]export FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[44]FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[45]export FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[45]FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[50]export PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[50]PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[51]export COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[51]COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[52]export COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[52]COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[53]export DMPDIR=/work/noaa/rstprod/dump -+++ config.base[53]DMPDIR=/work/noaa/rstprod/dump -+++ config.base[57]export COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[57]COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[58]export COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[58]COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[59]export COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[59]COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[62]export HOMEDIR=/work2/noaa/global/mterry -+++ config.base[62]HOMEDIR=/work2/noaa/global/mterry -+++ config.base[63]export STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[63]STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]export PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[65]export NOSCRUB=/work2/noaa/global/mterry -+++ config.base[65]NOSCRUB=/work2/noaa/global/mterry -+++ config.base[68]export BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[68]BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[71]export BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[71]BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[74]export DO_PREP_SFC=NO -+++ config.base[74]DO_PREP_SFC=NO -+++ config.base[77]export DO_GOES=NO -+++ config.base[77]DO_GOES=NO -+++ config.base[78]export DO_BUFRSND=NO -+++ config.base[78]DO_BUFRSND=NO -+++ config.base[79]export DO_GEMPAK=NO -+++ config.base[79]DO_GEMPAK=NO -+++ config.base[80]export DO_AWIPS=NO -+++ config.base[80]DO_AWIPS=NO -+++ config.base[81]export DO_NPOESS=NO -+++ config.base[81]DO_NPOESS=NO -+++ config.base[82]export DO_TRACKER=YES -+++ config.base[82]DO_TRACKER=YES -+++ config.base[83]export DO_GENESIS=YES -+++ config.base[83]DO_GENESIS=YES -+++ config.base[84]export DO_GENESIS_FSU=NO -+++ config.base[84]DO_GENESIS_FSU=NO -+++ config.base[85]export DO_VERFOZN=YES -+++ config.base[85]DO_VERFOZN=YES -+++ config.base[86]export DO_VERFRAD=YES -+++ config.base[86]DO_VERFRAD=YES -+++ config.base[87]export DO_VMINMON=YES -+++ config.base[87]DO_VMINMON=YES -+++ config.base[88]export DO_ANLSTAT=NO -+++ config.base[88]DO_ANLSTAT=NO -+++ config.base[91]export MODE=forecast-only -+++ config.base[91]MODE=forecast-only -+++ config.base[92]export DO_TEST_MODE=YES -+++ config.base[92]DO_TEST_MODE=YES -+++ config.base[101]export FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[101]FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[102]export HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[102]HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[103]export HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[103]HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[106]export NMV=/bin/mv -+++ config.base[106]NMV=/bin/mv -+++ config.base[107]export 'NLN=/bin/ln -sf' -+++ config.base[107]NLN='/bin/ln -sf' -+++ config.base[108]export VERBOSE=YES -+++ config.base[108]VERBOSE=YES -+++ config.base[109]export KEEPDATA=NO -+++ config.base[109]KEEPDATA=NO -+++ config.base[110]export DEBUG_POSTSCRIPT=NO -+++ config.base[110]DEBUG_POSTSCRIPT=NO -+++ config.base[111]export CHGRP_RSTPROD=YES -+++ config.base[111]CHGRP_RSTPROD=YES -+++ config.base[112]export 'CHGRP_CMD=chgrp rstprod' -+++ config.base[112]CHGRP_CMD='chgrp rstprod' -+++ config.base[113]export NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[113]NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[114]export NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[114]NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[117]export BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[117]BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[120]export SDATE=2021032312 -+++ config.base[120]SDATE=2021032312 -+++ config.base[121]export EDATE=2021032312 -+++ config.base[121]EDATE=2021032312 -+++ config.base[122]export EXP_WARM_START=.false. -+++ config.base[122]EXP_WARM_START=.false. -+++ config.base[123]export assim_freq=6 -+++ config.base[123]assim_freq=6 -+++ config.base[124]export PSLOT=C48_S2SW -+++ config.base[124]PSLOT=C48_S2SW -+++ config.base[125]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[125]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[126]export ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[126]ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[127]export DUMP_SUFFIX= -+++ config.base[127]DUMP_SUFFIX= -+++ config.base[128][[ 2021032312 -ge 2019092100 ]] -+++ config.base[128][[ 2021032312 -le 2019110700 ]] -+++ config.base[131]export ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[131]ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[132]export ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[132]ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[133]export FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[133]FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[136]export envir=prod -+++ config.base[136]envir=prod -+++ config.base[137]export NET=gfs -+++ config.base[137]NET=gfs -+++ config.base[138]export RUN=gfs -+++ config.base[138]RUN=gfs -+++ config.base[141]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.com -++++ config.com[4]echo 'BEGIN: config.com' -BEGIN: config.com -++++ config.com[38][[ emc == \n\c\o ]] -++++ config.com[43]COM_OBSPROC_TMPL='${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos' -++++ config.com[44]COM_RTOFS_TMPL='${DMPDIR}' -++++ config.com[45]COM_TCVITAL_TMPL='${DMPDIR}/${RUN}.${YMD}/${HH}/atmos' -++++ config.com[47]declare -rx 'COM_OBS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/obs' -++++ config.com[48]declare -rx COM_OBSPROC_TMPL COM_RTOFS_TMPL -++++ config.com[50]COM_BASE='${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}' -++++ config.com[52]declare -rx 'COM_TOP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}' -++++ config.com[54]declare -rx 'COM_CONF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf' -++++ config.com[55]declare -rx 'COM_OBS_JEDI=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi' -++++ config.com[57]declare -rx 'COM_ATMOS_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input' -++++ config.com[58]declare -rx 'COM_ATMOS_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart' -++++ config.com[59]declare -rx 'COM_ATMOS_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos' -++++ config.com[60]declare -rx 'COM_SNOW_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow' -++++ config.com[61]declare -rx 'COM_SNOW_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon' -++++ config.com[62]declare -rx 'COM_ATMOS_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history' -++++ config.com[63]declare -rx 'COM_ATMOS_MASTER_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master' -++++ config.com[64]declare -rx 'COM_ATMOS_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2' -++++ config.com[65]declare -rx 'COM_ATMOS_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}' -++++ config.com[66]declare -rx 'COM_ATMOS_BUFR_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr' -++++ config.com[67]declare -rx 'COM_ATMOS_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}' -++++ config.com[68]declare -rx 'COM_ATMOS_GENESIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital' -++++ config.com[69]declare -rx 'COM_ATMOS_TRACK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks' -++++ config.com[70]declare -rx 'COM_ATMOS_GOES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim' -++++ config.com[71]declare -rx 'COM_ATMOS_IMAGERY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery' -++++ config.com[72]declare -rx 'COM_ATMOS_OZNMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon' -++++ config.com[73]declare -rx 'COM_ATMOS_RADMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon' -++++ config.com[74]declare -rx 'COM_ATMOS_MINMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon' -++++ config.com[75]declare -rx 'COM_ATMOS_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon' -++++ config.com[76]declare -rx 'COM_ATMOS_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo' -++++ config.com[78]declare -rx 'COM_WAVE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart' -++++ config.com[79]declare -rx 'COM_WAVE_PREP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep' -++++ config.com[80]declare -rx 'COM_WAVE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history' -++++ config.com[81]declare -rx 'COM_WAVE_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded' -++++ config.com[82]declare -rx 'COM_WAVE_GRID_RES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}' -++++ config.com[83]declare -rx 'COM_WAVE_STATION_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station' -++++ config.com[84]declare -rx 'COM_WAVE_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak' -++++ config.com[85]declare -rx 'COM_WAVE_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo' -++++ config.com[87]declare -rx 'COM_OCEAN_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history' -++++ config.com[88]declare -rx 'COM_OCEAN_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart' -++++ config.com[89]declare -rx 'COM_OCEAN_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input' -++++ config.com[90]declare -rx 'COM_OCEAN_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean' -++++ config.com[91]declare -rx 'COM_OCEAN_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon' -++++ config.com[92]declare -rx 'COM_OCEAN_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf' -++++ config.com[93]declare -rx 'COM_OCEAN_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean' -++++ config.com[94]declare -rx 'COM_OCEAN_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf' -++++ config.com[95]declare -rx 'COM_OCEAN_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2' -++++ config.com[96]declare -rx 'COM_OCEAN_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}' -++++ config.com[98]declare -rx 'COM_ICE_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice' -++++ config.com[99]declare -rx 'COM_ICE_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf' -++++ config.com[100]declare -rx 'COM_ICE_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon' -++++ config.com[101]declare -rx 'COM_ICE_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice' -++++ config.com[102]declare -rx 'COM_ICE_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input' -++++ config.com[103]declare -rx 'COM_ICE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history' -++++ config.com[104]declare -rx 'COM_ICE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart' -++++ config.com[105]declare -rx 'COM_ICE_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf' -++++ config.com[106]declare -rx 'COM_ICE_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2' -++++ config.com[107]declare -rx 'COM_ICE_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}' -++++ config.com[109]declare -rx 'COM_CHEM_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history' -++++ config.com[110]declare -rx 'COM_CHEM_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem' -++++ config.com[111]declare -rx 'COM_CHEM_BMAT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix' -++++ config.com[112]declare -rx 'COM_CHEM_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon' -++++ config.com[114]declare -rx 'COM_MED_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart' -+++ config.base[143]export LOGSCRIPT= -+++ config.base[143]LOGSCRIPT= -+++ config.base[145]export 'REDOUT=1>' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wave -+++ config.wave[6]echo 'BEGIN: config.wave' -BEGIN: config.wave -+++ config.wave[13]export RUNRSTwave=gdas -+++ config.wave[13]RUNRSTwave=gdas -+++ config.wave[16]export MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[16]MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[19]case "${waveGRD}" in -+++ config.wave[64]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[64]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[65]export wavepostGRD= -+++ config.wave[65]wavepostGRD= -+++ config.wave[66]export waveuoutpGRD=uglo_100km -+++ config.wave[66]waveuoutpGRD=uglo_100km -+++ config.wave[75]export WAVEWND_DID= -+++ config.wave[75]WAVEWND_DID= -+++ config.wave[76]export WAVEWND_FID= -+++ config.wave[76]WAVEWND_FID= -+++ config.wave[79][[ gfs == \g\f\s ]] -+++ config.wave[80]export FHMAX_WAV=120 -+++ config.wave[80]FHMAX_WAV=120 -+++ config.wave[82]export WAVHINDH=0 -+++ config.wave[82]WAVHINDH=0 -+++ config.wave[83]export FHMAX_WAV_IBP=180 -+++ config.wave[83]FHMAX_WAV_IBP=180 -+++ config.wave[84](( FHMAX_WAV < FHMAX_WAV_IBP )) -+++ config.wave[84]export FHMAX_WAV_IBP=120 -+++ config.wave[84]FHMAX_WAV_IBP=120 -+++ config.wave[87]export DTFLD_WAV=3600 -+++ config.wave[87]DTFLD_WAV=3600 -+++ config.wave[88]export DTPNT_WAV=3600 -+++ config.wave[88]DTPNT_WAV=3600 -+++ config.wave[89]export FHINCP_WAV=1 -+++ config.wave[89]FHINCP_WAV=1 -+++ config.wave[92]export 'OUTPARS_WAV=WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[92]OUTPARS_WAV='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[95][[ gfs == \g\d\a\s ]] -+++ config.wave[99](( INTERVAL_GFS > 0 )) -+++ config.wave[100]export WAVHCYC=6 -+++ config.wave[100]WAVHCYC=6 -+++ config.wave[101]export FHMAX_WAV_CUR=192 -+++ config.wave[101]FHMAX_WAV_CUR=192 -+++ config.wave[109]export RSTTYPE_WAV=T -+++ config.wave[109]RSTTYPE_WAV=T -+++ config.wave[110][[ gfs != gfs ]] -+++ config.wave[115]rst_dt_gfs=43200 -+++ config.wave[116][[ 43200 -gt 0 ]] -+++ config.wave[117]export DT_1_RST_WAV=0 -+++ config.wave[117]DT_1_RST_WAV=0 -+++ config.wave[120]export DT_2_RST_WAV=43200 -+++ config.wave[120]DT_2_RST_WAV=43200 -+++ config.wave[126]export RSTIOFF_WAV=0 -+++ config.wave[126]RSTIOFF_WAV=0 -+++ config.wave[131]export RUNMEM=-1 -+++ config.wave[131]RUNMEM=-1 -+++ config.wave[134](( RUNMEM == -1 )) -+++ config.wave[136]export waveMEMB= -+++ config.wave[136]waveMEMB= -+++ config.wave[143]export WW3ATMINP=CPL -+++ config.wave[143]WW3ATMINP=CPL -+++ config.wave[144][[ YES == \Y\E\S ]] -+++ config.wave[145]export WW3ICEINP=CPL -+++ config.wave[145]WW3ICEINP=CPL -+++ config.wave[146]export WAVEICE_FID= -+++ config.wave[146]WAVEICE_FID= -+++ config.wave[152][[ YES == \Y\E\S ]] -+++ config.wave[153]export WW3CURINP=CPL -+++ config.wave[153]WW3CURINP=CPL -+++ config.wave[154]export WAVECUR_FID= -+++ config.wave[154]WAVECUR_FID= -+++ config.wave[161]export WW3ATMIENS=F -+++ config.wave[161]WW3ATMIENS=F -+++ config.wave[162]export WW3ICEIENS=F -+++ config.wave[162]WW3ICEIENS=F -+++ config.wave[163]export WW3CURIENS=F -+++ config.wave[163]WW3CURIENS=F -+++ config.wave[165]export GOFILETYPE=1 -+++ config.wave[165]GOFILETYPE=1 -+++ config.wave[166]export POFILETYPE=1 -+++ config.wave[166]POFILETYPE=1 -+++ config.wave[170]export FUNIPNT=T -+++ config.wave[170]FUNIPNT=T -+++ config.wave[172]export IOSRV=1 -+++ config.wave[172]IOSRV=1 -+++ config.wave[174]export FPNTPROC=T -+++ config.wave[174]FPNTPROC=T -+++ config.wave[176]export FGRDPROC=F -+++ config.wave[176]FGRDPROC=F -+++ config.wave[178]export FLAGMASKCOMP=F -+++ config.wave[178]FLAGMASKCOMP=F -+++ config.wave[180]export FLAGMASKOUT=F -+++ config.wave[180]FLAGMASKOUT=F -+++ config.wave[182]echo 'END: config.wave' -END: config.wave -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wavepostsbs -+++ config.wavepostsbs[6]echo 'BEGIN: config.wavepostsbs' -BEGIN: config.wavepostsbs -+++ config.wavepostsbs[9]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources wavepostsbs -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=wavepostsbs -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[156]ntasks=8 -++++ config.resources[157]threads_per_task=1 -++++ config.resources[158]tasks_per_node=80 -++++ config.resources[159]NTASKS=8 -++++ config.resources[160]memory=20GB -++++ config.resources[161]walltime=00:15:00 -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export NTASKS -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export memory -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.wavepostsbs[12]export MAX_TASKS=25 -+++ config.wavepostsbs[12]MAX_TASKS=25 -+++ config.wavepostsbs[15]export WAV_SUBGRBSRC= -+++ config.wavepostsbs[15]WAV_SUBGRBSRC= -+++ config.wavepostsbs[16]export WAV_SUBGRB= -+++ config.wavepostsbs[16]WAV_SUBGRB= -+++ config.wavepostsbs[19]export DOFLD_WAV=YES -+++ config.wavepostsbs[19]DOFLD_WAV=YES -+++ config.wavepostsbs[20]export DOPNT_WAV=YES -+++ config.wavepostsbs[20]DOPNT_WAV=YES -+++ config.wavepostsbs[21]export DOGRB_WAV=YES -+++ config.wavepostsbs[21]DOGRB_WAV=YES -+++ config.wavepostsbs[22][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+++ config.wavepostsbs[23]export DOGRI_WAV=YES -+++ config.wavepostsbs[23]DOGRI_WAV=YES -+++ config.wavepostsbs[27]export DOSPC_WAV=YES -+++ config.wavepostsbs[27]DOSPC_WAV=YES -+++ config.wavepostsbs[28]export DOBLL_WAV=YES -+++ config.wavepostsbs[28]DOBLL_WAV=YES -+++ config.wavepostsbs[30]echo 'END: config.wavepostsbs' -END: config.wavepostsbs -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env wavepostsbs -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=wavepostsbs -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 20971520 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 8 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[34]max_threads_per_task=1 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 1 ]] -+++ HERCULES.env[40][[ 1 -gt 1 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 8' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[63]export USE_CFP=YES -+++ HERCULES.env[63]USE_CFP=YES -+++ HERCULES.env[64][[ wavepostsbs == \w\a\v\e\p\r\e\p ]] -+++ HERCULES.env[65]export 'wavempexec=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[65]wavempexec='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[66]export 'wave_mpmd=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[66]wave_mpmd='--multi-prog --output=mpmd.%j.%t.out' -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[4]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ JGLOBAL_WAVE_POST_SBS[7]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[7]HH=12 -+ JGLOBAL_WAVE_POST_SBS[7]declare_from_tmpl -rx COMIN_WAVE_PREP:COM_WAVE_PREP_TMPL COMIN_WAVE_HISTORY:COM_WAVE_HISTORY_TMPL COMOUT_WAVE_GRID:COM_WAVE_GRID_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_WAVE_PREP=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep -declare_from_tmpl :: COMIN_WAVE_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history -declare_from_tmpl :: COMOUT_WAVE_GRID=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded -+ JGLOBAL_WAVE_POST_SBS[12][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n '' ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_atlocn_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_atlocn_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_atlocn_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_epacif_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_epacif_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_epacif_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_wcoast_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_wcoast_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_wcoast_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p50 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p50:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p50=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p50 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=arctic.9km -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_arctic_9km:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_arctic_9km=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_arctic_9km -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_gsouth_0p25:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_gsouth_0p25=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 ]] -+ JGLOBAL_WAVE_POST_SBS[28]/work2/noaa/global/mterry/global-workflow_forked/scripts/exgfs_wave_post_gridded_sbs.sh -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ exgfs_wave_post_gridded_sbs.sh[24]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ exgfs_wave_post_gridded_sbs.sh[26]DOGRI_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[27]DOGRB_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[29]export waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[29]waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[30]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[30]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[31]export wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[31]wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[33]cat - INFO: Grid information: - INFO: Native wave grids: uglo_100km - INFO: Interpolated grids: at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m - INFO: Post-process grids: -++ exgfs_wave_post_gridded_sbs.sh[40]printf %03i 99 -+ exgfs_wave_post_gridded_sbs.sh[40]fhr3=099 -++ exgfs_wave_post_gridded_sbs.sh[41]date -u -d '20210323 12 + 99 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[41]valid_time=2021032715 -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[49]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f099.bin ./out_grd.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f099.bin ./out_grd.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[52][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[62]cat - INFO: Summary: - INFO: Grid interp: DOGRI_WAV="YES" - INFO: Grib files: DOGRB_WAV="YES" - INFO: Fields to be included in grib files: - INFO: OUTPARS_WAV="WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA" -+ exgfs_wave_post_gridded_sbs.sh[70][[ YES == \N\O ]] -+ exgfs_wave_post_gridded_sbs.sh[76]rm -f 'cmdfile.*' cmdfile -+ exgfs_wave_post_gridded_sbs.sh[77]count=0 -+ exgfs_wave_post_gridded_sbs.sh[80][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[81]dt_int=3600. -+ exgfs_wave_post_gridded_sbs.sh[82]n_int=9999 -++ exgfs_wave_post_gridded_sbs.sh[83]date -u -d '20210327 15 - 0 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[83]ymdh_int=2021032715 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=1 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032715 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032715 99 atlocn 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib2_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib2_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=2 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032715 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032715 99 epacif 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib2_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib2_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=3 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032715 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032715 99 wcoast 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib2_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib2_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=4 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032715 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032715 99 global 0p50 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib2_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib2_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=5 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032715 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032715 99 arctic 9km '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib2_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib2_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=6 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032715 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032715 99 global 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib2_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib2_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=7 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032715 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032715 99 gsouth 0p25 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib2_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib2_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[100][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[113][[ 8 -lt 7 ]] -+ exgfs_wave_post_gridded_sbs.sh[121]echo 'INFO: Running MPMD job with 7 commands' -INFO: Running MPMD job with 7 commands -+ exgfs_wave_post_gridded_sbs.sh[122]/work2/noaa/global/mterry/global-workflow_forked/ush/run_mpmd.sh /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/cmdfile -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ run_mpmd.sh[31]source /work2/noaa/global/mterry/global-workflow_forked/ush/preamble.sh -++ preamble.sh[20]set +x -Begin run_mpmd.sh at Tue Jul 29 03:05:23 UTC 2025 -+ run_mpmd.sh[33]cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/cmdfile -+ run_mpmd.sh[36][[ YES != \Y\E\S ]] -+ run_mpmd.sh[46]export OMP_NUM_THREADS=1 -+ run_mpmd.sh[46]OMP_NUM_THREADS=1 -++ run_mpmd.sh[49]wc -l -+ run_mpmd.sh[49]nprocs=7 -+ run_mpmd.sh[52]mpmd_cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mpmd_cmdfile -+ run_mpmd.sh[53][[ -s /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mpmd_cmdfile ]] -+ run_mpmd.sh[55]cat - INFO: Executing MPMD job, STDOUT redirected for each process separately - INFO: On failure, logs for each job will be available in /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mpmd.proc_num.out - INFO: The proc_num corresponds to the line in '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mpmd_cmdfile' -+ run_mpmd.sh[61][[ srun -l --export=ALL --hint=nomultithread =~ ^srun.* ]] -+ run_mpmd.sh[65]nm=0 -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '0 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/cmdfile.1' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '1 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/cmdfile.2' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '2 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/cmdfile.3' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '3 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/cmdfile.4' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '4 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/cmdfile.5' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '5 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/cmdfile.6' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '6 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/cmdfile.7' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[72]set +e -+ run_mpmd.sh[74]srun -l --export=ALL --hint=nomultithread --multi-prog --output=mpmd.%j.%t.out -n 7 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mpmd_cmdfile -+ run_mpmd.sh[75]err=0 -+ run_mpmd.sh[76]set_strict -+ preamble.sh[35][[ YES == \Y\E\S ]] -+ preamble.sh[37]set -eu -+ run_mpmd.sh[103][[ 0 -eq 0 ]] -+ run_mpmd.sh[104]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mpmd_cmdfile -++ run_mpmd.sh[105]find . -name 'mpmd.*.out' -+ run_mpmd.sh[105]out_files='./mpmd.5951798.4.out -./mpmd.5951798.2.out -./mpmd.5951798.1.out -./mpmd.5951798.3.out -./mpmd.5951798.5.out -./mpmd.5951798.6.out -./mpmd.5951798.0.out' -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951798.4.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951798.4.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951798.2.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951798.2.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951798.1.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951798.1.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951798.3.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951798.3.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951798.5.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951798.5.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951798.6.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951798.6.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951798.0.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951798.0.out -+ run_mpmd.sh[110]cat mpmd.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + cmdfile.5[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032715 3600. 9999 -4: + cmdfile.5[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grid_interp_sbs.sh[25]grdID=aoc_9km -4: + wave_grid_interp_sbs.sh[26]valid_time=2021032715 -4: + wave_grid_interp_sbs.sh[27]dt=3600. -4: + wave_grid_interp_sbs.sh[28]nst=9999 -4: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676 -4: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.uglo_100km ./out_grd.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mod_def.uglo_100km ./mod_def.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mod_def.aoc_9km ./mod_def.aoc_9km -4: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km'\''' -4: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km' -4: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[51]weights_found=1 -4: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 150000' -4: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 150000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/aoc_9km/g ww3_gint.inp.tmpl -4: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -4: $ Input file for interpolation of uglo_100km to aoc_9km -4: $------------------------------------------------ -4: $ Start Time 3600. NSteps -4: 20210327 150000 3600. 9999 -4: $ Total number of grids -4: 2 -4: $ Grid extensions -4: 'uglo_100km' -4: 'aoc_9km' -4: $ -4: 0 -4: $ -4: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[70]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.321351 ']' -4: ++ prep_step[4]echo gfs_ww3_gint.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -4: INFO: Executing 'gfs_ww3_gint.x' -4: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[73]cat grid_interp.aoc_9km.out -4: -4: *** WAVEWATCH III Grid interpolation *** -4: =============================================== -4: -4: Comment character is '$' -4: -4: Time Information : -4: --------------------------------------------- -4: Starting Time : 2021/03/27 15:00:00 UTC -4: Interval (in sec) : 3600.00 -4: Number of requests : 9999 -4: --------------------------------------------- -4: Number of grids (including output grid) = 2 -4: -4: -4: Extension for grid 1 is --> uglo_100km -4: -4: Grid Particulars are : -4: Dimensions = 45166 1 -4: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Extension for grid 2 is --> aoc_9km -4: -4: Grid Particulars are : -4: Dimensions = 1006 1006 -4: Grid Type = 2 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -4: -4: -4: Preparing interpolation weights for output grid -4: Total number of wet points for interpolation 360052 -4: -4: -4: Variable: Grid Interpolation Map Units: 0.100E+01 -4: -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: *** | | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: 880 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 | -4: 754 | 0 0 0 | -4: | 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 628 | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 | -4: 502 | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 376 | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 250 | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 124 | 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 | -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: -4: -4: Interpolating fields .... -4: -4: Output group 1 -4: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -4: Output group 2 -4: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -4: Output group 3 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 4 -4: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -4: Output group 5 -4: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -4: Output group 6 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 7 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 8 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 9 -4: Output variables skipped -4: Output group 10 -4: Output variables skipped -4: ------------------------------------------------ -4: 1Current vel. -4: 1Wind speed -4: 1Ice concentration -4: 2Wave height -4: 2Mean wave period(+2) -4: 2Mean wave period(+1) -4: 2Peak frequency -4: 2Mean wave dir. a1b1 -4: 2Peak direction -4: 4Part. wave height -4: 4Part. peak period -4: 4Part. mean direction -4: 5Charnock parameter -4: ------------------------------------------------ -4: OUTPUT TIME : 2021/03/27 15:00:00 UTC -4: -4: End of file reached -4: -4: -4: *** End of Grid interpolation Routine *** -4: =============================================== -4: -4: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -4: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -4: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_aoc_9km/out_grd.aoc_9km /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.aoc_9km -4: + cmdfile.5[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032715 99 arctic 9km 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + cmdfile.5[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib2_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grib2_sbs.sh[30]grdID=aoc_9km -4: + wave_grib2_sbs.sh[31]GRIDNR=255 -4: + wave_grib2_sbs.sh[32]MODNR=11 -4: + wave_grib2_sbs.sh[33]valid_time=2021032715 -4: + wave_grib2_sbs.sh[34]fhr=99 -4: + wave_grib2_sbs.sh[35]grid_region=arctic -4: + wave_grib2_sbs.sh[36]grid_res=9km -4: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676 -4: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_aoc_9km -4: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_aoc_9km -4: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_aoc_9km -4: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_aoc_9km -4: ++ wave_grib2_sbs.sh[47]printf %03i 99 -4: + wave_grib2_sbs.sh[47]FH3=099 -4: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_arctic_9km -4: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.arctic.9km.f099.grib2 -4: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f099.grib2 ]] -4: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mod_def.aoc_9km ./mod_def.ww3 -4: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.aoc_9km ./out_grd.ww3 -4: + wave_grib2_sbs.sh[73]ngrib=1 -4: + wave_grib2_sbs.sh[74]dtgrib=3600 -4: + wave_grib2_sbs.sh[75]tstart='20210327 150000' -4: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 150000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.aoc_9km.inp.tmpl -4: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -4: $ WAVEWATCH-III gridded output input file -4: $ ---------------------------------------- -4: 20210327 150000 3600 1 -4: N -4: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -4: $ -4: 20210327 150000 7 11 255 0 20 -4: $ -4: 70 0 9.0 9.0 64 -4: $ 60 0 8.64919046313 8.64919046313 64 -4: $ end of input file -4: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[88]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.321351 ']' -4: ++ prep_step[4]echo gfs_ww3_grib.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[90]export err=0 -4: + wave_grib2_sbs.sh[90]err=0 -4: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[95]cat grib2_arctic_099.out -4: -4: *** WAVEWATCH III GRIB output postp. *** -4: ============================================== -4: -4: Comment character is '$' -4: -4: Grid name : Arctic Ocean PolarStereo 9km -4: -4: LINEIN: -4: 20210327 150000 3600 1 -4: -4: 20210327150000 3600 1 -4: GEN_PRO -99999 -4: -4: Output time data : -4: ----------------------------------------------------- -4: First time : 2021/03/27 15:00:00 UTC -4: Interval : 01:00:00 -4: Number of requests : 1 -4: Fields : Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: Requested output fields not yet available: -4: ----------------------------------------------------- -4: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -4: -4: Successfully requested output fields : -4: ----------------------------------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: -4: Additional GRIB parameters : -4: ----------------------------------------------------- -4: Run time : 2021/03/27 15:00:00 UTC -4: GRIB center ID : 7 -4: GRIB gen. proc. ID : 11 -4: GRIB grid ID : 255 -4: GRIB GDS parameter : 0 -4: Fields in file : -4: -------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: CHOSEN GRID TYPE: : POLARSTEREO -4: -4: -4: -4: Generating file -4: ----------------------------------------------------- -4: Data for 2021/03/27 15:00:00 UTC 0H forecast. -4: -4: End of program -4: ========================================= -4: WAVEWATCH III GRIB output -4: -4: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -4: + wave_grib2_sbs.sh[102][[ 99 -gt 0 ]] -4: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '99 hour fcst' -grib gfs.wave.t12z.arctic.9km.f099.grib2 -4: 1:0:d=2021032312:SPC:surface:99 hour fcst: -4: 2:190532:d=2021032312:DIRC:surface:99 hour fcst: -4: 3:660328:d=2021032312:UOGRD:surface:99 hour fcst: -4: 4:847256:d=2021032312:VOGRD:surface:99 hour fcst: -4: 5:1038341:d=2021032312:WIND:surface:99 hour fcst: -4: 6:1340236:d=2021032312:WDIR:surface:99 hour fcst: -4: 7:1772308:d=2021032312:UGRD:surface:99 hour fcst: -4: 8:2070334:d=2021032312:VGRD:surface:99 hour fcst: -4: 9:2372633:d=2021032312:ICEC:surface:99 hour fcst: -4: 10:2542389:d=2021032312:HTSGW:surface:99 hour fcst: -4: 11:2756307:d=2021032312:IMWF:surface:99 hour fcst: -4: 12:2974435:d=2021032312:MWSPER:surface:99 hour fcst: -4: 13:3194226:d=2021032312:PERPW:surface:99 hour fcst: -4: 14:3422636:d=2021032312:WWSDIR:surface:99 hour fcst: -4: 15:3729915:d=2021032312:DIRPW:surface:99 hour fcst: -4: 16:4043643:d=2021032312:WVHGT:surface:99 hour fcst: -4: 17:4253636:d=2021032312:SWELL:1 in sequence:99 hour fcst: -4: 18:4465938:d=2021032312:SWELL:2 in sequence:99 hour fcst: -4: 19:4630525:d=2021032312:SWELL:3 in sequence:99 hour fcst: -4: 20:4765318:d=2021032312:WVPER:surface:99 hour fcst: -4: 21:4985190:d=2021032312:SWPER:1 in sequence:99 hour fcst: -4: 22:5205869:d=2021032312:SWPER:2 in sequence:99 hour fcst: -4: 23:5382504:d=2021032312:SWPER:3 in sequence:99 hour fcst: -4: 24:5520714:d=2021032312:WVDIR:surface:99 hour fcst: -4: 25:5805276:d=2021032312:SWDIR:1 in sequence:99 hour fcst: -4: 26:6117917:d=2021032312:SWDIR:2 in sequence:99 hour fcst: -4: 27:6337650:d=2021032312:SWDIR:3 in sequence:99 hour fcst: -4: + wave_grib2_sbs.sh[104]err=0 -4: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.arctic.9km.f099.grib2 -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f099.grib2 ]] -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f099.grib2.idx ]] -4: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.arctic.9km.f099.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f099.grib2 -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f099.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f099.grib2 = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f099.grib2 ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f099.grib2 -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f099.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f099.grib2.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f099.grib2.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f099.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f099.grib2 -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.arctic.9km.f099.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f099.grib2.idx -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f099.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f099.grib2.idx = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f099.grib2.idx ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f099.grib2.idx -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f099.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f099.grib2.idx.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f099.grib2.idx.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f099.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f099.grib2.idx -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.arctic.9km.f099.grib2 and gfs.wave.t12z.arctic.9km.f099.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_aoc_9km to COM' -4: INFO: Copied gfs.wave.t12z.arctic.9km.f099.grib2 and gfs.wave.t12z.arctic.9km.f099.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_aoc_9km to COM -4: + wave_grib2_sbs.sh[130][[ aoc_9km == '' ]] -4: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -4: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.arctic.9km.f099.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -4: INFO: gfs.wave.t12z.arctic.9km.f099.grib2 is global.0p50 or SENDDBN is NO, no alert sent -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + cmdfile.3[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032715 3600. 9999 -2: + cmdfile.3[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grid_interp_sbs.sh[25]grdID=wc_10m -2: + wave_grid_interp_sbs.sh[26]valid_time=2021032715 -2: + wave_grid_interp_sbs.sh[27]dt=3600. -2: + wave_grid_interp_sbs.sh[28]nst=9999 -2: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676 -2: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.uglo_100km ./out_grd.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mod_def.uglo_100km ./mod_def.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mod_def.wc_10m ./mod_def.wc_10m -2: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ]] -2: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m'\''' -2: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m' -2: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[51]weights_found=1 -2: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 150000' -2: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 150000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/wc_10m/g ww3_gint.inp.tmpl -2: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -2: $ Input file for interpolation of uglo_100km to wc_10m -2: $------------------------------------------------ -2: $ Start Time 3600. NSteps -2: 20210327 150000 3600. 9999 -2: $ Total number of grids -2: 2 -2: $ Grid extensions -2: 'uglo_100km' -2: 'wc_10m' -2: $ -2: 0 -2: $ -2: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[70]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.321351 ']' -2: ++ prep_step[4]echo gfs_ww3_gint.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -2: INFO: Executing 'gfs_ww3_gint.x' -2: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[73]cat grid_interp.wc_10m.out -2: -2: *** WAVEWATCH III Grid interpolation *** -2: =============================================== -2: -2: Comment character is '$' -2: -2: Time Information : -2: --------------------------------------------- -2: Starting Time : 2021/03/27 15:00:00 UTC -2: Interval (in sec) : 3600.00 -2: Number of requests : 9999 -2: --------------------------------------------- -2: Number of grids (including output grid) = 2 -2: -2: -2: Extension for grid 1 is --> uglo_100km -2: -2: Grid Particulars are : -2: Dimensions = 45166 1 -2: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Extension for grid 2 is --> wc_10m -2: -2: Grid Particulars are : -2: Dimensions = 241 151 -2: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -2: -2: -2: Preparing interpolation weights for output grid -2: Total number of wet points for interpolation 11044 -2: -2: -2: Variable: Grid Interpolation Map Units: 0.100E+01 -2: -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: +---------------------------------------------------------------------------------------------------------------+ -2: 151 | 0 0 0 0 | -2: | 0 0 0 0 0 | -2: | 0 0 0 0 0 | -2: 130 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 109 | 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 88 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 | -2: 67 | 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 | -2: 46 | 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: 25 | 0 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 0 | -2: | | -2: 4 | | -2: +---------------------------------------------------------------------------------------------------------------+ -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: -2: -2: Interpolating fields .... -2: -2: Output group 1 -2: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -2: Output group 2 -2: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -2: Output group 3 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 4 -2: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -2: Output group 5 -2: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -2: Output group 6 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 7 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 8 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 9 -2: Output variables skipped -2: Output group 10 -2: Output variables skipped -2: ------------------------------------------------ -2: 1Current vel. -2: 1Wind speed -2: 1Ice concentration -2: 2Wave height -2: 2Mean wave period(+2) -2: 2Mean wave period(+1) -2: 2Peak frequency -2: 2Mean wave dir. a1b1 -2: 2Peak direction -2: 4Part. wave height -2: 4Part. peak period -2: 4Part. mean direction -2: 5Charnock parameter -2: ------------------------------------------------ -2: OUTPUT TIME : 2021/03/27 15:00:00 UTC -2: -2: End of file reached -2: -2: -2: *** End of Grid interpolation Routine *** -2: =============================================== -2: -2: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -2: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -2: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_wc_10m/out_grd.wc_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.wc_10m -2: + cmdfile.3[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032715 99 wcoast 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + cmdfile.3[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib2_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grib2_sbs.sh[30]grdID=wc_10m -2: + wave_grib2_sbs.sh[31]GRIDNR=255 -2: + wave_grib2_sbs.sh[32]MODNR=11 -2: + wave_grib2_sbs.sh[33]valid_time=2021032715 -2: + wave_grib2_sbs.sh[34]fhr=99 -2: + wave_grib2_sbs.sh[35]grid_region=wcoast -2: + wave_grib2_sbs.sh[36]grid_res=0p16 -2: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676 -2: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_wc_10m -2: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_wc_10m -2: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_wc_10m -2: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_wc_10m -2: ++ wave_grib2_sbs.sh[47]printf %03i 99 -2: + wave_grib2_sbs.sh[47]FH3=099 -2: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_wcoast_0p16 -2: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.wcoast.0p16.f099.grib2 -2: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f099.grib2 ]] -2: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mod_def.wc_10m ./mod_def.ww3 -2: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.wc_10m ./out_grd.ww3 -2: + wave_grib2_sbs.sh[73]ngrib=1 -2: + wave_grib2_sbs.sh[74]dtgrib=3600 -2: + wave_grib2_sbs.sh[75]tstart='20210327 150000' -2: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 150000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.wc_10m.inp.tmpl -2: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -2: $ WAVEWATCH-III gridded output input file -2: $ ---------------------------------------- -2: 20210327 150000 3600 1 -2: N -2: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -2: $ -2: 20210327 150000 7 11 255 0 0 -2: $ -2: $ end of input file -2: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[88]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.321351 ']' -2: ++ prep_step[4]echo gfs_ww3_grib.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[90]export err=0 -2: + wave_grib2_sbs.sh[90]err=0 -2: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[95]cat grib2_wcoast_099.out -2: -2: *** WAVEWATCH III GRIB output postp. *** -2: ============================================== -2: -2: Comment character is '$' -2: -2: Grid name : West Coast 10 min wave grid -2: -2: LINEIN: -2: 20210327 150000 3600 1 -2: -2: 20210327150000 3600 1 -2: GEN_PRO -99999 -2: -2: Output time data : -2: ----------------------------------------------------- -2: First time : 2021/03/27 15:00:00 UTC -2: Interval : 01:00:00 -2: Number of requests : 1 -2: Fields : Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: Requested output fields not yet available: -2: ----------------------------------------------------- -2: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -2: -2: Successfully requested output fields : -2: ----------------------------------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: -2: Additional GRIB parameters : -2: ----------------------------------------------------- -2: Run time : 2021/03/27 15:00:00 UTC -2: GRIB center ID : 7 -2: GRIB gen. proc. ID : 11 -2: GRIB grid ID : 255 -2: GRIB GDS parameter : 0 -2: Fields in file : -2: -------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: CHOSEN GRID TYPE: : LLRECTILINEAR -2: -2: -2: -2: Generating file -2: ----------------------------------------------------- -2: Data for 2021/03/27 15:00:00 UTC 0H forecast. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: -2: End of program -2: ========================================= -2: WAVEWATCH III GRIB output -2: -2: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -2: + wave_grib2_sbs.sh[102][[ 99 -gt 0 ]] -2: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '99 hour fcst' -grib gfs.wave.t12z.wcoast.0p16.f099.grib2 -2: 1:0:d=2021032312:SPC:surface:99 hour fcst: -2: 2:7432:d=2021032312:DIRC:surface:99 hour fcst: -2: 3:25086:d=2021032312:UOGRD:surface:99 hour fcst: -2: 4:32192:d=2021032312:VOGRD:surface:99 hour fcst: -2: 5:39761:d=2021032312:WIND:surface:99 hour fcst: -2: 6:51647:d=2021032312:WDIR:surface:99 hour fcst: -2: 7:68426:d=2021032312:UGRD:surface:99 hour fcst: -2: 8:79431:d=2021032312:VGRD:surface:99 hour fcst: -2: 9:91387:d=2021032312:ICEC:surface:99 hour fcst: -2: 10:96117:d=2021032312:HTSGW:surface:99 hour fcst: -2: 11:105599:d=2021032312:IMWF:surface:99 hour fcst: -2: 12:115262:d=2021032312:MWSPER:surface:99 hour fcst: -2: 13:125006:d=2021032312:PERPW:surface:99 hour fcst: -2: 14:135069:d=2021032312:WWSDIR:surface:99 hour fcst: -2: 15:150545:d=2021032312:DIRPW:surface:99 hour fcst: -2: 16:166418:d=2021032312:WVHGT:surface:99 hour fcst: -2: 17:174023:d=2021032312:SWELL:1 in sequence:99 hour fcst: -2: 18:182268:d=2021032312:SWELL:2 in sequence:99 hour fcst: -2: 19:187498:d=2021032312:SWELL:3 in sequence:99 hour fcst: -2: 20:192228:d=2021032312:WVPER:surface:99 hour fcst: -2: 21:200268:d=2021032312:SWPER:1 in sequence:99 hour fcst: -2: 22:208606:d=2021032312:SWPER:2 in sequence:99 hour fcst: -2: 23:214050:d=2021032312:SWPER:3 in sequence:99 hour fcst: -2: 24:218780:d=2021032312:WVDIR:surface:99 hour fcst: -2: 25:229542:d=2021032312:SWDIR:1 in sequence:99 hour fcst: -2: 26:241685:d=2021032312:SWDIR:2 in sequence:99 hour fcst: -2: 27:247421:d=2021032312:SWDIR:3 in sequence:99 hour fcst: -2: + wave_grib2_sbs.sh[104]err=0 -2: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.wcoast.0p16.f099.grib2 -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f099.grib2 ]] -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f099.grib2.idx ]] -2: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.wcoast.0p16.f099.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f099.grib2 -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f099.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f099.grib2 = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f099.grib2 ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f099.grib2 -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f099.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f099.grib2.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f099.grib2.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f099.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f099.grib2 -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.wcoast.0p16.f099.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f099.grib2.idx -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f099.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f099.grib2.idx = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f099.grib2.idx ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f099.grib2.idx -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f099.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f099.grib2.idx.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f099.grib2.idx.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f099.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f099.grib2.idx -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.wcoast.0p16.f099.grib2 and gfs.wave.t12z.wcoast.0p16.f099.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_wc_10m to COM' -2: INFO: Copied gfs.wave.t12z.wcoast.0p16.f099.grib2 and gfs.wave.t12z.wcoast.0p16.f099.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_wc_10m to COM -2: + wave_grib2_sbs.sh[130][[ wc_10m == '' ]] -2: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -2: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.wcoast.0p16.f099.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -2: INFO: gfs.wave.t12z.wcoast.0p16.f099.grib2 is global.0p50 or SENDDBN is NO, no alert sent -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + cmdfile.2[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032715 3600. 9999 -1: + cmdfile.2[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grid_interp_sbs.sh[25]grdID=ep_10m -1: + wave_grid_interp_sbs.sh[26]valid_time=2021032715 -1: + wave_grid_interp_sbs.sh[27]dt=3600. -1: + wave_grid_interp_sbs.sh[28]nst=9999 -1: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676 -1: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.uglo_100km ./out_grd.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mod_def.uglo_100km ./mod_def.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mod_def.ep_10m ./mod_def.ep_10m -1: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ]] -1: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m'\''' -1: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m' -1: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[51]weights_found=1 -1: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 150000' -1: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 150000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/ep_10m/g ww3_gint.inp.tmpl -1: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -1: $ Input file for interpolation of uglo_100km to ep_10m -1: $------------------------------------------------ -1: $ Start Time 3600. NSteps -1: 20210327 150000 3600. 9999 -1: $ Total number of grids -1: 2 -1: $ Grid extensions -1: 'uglo_100km' -1: 'ep_10m' -1: $ -1: 0 -1: $ -1: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[70]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.321351 ']' -1: ++ prep_step[4]echo gfs_ww3_gint.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -1: INFO: Executing 'gfs_ww3_gint.x' -1: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[73]cat grid_interp.ep_10m.out -1: -1: *** WAVEWATCH III Grid interpolation *** -1: =============================================== -1: -1: Comment character is '$' -1: -1: Time Information : -1: --------------------------------------------- -1: Starting Time : 2021/03/27 15:00:00 UTC -1: Interval (in sec) : 3600.00 -1: Number of requests : 9999 -1: --------------------------------------------- -1: Number of grids (including output grid) = 2 -1: -1: -1: Extension for grid 1 is --> uglo_100km -1: -1: Grid Particulars are : -1: Dimensions = 45166 1 -1: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Extension for grid 2 is --> ep_10m -1: -1: Grid Particulars are : -1: Dimensions = 511 301 -1: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -1: -1: -1: Preparing interpolation weights for output grid -1: Total number of wet points for interpolation 7439 -1: -1: -1: Variable: Grid Interpolation Map Units: 0.100E+01 -1: -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 301 | | -1: | | -1: | 0 0 | -1: 262 | 0 0 0 0 | -1: | 0 0 0 0 | -1: | 0 0 0 0 | -1: 223 | 0 0 0 | -1: | 0 | -1: | 0 | -1: 184 | | -1: | 0 0 0 | -1: | 0 0 0 0 | -1: 145 | 0 | -1: | | -1: | | -1: 106 | | -1: | | -1: | | -1: 67 | | -1: | 0 | -1: | | -1: 28 | 0 | -1: | | -1: | | -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: -1: -1: Interpolating fields .... -1: -1: Output group 1 -1: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -1: Output group 2 -1: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -1: Output group 3 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 4 -1: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -1: Output group 5 -1: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -1: Output group 6 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 7 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 8 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 9 -1: Output variables skipped -1: Output group 10 -1: Output variables skipped -1: ------------------------------------------------ -1: 1Current vel. -1: 1Wind speed -1: 1Ice concentration -1: 2Wave height -1: 2Mean wave period(+2) -1: 2Mean wave period(+1) -1: 2Peak frequency -1: 2Mean wave dir. a1b1 -1: 2Peak direction -1: 4Part. wave height -1: 4Part. peak period -1: 4Part. mean direction -1: 5Charnock parameter -1: ------------------------------------------------ -1: OUTPUT TIME : 2021/03/27 15:00:00 UTC -1: -1: End of file reached -1: -1: -1: *** End of Grid interpolation Routine *** -1: =============================================== -1: -1: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -1: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -1: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_ep_10m/out_grd.ep_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.ep_10m -1: + cmdfile.2[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032715 99 epacif 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + cmdfile.2[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib2_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grib2_sbs.sh[30]grdID=ep_10m -1: + wave_grib2_sbs.sh[31]GRIDNR=255 -1: + wave_grib2_sbs.sh[32]MODNR=11 -1: + wave_grib2_sbs.sh[33]valid_time=2021032715 -1: + wave_grib2_sbs.sh[34]fhr=99 -1: + wave_grib2_sbs.sh[35]grid_region=epacif -1: + wave_grib2_sbs.sh[36]grid_res=0p16 -1: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676 -1: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_ep_10m -1: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_ep_10m -1: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_ep_10m -1: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_ep_10m -1: ++ wave_grib2_sbs.sh[47]printf %03i 99 -1: + wave_grib2_sbs.sh[47]FH3=099 -1: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_epacif_0p16 -1: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.epacif.0p16.f099.grib2 -1: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f099.grib2 ]] -1: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mod_def.ep_10m ./mod_def.ww3 -1: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.ep_10m ./out_grd.ww3 -1: + wave_grib2_sbs.sh[73]ngrib=1 -1: + wave_grib2_sbs.sh[74]dtgrib=3600 -1: + wave_grib2_sbs.sh[75]tstart='20210327 150000' -1: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 150000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.ep_10m.inp.tmpl -1: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -1: $ WAVEWATCH-III gridded output input file -1: $ ---------------------------------------- -1: 20210327 150000 3600 1 -1: N -1: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -1: $ -1: 20210327 150000 7 11 255 0 0 -1: $ -1: $ end of input file -1: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[88]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.321351 ']' -1: ++ prep_step[4]echo gfs_ww3_grib.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[90]export err=0 -1: + wave_grib2_sbs.sh[90]err=0 -1: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[95]cat grib2_epacif_099.out -1: -1: *** WAVEWATCH III GRIB output postp. *** -1: ============================================== -1: -1: Comment character is '$' -1: -1: Grid name : East Pacific 10 min wave grid -1: -1: LINEIN: -1: 20210327 150000 3600 1 -1: -1: 20210327150000 3600 1 -1: GEN_PRO -99999 -1: -1: Output time data : -1: ----------------------------------------------------- -1: First time : 2021/03/27 15:00:00 UTC -1: Interval : 01:00:00 -1: Number of requests : 1 -1: Fields : Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: Requested output fields not yet available: -1: ----------------------------------------------------- -1: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -1: -1: Successfully requested output fields : -1: ----------------------------------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: -1: Additional GRIB parameters : -1: ----------------------------------------------------- -1: Run time : 2021/03/27 15:00:00 UTC -1: GRIB center ID : 7 -1: GRIB gen. proc. ID : 11 -1: GRIB grid ID : 255 -1: GRIB GDS parameter : 0 -1: Fields in file : -1: -------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: CHOSEN GRID TYPE: : LLRECTILINEAR -1: -1: -1: -1: Generating file -1: ----------------------------------------------------- -1: Data for 2021/03/27 15:00:00 UTC 0H forecast. -1: Warning: bitmask off everywhere. -1: Pretend one point in jpcpack to avoid crash. -1: Warning: bitmask off everywhere. -1: Pretend one point in jpcpack to avoid crash. -1: Warning: bitmask off everywhere. -1: Pretend one point in jpcpack to avoid crash. -1: -1: End of program -1: ========================================= -1: WAVEWATCH III GRIB output -1: -1: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -1: + wave_grib2_sbs.sh[102][[ 99 -gt 0 ]] -1: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '99 hour fcst' -grib gfs.wave.t12z.epacif.0p16.f099.grib2 -1: 1:0:d=2021032312:SPC:surface:99 hour fcst: -1: 2:21837:d=2021032312:DIRC:surface:99 hour fcst: -1: 3:50128:d=2021032312:UOGRD:surface:99 hour fcst: -1: 4:71952:d=2021032312:VOGRD:surface:99 hour fcst: -1: 5:93925:d=2021032312:WIND:surface:99 hour fcst: -1: 6:118550:d=2021032312:WDIR:surface:99 hour fcst: -1: 7:145836:d=2021032312:UGRD:surface:99 hour fcst: -1: 8:170474:d=2021032312:VGRD:surface:99 hour fcst: -1: 9:195072:d=2021032312:ICEC:surface:99 hour fcst: -1: 10:214480:d=2021032312:HTSGW:surface:99 hour fcst: -1: 11:236872:d=2021032312:IMWF:surface:99 hour fcst: -1: 12:259790:d=2021032312:MWSPER:surface:99 hour fcst: -1: 13:282766:d=2021032312:PERPW:surface:99 hour fcst: -1: 14:305655:d=2021032312:WWSDIR:surface:99 hour fcst: -1: 15:331586:d=2021032312:DIRPW:surface:99 hour fcst: -1: 16:357666:d=2021032312:WVHGT:surface:99 hour fcst: -1: 17:379522:d=2021032312:SWELL:1 in sequence:99 hour fcst: -1: 18:402429:d=2021032312:SWELL:2 in sequence:99 hour fcst: -1: 19:422845:d=2021032312:SWELL:3 in sequence:99 hour fcst: -1: 20:442253:d=2021032312:WVPER:surface:99 hour fcst: -1: 21:464147:d=2021032312:SWPER:1 in sequence:99 hour fcst: -1: 22:487124:d=2021032312:SWPER:2 in sequence:99 hour fcst: -1: 23:507479:d=2021032312:SWPER:3 in sequence:99 hour fcst: -1: 24:526887:d=2021032312:WVDIR:surface:99 hour fcst: -1: 25:550855:d=2021032312:SWDIR:1 in sequence:99 hour fcst: -1: 26:578087:d=2021032312:SWDIR:2 in sequence:99 hour fcst: -1: 27:599394:d=2021032312:SWDIR:3 in sequence:99 hour fcst: -1: + wave_grib2_sbs.sh[104]err=0 -1: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.epacif.0p16.f099.grib2 -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f099.grib2 ]] -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f099.grib2.idx ]] -1: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.epacif.0p16.f099.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f099.grib2 -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f099.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f099.grib2 = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f099.grib2 ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f099.grib2 -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f099.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f099.grib2.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f099.grib2.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f099.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f099.grib2 -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.epacif.0p16.f099.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f099.grib2.idx -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f099.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f099.grib2.idx = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f099.grib2.idx ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f099.grib2.idx -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f099.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f099.grib2.idx.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f099.grib2.idx.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f099.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f099.grib2.idx -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.epacif.0p16.f099.grib2 and gfs.wave.t12z.epacif.0p16.f099.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_ep_10m to COM' -1: INFO: Copied gfs.wave.t12z.epacif.0p16.f099.grib2 and gfs.wave.t12z.epacif.0p16.f099.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_ep_10m to COM -1: + wave_grib2_sbs.sh[130][[ ep_10m == '' ]] -1: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -1: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.epacif.0p16.f099.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -1: INFO: gfs.wave.t12z.epacif.0p16.f099.grib2 is global.0p50 or SENDDBN is NO, no alert sent -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + cmdfile.4[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032715 3600. 9999 -3: + cmdfile.4[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grid_interp_sbs.sh[25]grdID=glo_30m -3: + wave_grid_interp_sbs.sh[26]valid_time=2021032715 -3: + wave_grid_interp_sbs.sh[27]dt=3600. -3: + wave_grid_interp_sbs.sh[28]nst=9999 -3: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676 -3: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.uglo_100km ./out_grd.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mod_def.uglo_100km ./mod_def.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mod_def.glo_30m ./mod_def.glo_30m -3: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ]] -3: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m'\''' -3: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m' -3: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[51]weights_found=1 -3: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 150000' -3: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 150000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/glo_30m/g ww3_gint.inp.tmpl -3: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -3: $ Input file for interpolation of uglo_100km to glo_30m -3: $------------------------------------------------ -3: $ Start Time 3600. NSteps -3: 20210327 150000 3600. 9999 -3: $ Total number of grids -3: 2 -3: $ Grid extensions -3: 'uglo_100km' -3: 'glo_30m' -3: $ -3: 0 -3: $ -3: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[70]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.321351 ']' -3: ++ prep_step[4]echo gfs_ww3_gint.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -3: INFO: Executing 'gfs_ww3_gint.x' -3: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[73]cat grid_interp.glo_30m.out -3: -3: *** WAVEWATCH III Grid interpolation *** -3: =============================================== -3: -3: Comment character is '$' -3: -3: Time Information : -3: --------------------------------------------- -3: Starting Time : 2021/03/27 15:00:00 UTC -3: Interval (in sec) : 3600.00 -3: Number of requests : 9999 -3: --------------------------------------------- -3: Number of grids (including output grid) = 2 -3: -3: -3: Extension for grid 1 is --> uglo_100km -3: -3: Grid Particulars are : -3: Dimensions = 45166 1 -3: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Extension for grid 2 is --> glo_30m -3: -3: Grid Particulars are : -3: Dimensions = 720 336 -3: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -3: -3: -3: Preparing interpolation weights for output grid -3: Total number of wet points for interpolation 167619 -3: -3: -3: Variable: Grid Interpolation Map Units: 0.100E+01 -3: -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 336 | | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 291 | 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 | -3: 246 | 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 201 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 156 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 111 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 66 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 21 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 | -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: -3: -3: Interpolating fields .... -3: -3: Output group 1 -3: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -3: Output group 2 -3: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -3: Output group 3 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 4 -3: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -3: Output group 5 -3: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -3: Output group 6 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 7 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 8 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 9 -3: Output variables skipped -3: Output group 10 -3: Output variables skipped -3: ------------------------------------------------ -3: 1Current vel. -3: 1Wind speed -3: 1Ice concentration -3: 2Wave height -3: 2Mean wave period(+2) -3: 2Mean wave period(+1) -3: 2Peak frequency -3: 2Mean wave dir. a1b1 -3: 2Peak direction -3: 4Part. wave height -3: 4Part. peak period -3: 4Part. mean direction -3: 5Charnock parameter -3: ------------------------------------------------ -3: OUTPUT TIME : 2021/03/27 15:00:00 UTC -3: -3: End of file reached -3: -3: -3: *** End of Grid interpolation Routine *** -3: =============================================== -3: -3: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -3: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -3: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_glo_30m/out_grd.glo_30m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.glo_30m -3: + cmdfile.4[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032715 99 global 0p50 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + cmdfile.4[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib2_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grib2_sbs.sh[30]grdID=glo_30m -3: + wave_grib2_sbs.sh[31]GRIDNR=255 -3: + wave_grib2_sbs.sh[32]MODNR=11 -3: + wave_grib2_sbs.sh[33]valid_time=2021032715 -3: + wave_grib2_sbs.sh[34]fhr=99 -3: + wave_grib2_sbs.sh[35]grid_region=global -3: + wave_grib2_sbs.sh[36]grid_res=0p50 -3: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676 -3: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_glo_30m -3: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_glo_30m -3: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_glo_30m -3: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_glo_30m -3: ++ wave_grib2_sbs.sh[47]printf %03i 99 -3: + wave_grib2_sbs.sh[47]FH3=099 -3: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p50 -3: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p50.f099.grib2 -3: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f099.grib2 ]] -3: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mod_def.glo_30m ./mod_def.ww3 -3: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.glo_30m ./out_grd.ww3 -3: + wave_grib2_sbs.sh[73]ngrib=1 -3: + wave_grib2_sbs.sh[74]dtgrib=3600 -3: + wave_grib2_sbs.sh[75]tstart='20210327 150000' -3: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 150000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.glo_30m.inp.tmpl -3: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -3: $ WAVEWATCH-III gridded output input file -3: $ ---------------------------------------- -3: 20210327 150000 3600 1 -3: N -3: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -3: $ -3: 20210327 150000 7 11 255 0 0 -3: $ -3: $ end of input file -3: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[88]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.321351 ']' -3: ++ prep_step[4]echo gfs_ww3_grib.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[90]export err=0 -3: + wave_grib2_sbs.sh[90]err=0 -3: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[95]cat grib2_global_099.out -3: -3: *** WAVEWATCH III GRIB output postp. *** -3: ============================================== -3: -3: Comment character is '$' -3: -3: Grid name : Global 30 min wave grid -3: -3: LINEIN: -3: 20210327 150000 3600 1 -3: -3: 20210327150000 3600 1 -3: GEN_PRO -99999 -3: -3: Output time data : -3: ----------------------------------------------------- -3: First time : 2021/03/27 15:00:00 UTC -3: Interval : 01:00:00 -3: Number of requests : 1 -3: Fields : Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: Requested output fields not yet available: -3: ----------------------------------------------------- -3: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -3: -3: Successfully requested output fields : -3: ----------------------------------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: -3: Additional GRIB parameters : -3: ----------------------------------------------------- -3: Run time : 2021/03/27 15:00:00 UTC -3: GRIB center ID : 7 -3: GRIB gen. proc. ID : 11 -3: GRIB grid ID : 255 -3: GRIB GDS parameter : 0 -3: Fields in file : -3: -------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: CHOSEN GRID TYPE: : LLRECTILINEAR -3: -3: -3: -3: Generating file -3: ----------------------------------------------------- -3: Data for 2021/03/27 15:00:00 UTC 0H forecast. -3: -3: End of program -3: ========================================= -3: WAVEWATCH III GRIB output -3: -3: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -3: + wave_grib2_sbs.sh[102][[ 99 -gt 0 ]] -3: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '99 hour fcst' -grib gfs.wave.t12z.global.0p50.f099.grib2 -3: 1:0:d=2021032312:SPC:surface:99 hour fcst: -3: 2:77636:d=2021032312:DIRC:surface:99 hour fcst: -3: 3:322375:d=2021032312:UOGRD:surface:99 hour fcst: -3: 4:397673:d=2021032312:VOGRD:surface:99 hour fcst: -3: 5:477004:d=2021032312:WIND:surface:99 hour fcst: -3: 6:635652:d=2021032312:WDIR:surface:99 hour fcst: -3: 7:871776:d=2021032312:UGRD:surface:99 hour fcst: -3: 8:1026292:d=2021032312:VGRD:surface:99 hour fcst: -3: 9:1182740:d=2021032312:ICEC:surface:99 hour fcst: -3: 10:1222982:d=2021032312:HTSGW:surface:99 hour fcst: -3: 11:1321875:d=2021032312:IMWF:surface:99 hour fcst: -3: 12:1431701:d=2021032312:MWSPER:surface:99 hour fcst: -3: 13:1542261:d=2021032312:PERPW:surface:99 hour fcst: -3: 14:1659206:d=2021032312:WWSDIR:surface:99 hour fcst: -3: 15:1859550:d=2021032312:DIRPW:surface:99 hour fcst: -3: 16:2065312:d=2021032312:WVHGT:surface:99 hour fcst: -3: 17:2160566:d=2021032312:SWELL:1 in sequence:99 hour fcst: -3: 18:2265174:d=2021032312:SWELL:2 in sequence:99 hour fcst: -3: 19:2337921:d=2021032312:SWELL:3 in sequence:99 hour fcst: -3: 20:2381647:d=2021032312:WVPER:surface:99 hour fcst: -3: 21:2488249:d=2021032312:SWPER:1 in sequence:99 hour fcst: -3: 22:2606280:d=2021032312:SWPER:2 in sequence:99 hour fcst: -3: 23:2697860:d=2021032312:SWPER:3 in sequence:99 hour fcst: -3: 24:2749858:d=2021032312:WVDIR:surface:99 hour fcst: -3: 25:2911040:d=2021032312:SWDIR:1 in sequence:99 hour fcst: -3: 26:3113819:d=2021032312:SWDIR:2 in sequence:99 hour fcst: -3: 27:3258561:d=2021032312:SWDIR:3 in sequence:99 hour fcst: -3: + wave_grib2_sbs.sh[104]err=0 -3: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p50.f099.grib2 -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f099.grib2 ]] -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f099.grib2.idx ]] -3: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p50.f099.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f099.grib2 -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f099.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f099.grib2 = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f099.grib2 ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f099.grib2 -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f099.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f099.grib2.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f099.grib2.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f099.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f099.grib2 -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p50.f099.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f099.grib2.idx -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f099.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f099.grib2.idx = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f099.grib2.idx ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f099.grib2.idx -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f099.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f099.grib2.idx.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f099.grib2.idx.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f099.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f099.grib2.idx -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p50.f099.grib2 and gfs.wave.t12z.global.0p50.f099.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_glo_30m to COM' -3: INFO: Copied gfs.wave.t12z.global.0p50.f099.grib2 and gfs.wave.t12z.global.0p50.f099.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_glo_30m to COM -3: + wave_grib2_sbs.sh[130][[ glo_30m == '' ]] -3: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -3: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p50.f099.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -3: INFO: gfs.wave.t12z.global.0p50.f099.grib2 is global.0p50 or SENDDBN is NO, no alert sent -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + cmdfile.6[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032715 3600. 9999 -5: + cmdfile.6[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grid_interp_sbs.sh[25]grdID=gnh_10m -5: + wave_grid_interp_sbs.sh[26]valid_time=2021032715 -5: + wave_grid_interp_sbs.sh[27]dt=3600. -5: + wave_grid_interp_sbs.sh[28]nst=9999 -5: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676 -5: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.uglo_100km ./out_grd.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mod_def.uglo_100km ./mod_def.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mod_def.gnh_10m ./mod_def.gnh_10m -5: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m'\''' -5: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m' -5: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[51]weights_found=1 -5: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 150000' -5: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 150000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gnh_10m/g ww3_gint.inp.tmpl -5: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -5: $ Input file for interpolation of uglo_100km to gnh_10m -5: $------------------------------------------------ -5: $ Start Time 3600. NSteps -5: 20210327 150000 3600. 9999 -5: $ Total number of grids -5: 2 -5: $ Grid extensions -5: 'uglo_100km' -5: 'gnh_10m' -5: $ -5: 0 -5: $ -5: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[70]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.321351 ']' -5: ++ prep_step[4]echo gfs_ww3_gint.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -5: INFO: Executing 'gfs_ww3_gint.x' -5: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[73]cat grid_interp.gnh_10m.out -5: -5: *** WAVEWATCH III Grid interpolation *** -5: =============================================== -5: -5: Comment character is '$' -5: -5: Time Information : -5: --------------------------------------------- -5: Starting Time : 2021/03/27 15:00:00 UTC -5: Interval (in sec) : 3600.00 -5: Number of requests : 9999 -5: --------------------------------------------- -5: Number of grids (including output grid) = 2 -5: -5: -5: Extension for grid 1 is --> uglo_100km -5: -5: Grid Particulars are : -5: Dimensions = 45166 1 -5: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Extension for grid 2 is --> gnh_10m -5: -5: Grid Particulars are : -5: Dimensions = 2160 406 -5: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -5: -5: -5: Preparing interpolation weights for output grid -5: Total number of wet points for interpolation 571209 -5: -5: -5: Variable: Grid Interpolation Map Units: 0.100E+01 -5: -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 406 | | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: 355 | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 304 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 253 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 202 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 151 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 100 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 49 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: -5: -5: Interpolating fields .... -5: -5: Output group 1 -5: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -5: Output group 2 -5: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -5: Output group 3 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 4 -5: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -5: Output group 5 -5: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -5: Output group 6 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 7 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 8 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 9 -5: Output variables skipped -5: Output group 10 -5: Output variables skipped -5: ------------------------------------------------ -5: 1Current vel. -5: 1Wind speed -5: 1Ice concentration -5: 2Wave height -5: 2Mean wave period(+2) -5: 2Mean wave period(+1) -5: 2Peak frequency -5: 2Mean wave dir. a1b1 -5: 2Peak direction -5: 4Part. wave height -5: 4Part. peak period -5: 4Part. mean direction -5: 5Charnock parameter -5: ------------------------------------------------ -5: OUTPUT TIME : 2021/03/27 15:00:00 UTC -5: -5: End of file reached -5: -5: -5: *** End of Grid interpolation Routine *** -5: =============================================== -5: -5: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -5: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -5: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_gnh_10m/out_grd.gnh_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.gnh_10m -5: + cmdfile.6[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032715 99 global 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + cmdfile.6[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib2_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grib2_sbs.sh[30]grdID=gnh_10m -5: + wave_grib2_sbs.sh[31]GRIDNR=255 -5: + wave_grib2_sbs.sh[32]MODNR=11 -5: + wave_grib2_sbs.sh[33]valid_time=2021032715 -5: + wave_grib2_sbs.sh[34]fhr=99 -5: + wave_grib2_sbs.sh[35]grid_region=global -5: + wave_grib2_sbs.sh[36]grid_res=0p16 -5: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676 -5: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_gnh_10m -5: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_gnh_10m -5: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_gnh_10m -5: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_gnh_10m -5: ++ wave_grib2_sbs.sh[47]printf %03i 99 -5: + wave_grib2_sbs.sh[47]FH3=099 -5: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p16 -5: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p16.f099.grib2 -5: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f099.grib2 ]] -5: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mod_def.gnh_10m ./mod_def.ww3 -5: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.gnh_10m ./out_grd.ww3 -5: + wave_grib2_sbs.sh[73]ngrib=1 -5: + wave_grib2_sbs.sh[74]dtgrib=3600 -5: + wave_grib2_sbs.sh[75]tstart='20210327 150000' -5: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 150000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gnh_10m.inp.tmpl -5: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -5: $ WAVEWATCH-III gridded output input file -5: $ ---------------------------------------- -5: 20210327 150000 3600 1 -5: N -5: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -5: $ -5: 20210327 150000 7 11 255 0 0 -5: $ -5: $ end of input file -5: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[88]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.321351 ']' -5: ++ prep_step[4]echo gfs_ww3_grib.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[90]export err=0 -5: + wave_grib2_sbs.sh[90]err=0 -5: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[95]cat grib2_global_099.out -5: -5: *** WAVEWATCH III GRIB output postp. *** -5: ============================================== -5: -5: Comment character is '$' -5: -5: Grid name : GFSv16-wave N Hemisphere 1/6 d -5: -5: LINEIN: -5: 20210327 150000 3600 1 -5: -5: 20210327150000 3600 1 -5: GEN_PRO -99999 -5: -5: Output time data : -5: ----------------------------------------------------- -5: First time : 2021/03/27 15:00:00 UTC -5: Interval : 01:00:00 -5: Number of requests : 1 -5: Fields : Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: Requested output fields not yet available: -5: ----------------------------------------------------- -5: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -5: -5: Successfully requested output fields : -5: ----------------------------------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: -5: Additional GRIB parameters : -5: ----------------------------------------------------- -5: Run time : 2021/03/27 15:00:00 UTC -5: GRIB center ID : 7 -5: GRIB gen. proc. ID : 11 -5: GRIB grid ID : 255 -5: GRIB GDS parameter : 0 -5: Fields in file : -5: -------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: CHOSEN GRID TYPE: : LLRECTILINEAR -5: -5: -5: -5: Generating file -5: ----------------------------------------------------- -5: Data for 2021/03/27 15:00:00 UTC 0H forecast. -5: -5: End of program -5: ========================================= -5: WAVEWATCH III GRIB output -5: -5: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -5: + wave_grib2_sbs.sh[102][[ 99 -gt 0 ]] -5: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '99 hour fcst' -grib gfs.wave.t12z.global.0p16.f099.grib2 -5: 1:0:d=2021032312:SPC:surface:99 hour fcst: -5: 2:237276:d=2021032312:DIRC:surface:99 hour fcst: -5: 3:872941:d=2021032312:UOGRD:surface:99 hour fcst: -5: 4:1106311:d=2021032312:VOGRD:surface:99 hour fcst: -5: 5:1344336:d=2021032312:WIND:surface:99 hour fcst: -5: 6:1753491:d=2021032312:WDIR:surface:99 hour fcst: -5: 7:2389810:d=2021032312:UGRD:surface:99 hour fcst: -5: 8:2790157:d=2021032312:VGRD:surface:99 hour fcst: -5: 9:3192074:d=2021032312:ICEC:surface:99 hour fcst: -5: 10:3306600:d=2021032312:HTSGW:surface:99 hour fcst: -5: 11:3595533:d=2021032312:IMWF:surface:99 hour fcst: -5: 12:3907610:d=2021032312:MWSPER:surface:99 hour fcst: -5: 13:4222151:d=2021032312:PERPW:surface:99 hour fcst: -5: 14:4557968:d=2021032312:WWSDIR:surface:99 hour fcst: -5: 15:5137712:d=2021032312:DIRPW:surface:99 hour fcst: -5: 16:5741615:d=2021032312:WVHGT:surface:99 hour fcst: -5: 17:6017631:d=2021032312:SWELL:1 in sequence:99 hour fcst: -5: 18:6309494:d=2021032312:SWELL:2 in sequence:99 hour fcst: -5: 19:6506408:d=2021032312:SWELL:3 in sequence:99 hour fcst: -5: 20:6633839:d=2021032312:WVPER:surface:99 hour fcst: -5: 21:6942927:d=2021032312:SWPER:1 in sequence:99 hour fcst: -5: 22:7273487:d=2021032312:SWPER:2 in sequence:99 hour fcst: -5: 23:7515689:d=2021032312:SWPER:3 in sequence:99 hour fcst: -5: 24:7658212:d=2021032312:WVDIR:surface:99 hour fcst: -5: 25:8148359:d=2021032312:SWDIR:1 in sequence:99 hour fcst: -5: 26:8742698:d=2021032312:SWDIR:2 in sequence:99 hour fcst: -5: 27:9124543:d=2021032312:SWDIR:3 in sequence:99 hour fcst: -5: + wave_grib2_sbs.sh[104]err=0 -5: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p16.f099.grib2 -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f099.grib2 ]] -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f099.grib2.idx ]] -5: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p16.f099.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f099.grib2 -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f099.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f099.grib2 = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f099.grib2 ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f099.grib2 -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f099.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f099.grib2.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f099.grib2.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f099.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f099.grib2 -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p16.f099.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f099.grib2.idx -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f099.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f099.grib2.idx = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f099.grib2.idx ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f099.grib2.idx -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f099.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f099.grib2.idx.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f099.grib2.idx.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f099.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f099.grib2.idx -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p16.f099.grib2 and gfs.wave.t12z.global.0p16.f099.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_gnh_10m to COM' -5: INFO: Copied gfs.wave.t12z.global.0p16.f099.grib2 and gfs.wave.t12z.global.0p16.f099.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_gnh_10m to COM -5: + wave_grib2_sbs.sh[130][[ gnh_10m == '' ]] -5: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -5: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p16.f099.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -5: INFO: gfs.wave.t12z.global.0p16.f099.grib2 is global.0p50 or SENDDBN is NO, no alert sent -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + cmdfile.7[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032715 3600. 9999 -6: + cmdfile.7[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grid_interp_sbs.sh[25]grdID=gsh_15m -6: + wave_grid_interp_sbs.sh[26]valid_time=2021032715 -6: + wave_grid_interp_sbs.sh[27]dt=3600. -6: + wave_grid_interp_sbs.sh[28]nst=9999 -6: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676 -6: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.uglo_100km ./out_grd.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mod_def.uglo_100km ./mod_def.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mod_def.gsh_15m ./mod_def.gsh_15m -6: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m'\''' -6: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m' -6: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[51]weights_found=1 -6: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 150000' -6: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 150000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gsh_15m/g ww3_gint.inp.tmpl -6: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -6: $ Input file for interpolation of uglo_100km to gsh_15m -6: $------------------------------------------------ -6: $ Start Time 3600. NSteps -6: 20210327 150000 3600. 9999 -6: $ Total number of grids -6: 2 -6: $ Grid extensions -6: 'uglo_100km' -6: 'gsh_15m' -6: $ -6: 0 -6: $ -6: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[70]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.321351 ']' -6: ++ prep_step[4]echo gfs_ww3_gint.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -6: INFO: Executing 'gfs_ww3_gint.x' -6: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[73]cat grid_interp.gsh_15m.out -6: -6: *** WAVEWATCH III Grid interpolation *** -6: =============================================== -6: -6: Comment character is '$' -6: -6: Time Information : -6: --------------------------------------------- -6: Starting Time : 2021/03/27 15:00:00 UTC -6: Interval (in sec) : 3600.00 -6: Number of requests : 9999 -6: --------------------------------------------- -6: Number of grids (including output grid) = 2 -6: -6: -6: Extension for grid 1 is --> uglo_100km -6: -6: Grid Particulars are : -6: Dimensions = 45166 1 -6: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Extension for grid 2 is --> gsh_15m -6: -6: Grid Particulars are : -6: Dimensions = 1440 277 -6: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -6: -6: -6: Preparing interpolation weights for output grid -6: Total number of wet points for interpolation 317192 -6: -6: -6: Variable: Grid Interpolation Map Units: 0.100E+01 -6: -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 277 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 241 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 205 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 169 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 133 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 97 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 61 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 | -6: 25 | 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 | -6: | | -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: -6: -6: Interpolating fields .... -6: -6: Output group 1 -6: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -6: Output group 2 -6: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -6: Output group 3 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 4 -6: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -6: Output group 5 -6: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -6: Output group 6 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 7 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 8 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 9 -6: Output variables skipped -6: Output group 10 -6: Output variables skipped -6: ------------------------------------------------ -6: 1Current vel. -6: 1Wind speed -6: 1Ice concentration -6: 2Wave height -6: 2Mean wave period(+2) -6: 2Mean wave period(+1) -6: 2Peak frequency -6: 2Mean wave dir. a1b1 -6: 2Peak direction -6: 4Part. wave height -6: 4Part. peak period -6: 4Part. mean direction -6: 5Charnock parameter -6: ------------------------------------------------ -6: OUTPUT TIME : 2021/03/27 15:00:00 UTC -6: -6: End of file reached -6: -6: -6: *** End of Grid interpolation Routine *** -6: =============================================== -6: -6: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -6: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -6: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_gsh_15m/out_grd.gsh_15m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.gsh_15m -6: + cmdfile.7[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032715 99 gsouth 0p25 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + cmdfile.7[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib2_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grib2_sbs.sh[30]grdID=gsh_15m -6: + wave_grib2_sbs.sh[31]GRIDNR=255 -6: + wave_grib2_sbs.sh[32]MODNR=11 -6: + wave_grib2_sbs.sh[33]valid_time=2021032715 -6: + wave_grib2_sbs.sh[34]fhr=99 -6: + wave_grib2_sbs.sh[35]grid_region=gsouth -6: + wave_grib2_sbs.sh[36]grid_res=0p25 -6: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676 -6: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_gsh_15m -6: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_gsh_15m -6: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_gsh_15m -6: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_gsh_15m -6: ++ wave_grib2_sbs.sh[47]printf %03i 99 -6: + wave_grib2_sbs.sh[47]FH3=099 -6: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -6: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.gsouth.0p25.f099.grib2 -6: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f099.grib2 ]] -6: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mod_def.gsh_15m ./mod_def.ww3 -6: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.gsh_15m ./out_grd.ww3 -6: + wave_grib2_sbs.sh[73]ngrib=1 -6: + wave_grib2_sbs.sh[74]dtgrib=3600 -6: + wave_grib2_sbs.sh[75]tstart='20210327 150000' -6: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 150000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gsh_15m.inp.tmpl -6: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -6: $ WAVEWATCH-III gridded output input file -6: $ ---------------------------------------- -6: 20210327 150000 3600 1 -6: N -6: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -6: $ -6: 20210327 150000 7 11 255 0 0 -6: $ -6: $ end of input file -6: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[88]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.321351 ']' -6: ++ prep_step[4]echo gfs_ww3_grib.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[90]export err=0 -6: + wave_grib2_sbs.sh[90]err=0 -6: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[95]cat grib2_gsouth_099.out -6: -6: *** WAVEWATCH III GRIB output postp. *** -6: ============================================== -6: -6: Comment character is '$' -6: -6: Grid name : GFSv16-wave S Hemisphere 1/4 d -6: -6: LINEIN: -6: 20210327 150000 3600 1 -6: -6: 20210327150000 3600 1 -6: GEN_PRO -99999 -6: -6: Output time data : -6: ----------------------------------------------------- -6: First time : 2021/03/27 15:00:00 UTC -6: Interval : 01:00:00 -6: Number of requests : 1 -6: Fields : Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: Requested output fields not yet available: -6: ----------------------------------------------------- -6: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -6: -6: Successfully requested output fields : -6: ----------------------------------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: -6: Additional GRIB parameters : -6: ----------------------------------------------------- -6: Run time : 2021/03/27 15:00:00 UTC -6: GRIB center ID : 7 -6: GRIB gen. proc. ID : 11 -6: GRIB grid ID : 255 -6: GRIB GDS parameter : 0 -6: Fields in file : -6: -------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: CHOSEN GRID TYPE: : LLRECTILINEAR -6: -6: -6: -6: Generating file -6: ----------------------------------------------------- -6: Data for 2021/03/27 15:00:00 UTC 0H forecast. -6: -6: End of program -6: ========================================= -6: WAVEWATCH III GRIB output -6: -6: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -6: + wave_grib2_sbs.sh[102][[ 99 -gt 0 ]] -6: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '99 hour fcst' -grib gfs.wave.t12z.gsouth.0p25.f099.grib2 -6: 1:0:d=2021032312:SPC:surface:99 hour fcst: -6: 2:118174:d=2021032312:DIRC:surface:99 hour fcst: -6: 3:506578:d=2021032312:UOGRD:surface:99 hour fcst: -6: 4:620301:d=2021032312:VOGRD:surface:99 hour fcst: -6: 5:741660:d=2021032312:WIND:surface:99 hour fcst: -6: 6:985519:d=2021032312:WDIR:surface:99 hour fcst: -6: 7:1366760:d=2021032312:UGRD:surface:99 hour fcst: -6: 8:1603631:d=2021032312:VGRD:surface:99 hour fcst: -6: 9:1842423:d=2021032312:ICEC:surface:99 hour fcst: -6: 10:1903515:d=2021032312:HTSGW:surface:99 hour fcst: -6: 11:2059261:d=2021032312:IMWF:surface:99 hour fcst: -6: 12:2230613:d=2021032312:MWSPER:surface:99 hour fcst: -6: 13:2402068:d=2021032312:PERPW:surface:99 hour fcst: -6: 14:2585357:d=2021032312:WWSDIR:surface:99 hour fcst: -6: 15:2916419:d=2021032312:DIRPW:surface:99 hour fcst: -6: 16:3253152:d=2021032312:WVHGT:surface:99 hour fcst: -6: 17:3409005:d=2021032312:SWELL:1 in sequence:99 hour fcst: -6: 18:3582655:d=2021032312:SWELL:2 in sequence:99 hour fcst: -6: 19:3714003:d=2021032312:SWELL:3 in sequence:99 hour fcst: -6: 20:3794655:d=2021032312:WVPER:surface:99 hour fcst: -6: 21:3969447:d=2021032312:SWPER:1 in sequence:99 hour fcst: -6: 22:4165179:d=2021032312:SWPER:2 in sequence:99 hour fcst: -6: 23:4333621:d=2021032312:SWPER:3 in sequence:99 hour fcst: -6: 24:4433712:d=2021032312:WVDIR:surface:99 hour fcst: -6: 25:4704867:d=2021032312:SWDIR:1 in sequence:99 hour fcst: -6: 26:5055300:d=2021032312:SWDIR:2 in sequence:99 hour fcst: -6: 27:5340369:d=2021032312:SWDIR:3 in sequence:99 hour fcst: -6: + wave_grib2_sbs.sh[104]err=0 -6: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.gsouth.0p25.f099.grib2 -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f099.grib2 ]] -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f099.grib2.idx ]] -6: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.gsouth.0p25.f099.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f099.grib2 -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f099.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f099.grib2 = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f099.grib2 ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f099.grib2 -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f099.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f099.grib2.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f099.grib2.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f099.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f099.grib2 -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.gsouth.0p25.f099.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f099.grib2.idx -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f099.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f099.grib2.idx = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f099.grib2.idx ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f099.grib2.idx -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f099.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f099.grib2.idx.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f099.grib2.idx.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f099.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f099.grib2.idx -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.gsouth.0p25.f099.grib2 and gfs.wave.t12z.gsouth.0p25.f099.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_gsh_15m to COM' -6: INFO: Copied gfs.wave.t12z.gsouth.0p25.f099.grib2 and gfs.wave.t12z.gsouth.0p25.f099.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_gsh_15m to COM -6: + wave_grib2_sbs.sh[130][[ gsh_15m == '' ]] -6: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -6: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.gsouth.0p25.f099.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -6: INFO: gfs.wave.t12z.gsouth.0p25.f099.grib2 is global.0p50 or SENDDBN is NO, no alert sent -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + cmdfile.1[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032715 3600. 9999 -0: + cmdfile.1[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grid_interp_sbs.sh[25]grdID=at_10m -0: + wave_grid_interp_sbs.sh[26]valid_time=2021032715 -0: + wave_grid_interp_sbs.sh[27]dt=3600. -0: + wave_grid_interp_sbs.sh[28]nst=9999 -0: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676 -0: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.uglo_100km ./out_grd.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mod_def.uglo_100km ./mod_def.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mod_def.at_10m ./mod_def.at_10m -0: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ]] -0: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m'\''' -0: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m' -0: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[51]weights_found=1 -0: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 150000' -0: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 150000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/at_10m/g ww3_gint.inp.tmpl -0: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -0: $ Input file for interpolation of uglo_100km to at_10m -0: $------------------------------------------------ -0: $ Start Time 3600. NSteps -0: 20210327 150000 3600. 9999 -0: $ Total number of grids -0: 2 -0: $ Grid extensions -0: 'uglo_100km' -0: 'at_10m' -0: $ -0: 0 -0: $ -0: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[70]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.321351 ']' -0: ++ prep_step[4]echo gfs_ww3_gint.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -0: INFO: Executing 'gfs_ww3_gint.x' -0: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[73]cat grid_interp.at_10m.out -0: -0: *** WAVEWATCH III Grid interpolation *** -0: =============================================== -0: -0: Comment character is '$' -0: -0: Time Information : -0: --------------------------------------------- -0: Starting Time : 2021/03/27 15:00:00 UTC -0: Interval (in sec) : 3600.00 -0: Number of requests : 9999 -0: --------------------------------------------- -0: Number of grids (including output grid) = 2 -0: -0: -0: Extension for grid 1 is --> uglo_100km -0: -0: Grid Particulars are : -0: Dimensions = 45166 1 -0: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Extension for grid 2 is --> at_10m -0: -0: Grid Particulars are : -0: Dimensions = 301 331 -0: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -0: -0: -0: Preparing interpolation weights for output grid -0: Total number of wet points for interpolation 29591 -0: -0: -0: Variable: Grid Interpolation Map Units: 0.100E+01 -0: -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 331 | | -0: | | -0: | | -0: 289 | | -0: | | -0: | 0 0 0 | -0: 247 | 0 0 0 0 | -0: | 0 0 0 0 0 | -0: | 0 0 0 0 | -0: 205 | 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 163 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 121 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 79 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 | -0: | 0 0 | -0: 37 | 0 | -0: | | -0: | | -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: -0: -0: Interpolating fields .... -0: -0: Output group 1 -0: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -0: Output group 2 -0: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -0: Output group 3 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 4 -0: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -0: Output group 5 -0: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -0: Output group 6 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 7 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 8 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 9 -0: Output variables skipped -0: Output group 10 -0: Output variables skipped -0: ------------------------------------------------ -0: 1Current vel. -0: 1Wind speed -0: 1Ice concentration -0: 2Wave height -0: 2Mean wave period(+2) -0: 2Mean wave period(+1) -0: 2Peak frequency -0: 2Mean wave dir. a1b1 -0: 2Peak direction -0: 4Part. wave height -0: 4Part. peak period -0: 4Part. mean direction -0: 5Charnock parameter -0: ------------------------------------------------ -0: OUTPUT TIME : 2021/03/27 15:00:00 UTC -0: -0: End of file reached -0: -0: -0: *** End of Grid interpolation Routine *** -0: =============================================== -0: -0: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -0: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -0: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grid_interp_at_10m/out_grd.at_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.at_10m -0: + cmdfile.1[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032715 99 atlocn 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + cmdfile.1[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib2_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grib2_sbs.sh[30]grdID=at_10m -0: + wave_grib2_sbs.sh[31]GRIDNR=255 -0: + wave_grib2_sbs.sh[32]MODNR=11 -0: + wave_grib2_sbs.sh[33]valid_time=2021032715 -0: + wave_grib2_sbs.sh[34]fhr=99 -0: + wave_grib2_sbs.sh[35]grid_region=atlocn -0: + wave_grib2_sbs.sh[36]grid_res=0p16 -0: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676 -0: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_at_10m -0: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_at_10m -0: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_at_10m -0: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_at_10m -0: ++ wave_grib2_sbs.sh[47]printf %03i 99 -0: + wave_grib2_sbs.sh[47]FH3=099 -0: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_atlocn_0p16 -0: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.atlocn.0p16.f099.grib2 -0: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f099.grib2 ]] -0: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/mod_def.at_10m ./mod_def.ww3 -0: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/out_grd.at_10m ./out_grd.ww3 -0: + wave_grib2_sbs.sh[73]ngrib=1 -0: + wave_grib2_sbs.sh[74]dtgrib=3600 -0: + wave_grib2_sbs.sh[75]tstart='20210327 150000' -0: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 150000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.at_10m.inp.tmpl -0: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -0: $ WAVEWATCH-III gridded output input file -0: $ ---------------------------------------- -0: 20210327 150000 3600 1 -0: N -0: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -0: $ -0: 20210327 150000 7 11 255 0 0 -0: $ -0: $ end of input file -0: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[88]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.321351 ']' -0: ++ prep_step[4]echo gfs_ww3_grib.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[90]export err=0 -0: + wave_grib2_sbs.sh[90]err=0 -0: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[95]cat grib2_atlocn_099.out -0: -0: *** WAVEWATCH III GRIB output postp. *** -0: ============================================== -0: -0: Comment character is '$' -0: -0: Grid name : NW Atlantic 10 min wave grid -0: -0: LINEIN: -0: 20210327 150000 3600 1 -0: -0: 20210327150000 3600 1 -0: GEN_PRO -99999 -0: -0: Output time data : -0: ----------------------------------------------------- -0: First time : 2021/03/27 15:00:00 UTC -0: Interval : 01:00:00 -0: Number of requests : 1 -0: Fields : Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: Requested output fields not yet available: -0: ----------------------------------------------------- -0: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -0: -0: Successfully requested output fields : -0: ----------------------------------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: -0: Additional GRIB parameters : -0: ----------------------------------------------------- -0: Run time : 2021/03/27 15:00:00 UTC -0: GRIB center ID : 7 -0: GRIB gen. proc. ID : 11 -0: GRIB grid ID : 255 -0: GRIB GDS parameter : 0 -0: Fields in file : -0: -------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: CHOSEN GRID TYPE: : LLRECTILINEAR -0: -0: -0: -0: Generating file -0: ----------------------------------------------------- -0: Data for 2021/03/27 15:00:00 UTC 0H forecast. -0: -0: End of program -0: ========================================= -0: WAVEWATCH III GRIB output -0: -0: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -0: + wave_grib2_sbs.sh[102][[ 99 -gt 0 ]] -0: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '99 hour fcst' -grib gfs.wave.t12z.atlocn.0p16.f099.grib2 -0: 1:0:d=2021032312:SPC:surface:99 hour fcst: -0: 2:20179:d=2021032312:DIRC:surface:99 hour fcst: -0: 3:61157:d=2021032312:UOGRD:surface:99 hour fcst: -0: 4:80595:d=2021032312:VOGRD:surface:99 hour fcst: -0: 5:100768:d=2021032312:WIND:surface:99 hour fcst: -0: 6:131256:d=2021032312:WDIR:surface:99 hour fcst: -0: 7:171492:d=2021032312:UGRD:surface:99 hour fcst: -0: 8:201654:d=2021032312:VGRD:surface:99 hour fcst: -0: 9:231473:d=2021032312:ICEC:surface:99 hour fcst: -0: 10:244108:d=2021032312:HTSGW:surface:99 hour fcst: -0: 11:269595:d=2021032312:IMWF:surface:99 hour fcst: -0: 12:296130:d=2021032312:MWSPER:surface:99 hour fcst: -0: 13:322744:d=2021032312:PERPW:surface:99 hour fcst: -0: 14:350462:d=2021032312:WWSDIR:surface:99 hour fcst: -0: 15:390324:d=2021032312:DIRPW:surface:99 hour fcst: -0: 16:430940:d=2021032312:WVHGT:surface:99 hour fcst: -0: 17:455628:d=2021032312:SWELL:1 in sequence:99 hour fcst: -0: 18:476818:d=2021032312:SWELL:2 in sequence:99 hour fcst: -0: 19:492598:d=2021032312:SWELL:3 in sequence:99 hour fcst: -0: 20:506154:d=2021032312:WVPER:surface:99 hour fcst: -0: 21:532750:d=2021032312:SWPER:1 in sequence:99 hour fcst: -0: 22:555513:d=2021032312:SWPER:2 in sequence:99 hour fcst: -0: 23:573088:d=2021032312:SWPER:3 in sequence:99 hour fcst: -0: 24:587452:d=2021032312:WVDIR:surface:99 hour fcst: -0: 25:624903:d=2021032312:SWDIR:1 in sequence:99 hour fcst: -0: 26:657249:d=2021032312:SWDIR:2 in sequence:99 hour fcst: -0: 27:679264:d=2021032312:SWDIR:3 in sequence:99 hour fcst: -0: + wave_grib2_sbs.sh[104]err=0 -0: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.atlocn.0p16.f099.grib2 -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f099.grib2 ]] -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f099.grib2.idx ]] -0: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.atlocn.0p16.f099.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f099.grib2 -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f099.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f099.grib2 = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f099.grib2 ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f099.grib2 -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f099.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f099.grib2.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f099.grib2.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f099.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f099.grib2 -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.atlocn.0p16.f099.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f099.grib2.idx -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f099.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f099.grib2.idx = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f099.grib2.idx ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f099.grib2.idx -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f099.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f099.grib2.idx.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f099.grib2.idx.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f099.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f099.grib2.idx -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.atlocn.0p16.f099.grib2 and gfs.wave.t12z.atlocn.0p16.f099.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_at_10m to COM' -0: INFO: Copied gfs.wave.t12z.atlocn.0p16.f099.grib2 and gfs.wave.t12z.atlocn.0p16.f099.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676/grib_at_10m to COM -0: + wave_grib2_sbs.sh[130][[ at_10m == '' ]] -0: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -0: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.atlocn.0p16.f099.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -0: INFO: gfs.wave.t12z.atlocn.0p16.f099.grib2 is global.0p50 or SENDDBN is NO, no alert sent -+ run_mpmd.sh[113]exit 0 -+ run_mpmd.sh[1]postamble run_mpmd.sh 1753758323 0 -+ preamble.sh[62]set +x -End run_mpmd.sh at 03:05:34 with error code 0 (time elapsed: 00:00:11) -+ exgfs_wave_post_gridded_sbs.sh[122]true -+ exgfs_wave_post_gridded_sbs.sh[123]export err=0 -+ exgfs_wave_post_gridded_sbs.sh[123]err=0 -+ exgfs_wave_post_gridded_sbs.sh[124][[ 0 -ne 0 ]] -+ exgfs_wave_post_gridded_sbs.sh[130]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ exgfs_wave_post_gridded_sbs.sh[131]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ exgfs_wave_post_gridded_sbs.sh[132]gribchk=gfs.wave.t12z.gsouth.0p25.f099.grib2 -+ exgfs_wave_post_gridded_sbs.sh[133][[ ! -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f099.grib2 ]] -+ exgfs_wave_post_gridded_sbs.sh[138]exit 0 -+ JGLOBAL_WAVE_POST_SBS[28]true -+ JGLOBAL_WAVE_POST_SBS[29]export err=0 -+ JGLOBAL_WAVE_POST_SBS[29]err=0 -+ JGLOBAL_WAVE_POST_SBS[30][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[37]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312 -+ JGLOBAL_WAVE_POST_SBS[38][[ NO != \Y\E\S ]] -+ JGLOBAL_WAVE_POST_SBS[39]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f099.314676 -+ JGLOBAL_WAVE_POST_SBS[42]exit 0 -+ JGLOBAL_WAVE_POST_SBS[1]postamble /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS 1753758312 0 -+ preamble.sh[62]set +x -End /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at 03:05:35 with error code 0 (time elapsed: 00:00:23) diff --git a/ci/error_logs/HECULES_PR_LOGS/C48_S2SW_logs_2021032312_gfs_wavepostsbs_f102-f108.log b/ci/error_logs/HECULES_PR_LOGS/C48_S2SW_logs_2021032312_gfs_wavepostsbs_f102-f108.log deleted file mode 100644 index 0f8353d1..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/C48_S2SW_logs_2021032312_gfs_wavepostsbs_f102-f108.log +++ /dev/null @@ -1,12860 +0,0 @@ -+ source /work2/noaa/global/mterry/global-workflow_forked/ush/load_ufswm_modules.sh -++ [[ NO == \N\O ]] -++ echo 'Loading modules quietly...' -Loading modules quietly... -++ set +x -Running "module reset". Resetting modules to system default. The following $MODULEPATH directories have been removed: None - -Currently Loaded Modules: - 1) contrib/0.1 25) git-lfs/3.1.2 - 2) intel-oneapi-compilers/2023.1.0 26) crtm/2.4.0.1 - 3) stack-intel/2021.9.0 27) g2/3.5.1 - 4) intel-oneapi-mpi/2021.9.0 28) g2tmpl/1.13.0 - 5) stack-intel-oneapi-mpi/2021.9.0 29) ip/4.3.0 - 6) nghttp2/1.57.0 30) sp/2.5.0 - 7) curl/8.4.0 31) w3emc/2.10.0 - 8) cmake/3.23.1 32) gftl/1.10.0 - 9) libjpeg/2.1.0 33) gftl-shared/1.6.1 - 10) jasper/2.0.32 34) fargparse/1.5.0 - 11) zlib/1.2.13 35) pigz/2.7 - 12) libpng/1.6.37 36) tar/1.34 - 13) hdf5/1.14.0 37) gettext/0.21.1 - 14) snappy/1.1.10 38) libxcrypt/4.4.35 - 15) zstd/1.5.2 39) sqlite/3.43.2 - 16) c-blosc/1.21.5 40) util-linux-uuid/2.38.1 - 17) netcdf-c/4.9.2 41) python/3.10.13 - 18) netcdf-fortran/4.6.1 42) mapl/2.40.3-esmf-8.6.0 - 19) parallel-netcdf/1.12.2 43) scotch/7.0.4 - 20) parallelio/2.5.10 44) ufs_common - 21) esmf/8.6.0 45) nccmp/1.9.0.1 - 22) fms/2024.01.02 46) ufs_hercules.intel - 23) bacio/2.4.1 47) prod_util/2.1.1 - 24) crtm-fix/2.4.0.1_emc 48) wgrib2/3.1.1 - - - -Begin /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at Tue Jul 29 03:04:27 UTC 2025 -++ jjob_header.sh[46]OPTIND=1 -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[50]env_job=wavepostsbs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[49]read -ra configs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[61]shift 4 -++ jjob_header.sh[63][[ -z wavepostsbs ]] -++ jjob_header.sh[71]export DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010 -++ jjob_header.sh[71]DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010 -++ jjob_header.sh[72][[ YES == \Y\E\S ]] -++ jjob_header.sh[73]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010 -++ jjob_header.sh[75]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010 -++ jjob_header.sh[76]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010 -++ jjob_header.sh[85]export pid=306705 -++ jjob_header.sh[85]pid=306705 -++ jjob_header.sh[86]export pgmout=OUTPUT.306705 -++ jjob_header.sh[86]pgmout=OUTPUT.306705 -++ jjob_header.sh[87]export pgmerr=errfile -++ jjob_header.sh[87]pgmerr=errfile -++ jjob_header.sh[90]export pgm= -++ jjob_header.sh[90]pgm= -++ jjob_header.sh[96]export cycle=t12z -++ jjob_header.sh[96]cycle=t12z -++ jjob_header.sh[97]setpdy.sh -+ setpdy.sh[20]'[' /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010 == /home/mterry ']' -+ setpdy.sh[25][[ ! t12z =~ t??z ]] -+ setpdy.sh[30]case $# in -+ setpdy.sh[31]dates_before_PDY=7 -+ setpdy.sh[32]dates_after_PDY=7 -+ setpdy.sh[50]COMDATEROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+ setpdy.sh[53]'[' -z 20210323 ']' -+ setpdy.sh[57]sed 's/[0-9]\{8\}/20210323/' /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z -sed: can't read /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z: No such file or directory -++ jjob_header.sh[97]true -++ jjob_header.sh[98]source ./PDY -/work2/noaa/global/mterry/global-workflow_forked/ush/jjob_header.sh: line 98: ./PDY: No such file or directory -++ jjob_header.sh[98]true -++ jjob_header.sh[104]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[104]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.base -+++ config.base[6]echo 'BEGIN: config.base' -BEGIN: config.base -+++ config.base[9]export machine=HERCULES -+++ config.base[9]machine=HERCULES -+++ config.base[12]export RUN_ENVIR=emc -+++ config.base[12]RUN_ENVIR=emc -+++ config.base[15]export ACCOUNT=fv3-cpu -+++ config.base[15]ACCOUNT=fv3-cpu -+++ config.base[16]export QUEUE=batch -+++ config.base[16]QUEUE=batch -+++ config.base[17]export QUEUE_SERVICE=batch -+++ config.base[17]QUEUE_SERVICE=batch -+++ config.base[18]export QUEUE_DTN=batch -+++ config.base[18]QUEUE_DTN=batch -+++ config.base[19]export PARTITION_BATCH=hercules -+++ config.base[19]PARTITION_BATCH=hercules -+++ config.base[20]export PARTITION_SERVICE=service -+++ config.base[20]PARTITION_SERVICE=service -+++ config.base[21]export PARTITION_DTN= -+++ config.base[21]PARTITION_DTN= -+++ config.base[22]export RESERVATION= -+++ config.base[22]RESERVATION= -+++ config.base[23]export CLUSTERS= -+++ config.base[23]CLUSTERS= -+++ config.base[24]export CLUSTERS_SERVICE= -+++ config.base[24]CLUSTERS_SERVICE= -+++ config.base[25]export CLUSTERS_DTN= -+++ config.base[25]CLUSTERS_DTN= -+++ config.base[28]export HPSS_PROJECT=emc-global -+++ config.base[28]HPSS_PROJECT=emc-global -+++ config.base[31]export HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[31]HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[32]export EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[32]EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[33]export FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[33]FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[34]export PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[34]PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[35]export SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[35]SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[36]export USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[36]USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[38]export FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[38]FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[39]export FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[39]FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[40]export FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[40]FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[41]export FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[41]FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[42]export FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[42]FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[43]export FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[43]FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[44]export FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[44]FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[45]export FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[45]FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[50]export PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[50]PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[51]export COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[51]COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[52]export COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[52]COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[53]export DMPDIR=/work/noaa/rstprod/dump -+++ config.base[53]DMPDIR=/work/noaa/rstprod/dump -+++ config.base[57]export COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[57]COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[58]export COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[58]COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[59]export COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[59]COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[62]export HOMEDIR=/work2/noaa/global/mterry -+++ config.base[62]HOMEDIR=/work2/noaa/global/mterry -+++ config.base[63]export STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[63]STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]export PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[65]export NOSCRUB=/work2/noaa/global/mterry -+++ config.base[65]NOSCRUB=/work2/noaa/global/mterry -+++ config.base[68]export BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[68]BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[71]export BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[71]BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[74]export DO_PREP_SFC=NO -+++ config.base[74]DO_PREP_SFC=NO -+++ config.base[77]export DO_GOES=NO -+++ config.base[77]DO_GOES=NO -+++ config.base[78]export DO_BUFRSND=NO -+++ config.base[78]DO_BUFRSND=NO -+++ config.base[79]export DO_GEMPAK=NO -+++ config.base[79]DO_GEMPAK=NO -+++ config.base[80]export DO_AWIPS=NO -+++ config.base[80]DO_AWIPS=NO -+++ config.base[81]export DO_NPOESS=NO -+++ config.base[81]DO_NPOESS=NO -+++ config.base[82]export DO_TRACKER=YES -+++ config.base[82]DO_TRACKER=YES -+++ config.base[83]export DO_GENESIS=YES -+++ config.base[83]DO_GENESIS=YES -+++ config.base[84]export DO_GENESIS_FSU=NO -+++ config.base[84]DO_GENESIS_FSU=NO -+++ config.base[85]export DO_VERFOZN=YES -+++ config.base[85]DO_VERFOZN=YES -+++ config.base[86]export DO_VERFRAD=YES -+++ config.base[86]DO_VERFRAD=YES -+++ config.base[87]export DO_VMINMON=YES -+++ config.base[87]DO_VMINMON=YES -+++ config.base[88]export DO_ANLSTAT=NO -+++ config.base[88]DO_ANLSTAT=NO -+++ config.base[91]export MODE=forecast-only -+++ config.base[91]MODE=forecast-only -+++ config.base[92]export DO_TEST_MODE=YES -+++ config.base[92]DO_TEST_MODE=YES -+++ config.base[101]export FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[101]FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[102]export HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[102]HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[103]export HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[103]HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[106]export NMV=/bin/mv -+++ config.base[106]NMV=/bin/mv -+++ config.base[107]export 'NLN=/bin/ln -sf' -+++ config.base[107]NLN='/bin/ln -sf' -+++ config.base[108]export VERBOSE=YES -+++ config.base[108]VERBOSE=YES -+++ config.base[109]export KEEPDATA=NO -+++ config.base[109]KEEPDATA=NO -+++ config.base[110]export DEBUG_POSTSCRIPT=NO -+++ config.base[110]DEBUG_POSTSCRIPT=NO -+++ config.base[111]export CHGRP_RSTPROD=YES -+++ config.base[111]CHGRP_RSTPROD=YES -+++ config.base[112]export 'CHGRP_CMD=chgrp rstprod' -+++ config.base[112]CHGRP_CMD='chgrp rstprod' -+++ config.base[113]export NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[113]NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[114]export NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[114]NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[117]export BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[117]BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[120]export SDATE=2021032312 -+++ config.base[120]SDATE=2021032312 -+++ config.base[121]export EDATE=2021032312 -+++ config.base[121]EDATE=2021032312 -+++ config.base[122]export EXP_WARM_START=.false. -+++ config.base[122]EXP_WARM_START=.false. -+++ config.base[123]export assim_freq=6 -+++ config.base[123]assim_freq=6 -+++ config.base[124]export PSLOT=C48_S2SW -+++ config.base[124]PSLOT=C48_S2SW -+++ config.base[125]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[125]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[126]export ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[126]ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[127]export DUMP_SUFFIX= -+++ config.base[127]DUMP_SUFFIX= -+++ config.base[128][[ 2021032312 -ge 2019092100 ]] -+++ config.base[128][[ 2021032312 -le 2019110700 ]] -+++ config.base[131]export ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[131]ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[132]export ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[132]ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[133]export FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[133]FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[136]export envir=prod -+++ config.base[136]envir=prod -+++ config.base[137]export NET=gfs -+++ config.base[137]NET=gfs -+++ config.base[138]export RUN=gfs -+++ config.base[138]RUN=gfs -+++ config.base[141]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.com -++++ config.com[4]echo 'BEGIN: config.com' -BEGIN: config.com -++++ config.com[38][[ emc == \n\c\o ]] -++++ config.com[43]COM_OBSPROC_TMPL='${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos' -++++ config.com[44]COM_RTOFS_TMPL='${DMPDIR}' -++++ config.com[45]COM_TCVITAL_TMPL='${DMPDIR}/${RUN}.${YMD}/${HH}/atmos' -++++ config.com[47]declare -rx 'COM_OBS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/obs' -++++ config.com[48]declare -rx COM_OBSPROC_TMPL COM_RTOFS_TMPL -++++ config.com[50]COM_BASE='${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}' -++++ config.com[52]declare -rx 'COM_TOP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}' -++++ config.com[54]declare -rx 'COM_CONF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf' -++++ config.com[55]declare -rx 'COM_OBS_JEDI=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi' -++++ config.com[57]declare -rx 'COM_ATMOS_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input' -++++ config.com[58]declare -rx 'COM_ATMOS_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart' -++++ config.com[59]declare -rx 'COM_ATMOS_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos' -++++ config.com[60]declare -rx 'COM_SNOW_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow' -++++ config.com[61]declare -rx 'COM_SNOW_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon' -++++ config.com[62]declare -rx 'COM_ATMOS_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history' -++++ config.com[63]declare -rx 'COM_ATMOS_MASTER_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master' -++++ config.com[64]declare -rx 'COM_ATMOS_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2' -++++ config.com[65]declare -rx 'COM_ATMOS_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}' -++++ config.com[66]declare -rx 'COM_ATMOS_BUFR_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr' -++++ config.com[67]declare -rx 'COM_ATMOS_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}' -++++ config.com[68]declare -rx 'COM_ATMOS_GENESIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital' -++++ config.com[69]declare -rx 'COM_ATMOS_TRACK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks' -++++ config.com[70]declare -rx 'COM_ATMOS_GOES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim' -++++ config.com[71]declare -rx 'COM_ATMOS_IMAGERY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery' -++++ config.com[72]declare -rx 'COM_ATMOS_OZNMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon' -++++ config.com[73]declare -rx 'COM_ATMOS_RADMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon' -++++ config.com[74]declare -rx 'COM_ATMOS_MINMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon' -++++ config.com[75]declare -rx 'COM_ATMOS_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon' -++++ config.com[76]declare -rx 'COM_ATMOS_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo' -++++ config.com[78]declare -rx 'COM_WAVE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart' -++++ config.com[79]declare -rx 'COM_WAVE_PREP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep' -++++ config.com[80]declare -rx 'COM_WAVE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history' -++++ config.com[81]declare -rx 'COM_WAVE_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded' -++++ config.com[82]declare -rx 'COM_WAVE_GRID_RES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}' -++++ config.com[83]declare -rx 'COM_WAVE_STATION_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station' -++++ config.com[84]declare -rx 'COM_WAVE_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak' -++++ config.com[85]declare -rx 'COM_WAVE_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo' -++++ config.com[87]declare -rx 'COM_OCEAN_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history' -++++ config.com[88]declare -rx 'COM_OCEAN_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart' -++++ config.com[89]declare -rx 'COM_OCEAN_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input' -++++ config.com[90]declare -rx 'COM_OCEAN_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean' -++++ config.com[91]declare -rx 'COM_OCEAN_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon' -++++ config.com[92]declare -rx 'COM_OCEAN_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf' -++++ config.com[93]declare -rx 'COM_OCEAN_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean' -++++ config.com[94]declare -rx 'COM_OCEAN_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf' -++++ config.com[95]declare -rx 'COM_OCEAN_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2' -++++ config.com[96]declare -rx 'COM_OCEAN_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}' -++++ config.com[98]declare -rx 'COM_ICE_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice' -++++ config.com[99]declare -rx 'COM_ICE_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf' -++++ config.com[100]declare -rx 'COM_ICE_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon' -++++ config.com[101]declare -rx 'COM_ICE_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice' -++++ config.com[102]declare -rx 'COM_ICE_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input' -++++ config.com[103]declare -rx 'COM_ICE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history' -++++ config.com[104]declare -rx 'COM_ICE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart' -++++ config.com[105]declare -rx 'COM_ICE_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf' -++++ config.com[106]declare -rx 'COM_ICE_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2' -++++ config.com[107]declare -rx 'COM_ICE_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}' -++++ config.com[109]declare -rx 'COM_CHEM_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history' -++++ config.com[110]declare -rx 'COM_CHEM_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem' -++++ config.com[111]declare -rx 'COM_CHEM_BMAT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix' -++++ config.com[112]declare -rx 'COM_CHEM_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon' -++++ config.com[114]declare -rx 'COM_MED_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart' -+++ config.base[143]export LOGSCRIPT= -+++ config.base[143]LOGSCRIPT= -+++ config.base[145]export 'REDOUT=1>' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wave -+++ config.wave[6]echo 'BEGIN: config.wave' -BEGIN: config.wave -+++ config.wave[13]export RUNRSTwave=gdas -+++ config.wave[13]RUNRSTwave=gdas -+++ config.wave[16]export MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[16]MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[19]case "${waveGRD}" in -+++ config.wave[64]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[64]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[65]export wavepostGRD= -+++ config.wave[65]wavepostGRD= -+++ config.wave[66]export waveuoutpGRD=uglo_100km -+++ config.wave[66]waveuoutpGRD=uglo_100km -+++ config.wave[75]export WAVEWND_DID= -+++ config.wave[75]WAVEWND_DID= -+++ config.wave[76]export WAVEWND_FID= -+++ config.wave[76]WAVEWND_FID= -+++ config.wave[79][[ gfs == \g\f\s ]] -+++ config.wave[80]export FHMAX_WAV=120 -+++ config.wave[80]FHMAX_WAV=120 -+++ config.wave[82]export WAVHINDH=0 -+++ config.wave[82]WAVHINDH=0 -+++ config.wave[83]export FHMAX_WAV_IBP=180 -+++ config.wave[83]FHMAX_WAV_IBP=180 -+++ config.wave[84](( FHMAX_WAV < FHMAX_WAV_IBP )) -+++ config.wave[84]export FHMAX_WAV_IBP=120 -+++ config.wave[84]FHMAX_WAV_IBP=120 -+++ config.wave[87]export DTFLD_WAV=3600 -+++ config.wave[87]DTFLD_WAV=3600 -+++ config.wave[88]export DTPNT_WAV=3600 -+++ config.wave[88]DTPNT_WAV=3600 -+++ config.wave[89]export FHINCP_WAV=1 -+++ config.wave[89]FHINCP_WAV=1 -+++ config.wave[92]export 'OUTPARS_WAV=WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[92]OUTPARS_WAV='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[95][[ gfs == \g\d\a\s ]] -+++ config.wave[99](( INTERVAL_GFS > 0 )) -+++ config.wave[100]export WAVHCYC=6 -+++ config.wave[100]WAVHCYC=6 -+++ config.wave[101]export FHMAX_WAV_CUR=192 -+++ config.wave[101]FHMAX_WAV_CUR=192 -+++ config.wave[109]export RSTTYPE_WAV=T -+++ config.wave[109]RSTTYPE_WAV=T -+++ config.wave[110][[ gfs != gfs ]] -+++ config.wave[115]rst_dt_gfs=43200 -+++ config.wave[116][[ 43200 -gt 0 ]] -+++ config.wave[117]export DT_1_RST_WAV=0 -+++ config.wave[117]DT_1_RST_WAV=0 -+++ config.wave[120]export DT_2_RST_WAV=43200 -+++ config.wave[120]DT_2_RST_WAV=43200 -+++ config.wave[126]export RSTIOFF_WAV=0 -+++ config.wave[126]RSTIOFF_WAV=0 -+++ config.wave[131]export RUNMEM=-1 -+++ config.wave[131]RUNMEM=-1 -+++ config.wave[134](( RUNMEM == -1 )) -+++ config.wave[136]export waveMEMB= -+++ config.wave[136]waveMEMB= -+++ config.wave[143]export WW3ATMINP=CPL -+++ config.wave[143]WW3ATMINP=CPL -+++ config.wave[144][[ YES == \Y\E\S ]] -+++ config.wave[145]export WW3ICEINP=CPL -+++ config.wave[145]WW3ICEINP=CPL -+++ config.wave[146]export WAVEICE_FID= -+++ config.wave[146]WAVEICE_FID= -+++ config.wave[152][[ YES == \Y\E\S ]] -+++ config.wave[153]export WW3CURINP=CPL -+++ config.wave[153]WW3CURINP=CPL -+++ config.wave[154]export WAVECUR_FID= -+++ config.wave[154]WAVECUR_FID= -+++ config.wave[161]export WW3ATMIENS=F -+++ config.wave[161]WW3ATMIENS=F -+++ config.wave[162]export WW3ICEIENS=F -+++ config.wave[162]WW3ICEIENS=F -+++ config.wave[163]export WW3CURIENS=F -+++ config.wave[163]WW3CURIENS=F -+++ config.wave[165]export GOFILETYPE=1 -+++ config.wave[165]GOFILETYPE=1 -+++ config.wave[166]export POFILETYPE=1 -+++ config.wave[166]POFILETYPE=1 -+++ config.wave[170]export FUNIPNT=T -+++ config.wave[170]FUNIPNT=T -+++ config.wave[172]export IOSRV=1 -+++ config.wave[172]IOSRV=1 -+++ config.wave[174]export FPNTPROC=T -+++ config.wave[174]FPNTPROC=T -+++ config.wave[176]export FGRDPROC=F -+++ config.wave[176]FGRDPROC=F -+++ config.wave[178]export FLAGMASKCOMP=F -+++ config.wave[178]FLAGMASKCOMP=F -+++ config.wave[180]export FLAGMASKOUT=F -+++ config.wave[180]FLAGMASKOUT=F -+++ config.wave[182]echo 'END: config.wave' -END: config.wave -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wavepostsbs -+++ config.wavepostsbs[6]echo 'BEGIN: config.wavepostsbs' -BEGIN: config.wavepostsbs -+++ config.wavepostsbs[9]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources wavepostsbs -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=wavepostsbs -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[156]ntasks=8 -++++ config.resources[157]threads_per_task=1 -++++ config.resources[158]tasks_per_node=80 -++++ config.resources[159]NTASKS=8 -++++ config.resources[160]memory=20GB -++++ config.resources[161]walltime=00:15:00 -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export NTASKS -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export memory -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.wavepostsbs[12]export MAX_TASKS=25 -+++ config.wavepostsbs[12]MAX_TASKS=25 -+++ config.wavepostsbs[15]export WAV_SUBGRBSRC= -+++ config.wavepostsbs[15]WAV_SUBGRBSRC= -+++ config.wavepostsbs[16]export WAV_SUBGRB= -+++ config.wavepostsbs[16]WAV_SUBGRB= -+++ config.wavepostsbs[19]export DOFLD_WAV=YES -+++ config.wavepostsbs[19]DOFLD_WAV=YES -+++ config.wavepostsbs[20]export DOPNT_WAV=YES -+++ config.wavepostsbs[20]DOPNT_WAV=YES -+++ config.wavepostsbs[21]export DOGRB_WAV=YES -+++ config.wavepostsbs[21]DOGRB_WAV=YES -+++ config.wavepostsbs[22][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+++ config.wavepostsbs[23]export DOGRI_WAV=YES -+++ config.wavepostsbs[23]DOGRI_WAV=YES -+++ config.wavepostsbs[27]export DOSPC_WAV=YES -+++ config.wavepostsbs[27]DOSPC_WAV=YES -+++ config.wavepostsbs[28]export DOBLL_WAV=YES -+++ config.wavepostsbs[28]DOBLL_WAV=YES -+++ config.wavepostsbs[30]echo 'END: config.wavepostsbs' -END: config.wavepostsbs -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env wavepostsbs -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=wavepostsbs -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 20971520 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 8 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[34]max_threads_per_task=1 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 1 ]] -+++ HERCULES.env[40][[ 1 -gt 1 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 8' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[63]export USE_CFP=YES -+++ HERCULES.env[63]USE_CFP=YES -+++ HERCULES.env[64][[ wavepostsbs == \w\a\v\e\p\r\e\p ]] -+++ HERCULES.env[65]export 'wavempexec=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[65]wavempexec='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[66]export 'wave_mpmd=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[66]wave_mpmd='--multi-prog --output=mpmd.%j.%t.out' -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[4]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ JGLOBAL_WAVE_POST_SBS[7]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[7]HH=12 -+ JGLOBAL_WAVE_POST_SBS[7]declare_from_tmpl -rx COMIN_WAVE_PREP:COM_WAVE_PREP_TMPL COMIN_WAVE_HISTORY:COM_WAVE_HISTORY_TMPL COMOUT_WAVE_GRID:COM_WAVE_GRID_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_WAVE_PREP=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep -declare_from_tmpl :: COMIN_WAVE_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history -declare_from_tmpl :: COMOUT_WAVE_GRID=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded -+ JGLOBAL_WAVE_POST_SBS[12][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n '' ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_atlocn_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_atlocn_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_atlocn_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_epacif_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_epacif_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_epacif_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_wcoast_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_wcoast_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_wcoast_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p50 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p50:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p50=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p50 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=arctic.9km -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_arctic_9km:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_arctic_9km=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_arctic_9km -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_gsouth_0p25:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_gsouth_0p25=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 ]] -+ JGLOBAL_WAVE_POST_SBS[28]/work2/noaa/global/mterry/global-workflow_forked/scripts/exgfs_wave_post_gridded_sbs.sh -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ exgfs_wave_post_gridded_sbs.sh[24]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ exgfs_wave_post_gridded_sbs.sh[26]DOGRI_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[27]DOGRB_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[29]export waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[29]waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[30]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[30]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[31]export wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[31]wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[33]cat - INFO: Grid information: - INFO: Native wave grids: uglo_100km - INFO: Interpolated grids: at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m - INFO: Post-process grids: -++ exgfs_wave_post_gridded_sbs.sh[40]printf %03i 102 -+ exgfs_wave_post_gridded_sbs.sh[40]fhr3=102 -++ exgfs_wave_post_gridded_sbs.sh[41]date -u -d '20210323 12 + 102 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[41]valid_time=2021032718 -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[49]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f102.bin ./out_grd.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f102.bin ./out_grd.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[52][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[62]cat - INFO: Summary: - INFO: Grid interp: DOGRI_WAV="YES" - INFO: Grib files: DOGRB_WAV="YES" - INFO: Fields to be included in grib files: - INFO: OUTPARS_WAV="WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA" -+ exgfs_wave_post_gridded_sbs.sh[70][[ YES == \N\O ]] -+ exgfs_wave_post_gridded_sbs.sh[76]rm -f 'cmdfile.*' cmdfile -+ exgfs_wave_post_gridded_sbs.sh[77]count=0 -+ exgfs_wave_post_gridded_sbs.sh[80][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[81]dt_int=3600. -+ exgfs_wave_post_gridded_sbs.sh[82]n_int=9999 -++ exgfs_wave_post_gridded_sbs.sh[83]date -u -d '20210327 18 - 0 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[83]ymdh_int=2021032718 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=1 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032718 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032718 102 atlocn 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib2_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib2_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=2 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032718 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032718 102 epacif 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib2_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib2_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=3 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032718 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032718 102 wcoast 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib2_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib2_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=4 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032718 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032718 102 global 0p50 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib2_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib2_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=5 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032718 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032718 102 arctic 9km '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib2_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib2_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=6 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032718 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032718 102 global 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib2_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib2_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=7 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032718 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032718 102 gsouth 0p25 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib2_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib2_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[100][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[113][[ 8 -lt 7 ]] -+ exgfs_wave_post_gridded_sbs.sh[121]echo 'INFO: Running MPMD job with 7 commands' -INFO: Running MPMD job with 7 commands -+ exgfs_wave_post_gridded_sbs.sh[122]/work2/noaa/global/mterry/global-workflow_forked/ush/run_mpmd.sh /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/cmdfile -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ run_mpmd.sh[31]source /work2/noaa/global/mterry/global-workflow_forked/ush/preamble.sh -++ preamble.sh[20]set +x -Begin run_mpmd.sh at Tue Jul 29 03:04:39 UTC 2025 -+ run_mpmd.sh[33]cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/cmdfile -+ run_mpmd.sh[36][[ YES != \Y\E\S ]] -+ run_mpmd.sh[46]export OMP_NUM_THREADS=1 -+ run_mpmd.sh[46]OMP_NUM_THREADS=1 -++ run_mpmd.sh[49]wc -l -+ run_mpmd.sh[49]nprocs=7 -+ run_mpmd.sh[52]mpmd_cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mpmd_cmdfile -+ run_mpmd.sh[53][[ -s /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mpmd_cmdfile ]] -+ run_mpmd.sh[55]cat - INFO: Executing MPMD job, STDOUT redirected for each process separately - INFO: On failure, logs for each job will be available in /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mpmd.proc_num.out - INFO: The proc_num corresponds to the line in '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mpmd_cmdfile' -+ run_mpmd.sh[61][[ srun -l --export=ALL --hint=nomultithread =~ ^srun.* ]] -+ run_mpmd.sh[65]nm=0 -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '0 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/cmdfile.1' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '1 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/cmdfile.2' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '2 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/cmdfile.3' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '3 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/cmdfile.4' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '4 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/cmdfile.5' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '5 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/cmdfile.6' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '6 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/cmdfile.7' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[72]set +e -+ run_mpmd.sh[74]srun -l --export=ALL --hint=nomultithread --multi-prog --output=mpmd.%j.%t.out -n 7 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mpmd_cmdfile -+ run_mpmd.sh[75]err=0 -+ run_mpmd.sh[76]set_strict -+ preamble.sh[35][[ YES == \Y\E\S ]] -+ preamble.sh[37]set -eu -+ run_mpmd.sh[103][[ 0 -eq 0 ]] -+ run_mpmd.sh[104]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mpmd_cmdfile -++ run_mpmd.sh[105]find . -name 'mpmd.*.out' -+ run_mpmd.sh[105]out_files='./mpmd.5951799.1.out -./mpmd.5951799.6.out -./mpmd.5951799.0.out -./mpmd.5951799.5.out -./mpmd.5951799.3.out -./mpmd.5951799.4.out -./mpmd.5951799.2.out' -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951799.1.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951799.1.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951799.6.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951799.6.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951799.0.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951799.0.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951799.5.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951799.5.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951799.3.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951799.3.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951799.4.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951799.4.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951799.2.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951799.2.out -+ run_mpmd.sh[110]cat mpmd.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + cmdfile.2[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032718 3600. 9999 -1: + cmdfile.2[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grid_interp_sbs.sh[25]grdID=ep_10m -1: + wave_grid_interp_sbs.sh[26]valid_time=2021032718 -1: + wave_grid_interp_sbs.sh[27]dt=3600. -1: + wave_grid_interp_sbs.sh[28]nst=9999 -1: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010 -1: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.uglo_100km ./out_grd.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mod_def.uglo_100km ./mod_def.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mod_def.ep_10m ./mod_def.ep_10m -1: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ]] -1: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m'\''' -1: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m' -1: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[51]weights_found=1 -1: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 180000' -1: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 180000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/ep_10m/g ww3_gint.inp.tmpl -1: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -1: $ Input file for interpolation of uglo_100km to ep_10m -1: $------------------------------------------------ -1: $ Start Time 3600. NSteps -1: 20210327 180000 3600. 9999 -1: $ Total number of grids -1: 2 -1: $ Grid extensions -1: 'uglo_100km' -1: 'ep_10m' -1: $ -1: 0 -1: $ -1: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[70]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.306705 ']' -1: ++ prep_step[4]echo gfs_ww3_gint.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -1: INFO: Executing 'gfs_ww3_gint.x' -1: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[73]cat grid_interp.ep_10m.out -1: -1: *** WAVEWATCH III Grid interpolation *** -1: =============================================== -1: -1: Comment character is '$' -1: -1: Time Information : -1: --------------------------------------------- -1: Starting Time : 2021/03/27 18:00:00 UTC -1: Interval (in sec) : 3600.00 -1: Number of requests : 9999 -1: --------------------------------------------- -1: Number of grids (including output grid) = 2 -1: -1: -1: Extension for grid 1 is --> uglo_100km -1: -1: Grid Particulars are : -1: Dimensions = 45166 1 -1: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Extension for grid 2 is --> ep_10m -1: -1: Grid Particulars are : -1: Dimensions = 511 301 -1: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -1: -1: -1: Preparing interpolation weights for output grid -1: Total number of wet points for interpolation 7439 -1: -1: -1: Variable: Grid Interpolation Map Units: 0.100E+01 -1: -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 301 | | -1: | | -1: | 0 0 | -1: 262 | 0 0 0 0 | -1: | 0 0 0 0 | -1: | 0 0 0 0 | -1: 223 | 0 0 0 | -1: | 0 | -1: | 0 | -1: 184 | | -1: | 0 0 0 | -1: | 0 0 0 0 | -1: 145 | 0 | -1: | | -1: | | -1: 106 | | -1: | | -1: | | -1: 67 | | -1: | 0 | -1: | | -1: 28 | 0 | -1: | | -1: | | -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: -1: -1: Interpolating fields .... -1: -1: Output group 1 -1: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -1: Output group 2 -1: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -1: Output group 3 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 4 -1: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -1: Output group 5 -1: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -1: Output group 6 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 7 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 8 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 9 -1: Output variables skipped -1: Output group 10 -1: Output variables skipped -1: ------------------------------------------------ -1: 1Current vel. -1: 1Wind speed -1: 1Ice concentration -1: 2Wave height -1: 2Mean wave period(+2) -1: 2Mean wave period(+1) -1: 2Peak frequency -1: 2Mean wave dir. a1b1 -1: 2Peak direction -1: 4Part. wave height -1: 4Part. peak period -1: 4Part. mean direction -1: 5Charnock parameter -1: ------------------------------------------------ -1: OUTPUT TIME : 2021/03/27 18:00:00 UTC -1: -1: End of file reached -1: -1: -1: *** End of Grid interpolation Routine *** -1: =============================================== -1: -1: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -1: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -1: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_ep_10m/out_grd.ep_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.ep_10m -1: + cmdfile.2[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032718 102 epacif 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + cmdfile.2[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib2_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grib2_sbs.sh[30]grdID=ep_10m -1: + wave_grib2_sbs.sh[31]GRIDNR=255 -1: + wave_grib2_sbs.sh[32]MODNR=11 -1: + wave_grib2_sbs.sh[33]valid_time=2021032718 -1: + wave_grib2_sbs.sh[34]fhr=102 -1: + wave_grib2_sbs.sh[35]grid_region=epacif -1: + wave_grib2_sbs.sh[36]grid_res=0p16 -1: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010 -1: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_ep_10m -1: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_ep_10m -1: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_ep_10m -1: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_ep_10m -1: ++ wave_grib2_sbs.sh[47]printf %03i 102 -1: + wave_grib2_sbs.sh[47]FH3=102 -1: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_epacif_0p16 -1: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.epacif.0p16.f102.grib2 -1: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f102.grib2 ]] -1: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mod_def.ep_10m ./mod_def.ww3 -1: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.ep_10m ./out_grd.ww3 -1: + wave_grib2_sbs.sh[73]ngrib=1 -1: + wave_grib2_sbs.sh[74]dtgrib=3600 -1: + wave_grib2_sbs.sh[75]tstart='20210327 180000' -1: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 180000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.ep_10m.inp.tmpl -1: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -1: $ WAVEWATCH-III gridded output input file -1: $ ---------------------------------------- -1: 20210327 180000 3600 1 -1: N -1: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -1: $ -1: 20210327 180000 7 11 255 0 0 -1: $ -1: $ end of input file -1: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[88]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.306705 ']' -1: ++ prep_step[4]echo gfs_ww3_grib.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[90]export err=0 -1: + wave_grib2_sbs.sh[90]err=0 -1: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[95]cat grib2_epacif_102.out -1: -1: *** WAVEWATCH III GRIB output postp. *** -1: ============================================== -1: -1: Comment character is '$' -1: -1: Grid name : East Pacific 10 min wave grid -1: -1: LINEIN: -1: 20210327 180000 3600 1 -1: -1: 20210327180000 3600 1 -1: GEN_PRO -99999 -1: -1: Output time data : -1: ----------------------------------------------------- -1: First time : 2021/03/27 18:00:00 UTC -1: Interval : 01:00:00 -1: Number of requests : 1 -1: Fields : Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: Requested output fields not yet available: -1: ----------------------------------------------------- -1: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -1: -1: Successfully requested output fields : -1: ----------------------------------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: -1: Additional GRIB parameters : -1: ----------------------------------------------------- -1: Run time : 2021/03/27 18:00:00 UTC -1: GRIB center ID : 7 -1: GRIB gen. proc. ID : 11 -1: GRIB grid ID : 255 -1: GRIB GDS parameter : 0 -1: Fields in file : -1: -------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: CHOSEN GRID TYPE: : LLRECTILINEAR -1: -1: -1: -1: Generating file -1: ----------------------------------------------------- -1: Data for 2021/03/27 18:00:00 UTC 0H forecast. -1: Warning: bitmask off everywhere. -1: Pretend one point in jpcpack to avoid crash. -1: Warning: bitmask off everywhere. -1: Pretend one point in jpcpack to avoid crash. -1: Warning: bitmask off everywhere. -1: Pretend one point in jpcpack to avoid crash. -1: -1: End of program -1: ========================================= -1: WAVEWATCH III GRIB output -1: -1: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -1: + wave_grib2_sbs.sh[102][[ 102 -gt 0 ]] -1: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '102 hour fcst' -grib gfs.wave.t12z.epacif.0p16.f102.grib2 -1: 1:0:d=2021032312:SPC:surface:102 hour fcst: -1: 2:21982:d=2021032312:DIRC:surface:102 hour fcst: -1: 3:50437:d=2021032312:UOGRD:surface:102 hour fcst: -1: 4:72271:d=2021032312:VOGRD:surface:102 hour fcst: -1: 5:94321:d=2021032312:WIND:surface:102 hour fcst: -1: 6:118899:d=2021032312:WDIR:surface:102 hour fcst: -1: 7:146056:d=2021032312:UGRD:surface:102 hour fcst: -1: 8:170627:d=2021032312:VGRD:surface:102 hour fcst: -1: 9:195093:d=2021032312:ICEC:surface:102 hour fcst: -1: 10:214501:d=2021032312:HTSGW:surface:102 hour fcst: -1: 11:236871:d=2021032312:IMWF:surface:102 hour fcst: -1: 12:259845:d=2021032312:MWSPER:surface:102 hour fcst: -1: 13:282836:d=2021032312:PERPW:surface:102 hour fcst: -1: 14:305764:d=2021032312:WWSDIR:surface:102 hour fcst: -1: 15:331755:d=2021032312:DIRPW:surface:102 hour fcst: -1: 16:357905:d=2021032312:WVHGT:surface:102 hour fcst: -1: 17:380038:d=2021032312:SWELL:1 in sequence:102 hour fcst: -1: 18:402945:d=2021032312:SWELL:2 in sequence:102 hour fcst: -1: 19:423422:d=2021032312:SWELL:3 in sequence:102 hour fcst: -1: 20:442830:d=2021032312:WVPER:surface:102 hour fcst: -1: 21:465046:d=2021032312:SWPER:1 in sequence:102 hour fcst: -1: 22:488106:d=2021032312:SWPER:2 in sequence:102 hour fcst: -1: 23:508564:d=2021032312:SWPER:3 in sequence:102 hour fcst: -1: 24:527972:d=2021032312:WVDIR:surface:102 hour fcst: -1: 25:552329:d=2021032312:SWDIR:1 in sequence:102 hour fcst: -1: 26:579544:d=2021032312:SWDIR:2 in sequence:102 hour fcst: -1: 27:601044:d=2021032312:SWDIR:3 in sequence:102 hour fcst: -1: + wave_grib2_sbs.sh[104]err=0 -1: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.epacif.0p16.f102.grib2 -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f102.grib2 ]] -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f102.grib2.idx ]] -1: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.epacif.0p16.f102.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f102.grib2 -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f102.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f102.grib2 = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f102.grib2 ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f102.grib2 -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f102.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f102.grib2.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f102.grib2.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f102.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f102.grib2 -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.epacif.0p16.f102.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f102.grib2.idx -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f102.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f102.grib2.idx = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f102.grib2.idx ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f102.grib2.idx -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f102.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f102.grib2.idx.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f102.grib2.idx.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f102.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f102.grib2.idx -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.epacif.0p16.f102.grib2 and gfs.wave.t12z.epacif.0p16.f102.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_ep_10m to COM' -1: INFO: Copied gfs.wave.t12z.epacif.0p16.f102.grib2 and gfs.wave.t12z.epacif.0p16.f102.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_ep_10m to COM -1: + wave_grib2_sbs.sh[130][[ ep_10m == '' ]] -1: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -1: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.epacif.0p16.f102.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -1: INFO: gfs.wave.t12z.epacif.0p16.f102.grib2 is global.0p50 or SENDDBN is NO, no alert sent -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + cmdfile.7[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032718 3600. 9999 -6: + cmdfile.7[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grid_interp_sbs.sh[25]grdID=gsh_15m -6: + wave_grid_interp_sbs.sh[26]valid_time=2021032718 -6: + wave_grid_interp_sbs.sh[27]dt=3600. -6: + wave_grid_interp_sbs.sh[28]nst=9999 -6: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010 -6: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.uglo_100km ./out_grd.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mod_def.uglo_100km ./mod_def.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mod_def.gsh_15m ./mod_def.gsh_15m -6: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m'\''' -6: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m' -6: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[51]weights_found=1 -6: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 180000' -6: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 180000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gsh_15m/g ww3_gint.inp.tmpl -6: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -6: $ Input file for interpolation of uglo_100km to gsh_15m -6: $------------------------------------------------ -6: $ Start Time 3600. NSteps -6: 20210327 180000 3600. 9999 -6: $ Total number of grids -6: 2 -6: $ Grid extensions -6: 'uglo_100km' -6: 'gsh_15m' -6: $ -6: 0 -6: $ -6: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[70]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.306705 ']' -6: ++ prep_step[4]echo gfs_ww3_gint.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -6: INFO: Executing 'gfs_ww3_gint.x' -6: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[73]cat grid_interp.gsh_15m.out -6: -6: *** WAVEWATCH III Grid interpolation *** -6: =============================================== -6: -6: Comment character is '$' -6: -6: Time Information : -6: --------------------------------------------- -6: Starting Time : 2021/03/27 18:00:00 UTC -6: Interval (in sec) : 3600.00 -6: Number of requests : 9999 -6: --------------------------------------------- -6: Number of grids (including output grid) = 2 -6: -6: -6: Extension for grid 1 is --> uglo_100km -6: -6: Grid Particulars are : -6: Dimensions = 45166 1 -6: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Extension for grid 2 is --> gsh_15m -6: -6: Grid Particulars are : -6: Dimensions = 1440 277 -6: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -6: -6: -6: Preparing interpolation weights for output grid -6: Total number of wet points for interpolation 317192 -6: -6: -6: Variable: Grid Interpolation Map Units: 0.100E+01 -6: -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 277 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 241 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 205 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 169 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 133 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 97 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 61 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 | -6: 25 | 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 | -6: | | -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: -6: -6: Interpolating fields .... -6: -6: Output group 1 -6: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -6: Output group 2 -6: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -6: Output group 3 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 4 -6: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -6: Output group 5 -6: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -6: Output group 6 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 7 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 8 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 9 -6: Output variables skipped -6: Output group 10 -6: Output variables skipped -6: ------------------------------------------------ -6: 1Current vel. -6: 1Wind speed -6: 1Ice concentration -6: 2Wave height -6: 2Mean wave period(+2) -6: 2Mean wave period(+1) -6: 2Peak frequency -6: 2Mean wave dir. a1b1 -6: 2Peak direction -6: 4Part. wave height -6: 4Part. peak period -6: 4Part. mean direction -6: 5Charnock parameter -6: ------------------------------------------------ -6: OUTPUT TIME : 2021/03/27 18:00:00 UTC -6: -6: End of file reached -6: -6: -6: *** End of Grid interpolation Routine *** -6: =============================================== -6: -6: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -6: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -6: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_gsh_15m/out_grd.gsh_15m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.gsh_15m -6: + cmdfile.7[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032718 102 gsouth 0p25 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + cmdfile.7[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib2_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grib2_sbs.sh[30]grdID=gsh_15m -6: + wave_grib2_sbs.sh[31]GRIDNR=255 -6: + wave_grib2_sbs.sh[32]MODNR=11 -6: + wave_grib2_sbs.sh[33]valid_time=2021032718 -6: + wave_grib2_sbs.sh[34]fhr=102 -6: + wave_grib2_sbs.sh[35]grid_region=gsouth -6: + wave_grib2_sbs.sh[36]grid_res=0p25 -6: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010 -6: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_gsh_15m -6: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_gsh_15m -6: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_gsh_15m -6: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_gsh_15m -6: ++ wave_grib2_sbs.sh[47]printf %03i 102 -6: + wave_grib2_sbs.sh[47]FH3=102 -6: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -6: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.gsouth.0p25.f102.grib2 -6: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f102.grib2 ]] -6: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mod_def.gsh_15m ./mod_def.ww3 -6: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.gsh_15m ./out_grd.ww3 -6: + wave_grib2_sbs.sh[73]ngrib=1 -6: + wave_grib2_sbs.sh[74]dtgrib=3600 -6: + wave_grib2_sbs.sh[75]tstart='20210327 180000' -6: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 180000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gsh_15m.inp.tmpl -6: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -6: $ WAVEWATCH-III gridded output input file -6: $ ---------------------------------------- -6: 20210327 180000 3600 1 -6: N -6: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -6: $ -6: 20210327 180000 7 11 255 0 0 -6: $ -6: $ end of input file -6: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[88]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.306705 ']' -6: ++ prep_step[4]echo gfs_ww3_grib.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[90]export err=0 -6: + wave_grib2_sbs.sh[90]err=0 -6: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[95]cat grib2_gsouth_102.out -6: -6: *** WAVEWATCH III GRIB output postp. *** -6: ============================================== -6: -6: Comment character is '$' -6: -6: Grid name : GFSv16-wave S Hemisphere 1/4 d -6: -6: LINEIN: -6: 20210327 180000 3600 1 -6: -6: 20210327180000 3600 1 -6: GEN_PRO -99999 -6: -6: Output time data : -6: ----------------------------------------------------- -6: First time : 2021/03/27 18:00:00 UTC -6: Interval : 01:00:00 -6: Number of requests : 1 -6: Fields : Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: Requested output fields not yet available: -6: ----------------------------------------------------- -6: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -6: -6: Successfully requested output fields : -6: ----------------------------------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: -6: Additional GRIB parameters : -6: ----------------------------------------------------- -6: Run time : 2021/03/27 18:00:00 UTC -6: GRIB center ID : 7 -6: GRIB gen. proc. ID : 11 -6: GRIB grid ID : 255 -6: GRIB GDS parameter : 0 -6: Fields in file : -6: -------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: CHOSEN GRID TYPE: : LLRECTILINEAR -6: -6: -6: -6: Generating file -6: ----------------------------------------------------- -6: Data for 2021/03/27 18:00:00 UTC 0H forecast. -6: -6: End of program -6: ========================================= -6: WAVEWATCH III GRIB output -6: -6: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -6: + wave_grib2_sbs.sh[102][[ 102 -gt 0 ]] -6: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '102 hour fcst' -grib gfs.wave.t12z.gsouth.0p25.f102.grib2 -6: 1:0:d=2021032312:SPC:surface:102 hour fcst: -6: 2:118625:d=2021032312:DIRC:surface:102 hour fcst: -6: 3:506620:d=2021032312:UOGRD:surface:102 hour fcst: -6: 4:620271:d=2021032312:VOGRD:surface:102 hour fcst: -6: 5:741914:d=2021032312:WIND:surface:102 hour fcst: -6: 6:985584:d=2021032312:WDIR:surface:102 hour fcst: -6: 7:1366505:d=2021032312:UGRD:surface:102 hour fcst: -6: 8:1603884:d=2021032312:VGRD:surface:102 hour fcst: -6: 9:1842386:d=2021032312:ICEC:surface:102 hour fcst: -6: 10:1903469:d=2021032312:HTSGW:surface:102 hour fcst: -6: 11:2058534:d=2021032312:IMWF:surface:102 hour fcst: -6: 12:2230443:d=2021032312:MWSPER:surface:102 hour fcst: -6: 13:2402591:d=2021032312:PERPW:surface:102 hour fcst: -6: 14:2586653:d=2021032312:WWSDIR:surface:102 hour fcst: -6: 15:2918105:d=2021032312:DIRPW:surface:102 hour fcst: -6: 16:3254952:d=2021032312:WVHGT:surface:102 hour fcst: -6: 17:3409958:d=2021032312:SWELL:1 in sequence:102 hour fcst: -6: 18:3583743:d=2021032312:SWELL:2 in sequence:102 hour fcst: -6: 19:3717046:d=2021032312:SWELL:3 in sequence:102 hour fcst: -6: 20:3799082:d=2021032312:WVPER:surface:102 hour fcst: -6: 21:3973488:d=2021032312:SWPER:1 in sequence:102 hour fcst: -6: 22:4169346:d=2021032312:SWPER:2 in sequence:102 hour fcst: -6: 23:4341227:d=2021032312:SWPER:3 in sequence:102 hour fcst: -6: 24:4442568:d=2021032312:WVDIR:surface:102 hour fcst: -6: 25:4712305:d=2021032312:SWDIR:1 in sequence:102 hour fcst: -6: 26:5063601:d=2021032312:SWDIR:2 in sequence:102 hour fcst: -6: 27:5354849:d=2021032312:SWDIR:3 in sequence:102 hour fcst: -6: + wave_grib2_sbs.sh[104]err=0 -6: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.gsouth.0p25.f102.grib2 -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f102.grib2 ]] -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f102.grib2.idx ]] -6: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.gsouth.0p25.f102.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f102.grib2 -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f102.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f102.grib2 = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f102.grib2 ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f102.grib2 -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f102.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f102.grib2.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f102.grib2.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f102.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f102.grib2 -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.gsouth.0p25.f102.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f102.grib2.idx -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f102.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f102.grib2.idx = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f102.grib2.idx ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f102.grib2.idx -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f102.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f102.grib2.idx.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f102.grib2.idx.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f102.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f102.grib2.idx -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.gsouth.0p25.f102.grib2 and gfs.wave.t12z.gsouth.0p25.f102.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_gsh_15m to COM' -6: INFO: Copied gfs.wave.t12z.gsouth.0p25.f102.grib2 and gfs.wave.t12z.gsouth.0p25.f102.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_gsh_15m to COM -6: + wave_grib2_sbs.sh[130][[ gsh_15m == '' ]] -6: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -6: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.gsouth.0p25.f102.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -6: INFO: gfs.wave.t12z.gsouth.0p25.f102.grib2 is global.0p50 or SENDDBN is NO, no alert sent -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + cmdfile.1[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032718 3600. 9999 -0: + cmdfile.1[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grid_interp_sbs.sh[25]grdID=at_10m -0: + wave_grid_interp_sbs.sh[26]valid_time=2021032718 -0: + wave_grid_interp_sbs.sh[27]dt=3600. -0: + wave_grid_interp_sbs.sh[28]nst=9999 -0: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010 -0: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.uglo_100km ./out_grd.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mod_def.uglo_100km ./mod_def.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mod_def.at_10m ./mod_def.at_10m -0: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ]] -0: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m'\''' -0: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m' -0: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[51]weights_found=1 -0: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 180000' -0: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 180000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/at_10m/g ww3_gint.inp.tmpl -0: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -0: $ Input file for interpolation of uglo_100km to at_10m -0: $------------------------------------------------ -0: $ Start Time 3600. NSteps -0: 20210327 180000 3600. 9999 -0: $ Total number of grids -0: 2 -0: $ Grid extensions -0: 'uglo_100km' -0: 'at_10m' -0: $ -0: 0 -0: $ -0: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[70]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.306705 ']' -0: ++ prep_step[4]echo gfs_ww3_gint.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -0: INFO: Executing 'gfs_ww3_gint.x' -0: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[73]cat grid_interp.at_10m.out -0: -0: *** WAVEWATCH III Grid interpolation *** -0: =============================================== -0: -0: Comment character is '$' -0: -0: Time Information : -0: --------------------------------------------- -0: Starting Time : 2021/03/27 18:00:00 UTC -0: Interval (in sec) : 3600.00 -0: Number of requests : 9999 -0: --------------------------------------------- -0: Number of grids (including output grid) = 2 -0: -0: -0: Extension for grid 1 is --> uglo_100km -0: -0: Grid Particulars are : -0: Dimensions = 45166 1 -0: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Extension for grid 2 is --> at_10m -0: -0: Grid Particulars are : -0: Dimensions = 301 331 -0: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -0: -0: -0: Preparing interpolation weights for output grid -0: Total number of wet points for interpolation 29591 -0: -0: -0: Variable: Grid Interpolation Map Units: 0.100E+01 -0: -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 331 | | -0: | | -0: | | -0: 289 | | -0: | | -0: | 0 0 0 | -0: 247 | 0 0 0 0 | -0: | 0 0 0 0 0 | -0: | 0 0 0 0 | -0: 205 | 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 163 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 121 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 79 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 | -0: | 0 0 | -0: 37 | 0 | -0: | | -0: | | -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: -0: -0: Interpolating fields .... -0: -0: Output group 1 -0: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -0: Output group 2 -0: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -0: Output group 3 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 4 -0: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -0: Output group 5 -0: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -0: Output group 6 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 7 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 8 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 9 -0: Output variables skipped -0: Output group 10 -0: Output variables skipped -0: ------------------------------------------------ -0: 1Current vel. -0: 1Wind speed -0: 1Ice concentration -0: 2Wave height -0: 2Mean wave period(+2) -0: 2Mean wave period(+1) -0: 2Peak frequency -0: 2Mean wave dir. a1b1 -0: 2Peak direction -0: 4Part. wave height -0: 4Part. peak period -0: 4Part. mean direction -0: 5Charnock parameter -0: ------------------------------------------------ -0: OUTPUT TIME : 2021/03/27 18:00:00 UTC -0: -0: End of file reached -0: -0: -0: *** End of Grid interpolation Routine *** -0: =============================================== -0: -0: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -0: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -0: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_at_10m/out_grd.at_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.at_10m -0: + cmdfile.1[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032718 102 atlocn 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + cmdfile.1[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib2_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grib2_sbs.sh[30]grdID=at_10m -0: + wave_grib2_sbs.sh[31]GRIDNR=255 -0: + wave_grib2_sbs.sh[32]MODNR=11 -0: + wave_grib2_sbs.sh[33]valid_time=2021032718 -0: + wave_grib2_sbs.sh[34]fhr=102 -0: + wave_grib2_sbs.sh[35]grid_region=atlocn -0: + wave_grib2_sbs.sh[36]grid_res=0p16 -0: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010 -0: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_at_10m -0: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_at_10m -0: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_at_10m -0: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_at_10m -0: ++ wave_grib2_sbs.sh[47]printf %03i 102 -0: + wave_grib2_sbs.sh[47]FH3=102 -0: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_atlocn_0p16 -0: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.atlocn.0p16.f102.grib2 -0: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f102.grib2 ]] -0: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mod_def.at_10m ./mod_def.ww3 -0: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.at_10m ./out_grd.ww3 -0: + wave_grib2_sbs.sh[73]ngrib=1 -0: + wave_grib2_sbs.sh[74]dtgrib=3600 -0: + wave_grib2_sbs.sh[75]tstart='20210327 180000' -0: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 180000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.at_10m.inp.tmpl -0: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -0: $ WAVEWATCH-III gridded output input file -0: $ ---------------------------------------- -0: 20210327 180000 3600 1 -0: N -0: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -0: $ -0: 20210327 180000 7 11 255 0 0 -0: $ -0: $ end of input file -0: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[88]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.306705 ']' -0: ++ prep_step[4]echo gfs_ww3_grib.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[90]export err=0 -0: + wave_grib2_sbs.sh[90]err=0 -0: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[95]cat grib2_atlocn_102.out -0: -0: *** WAVEWATCH III GRIB output postp. *** -0: ============================================== -0: -0: Comment character is '$' -0: -0: Grid name : NW Atlantic 10 min wave grid -0: -0: LINEIN: -0: 20210327 180000 3600 1 -0: -0: 20210327180000 3600 1 -0: GEN_PRO -99999 -0: -0: Output time data : -0: ----------------------------------------------------- -0: First time : 2021/03/27 18:00:00 UTC -0: Interval : 01:00:00 -0: Number of requests : 1 -0: Fields : Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: Requested output fields not yet available: -0: ----------------------------------------------------- -0: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -0: -0: Successfully requested output fields : -0: ----------------------------------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: -0: Additional GRIB parameters : -0: ----------------------------------------------------- -0: Run time : 2021/03/27 18:00:00 UTC -0: GRIB center ID : 7 -0: GRIB gen. proc. ID : 11 -0: GRIB grid ID : 255 -0: GRIB GDS parameter : 0 -0: Fields in file : -0: -------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: CHOSEN GRID TYPE: : LLRECTILINEAR -0: -0: -0: -0: Generating file -0: ----------------------------------------------------- -0: Data for 2021/03/27 18:00:00 UTC 0H forecast. -0: -0: End of program -0: ========================================= -0: WAVEWATCH III GRIB output -0: -0: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -0: + wave_grib2_sbs.sh[102][[ 102 -gt 0 ]] -0: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '102 hour fcst' -grib gfs.wave.t12z.atlocn.0p16.f102.grib2 -0: 1:0:d=2021032312:SPC:surface:102 hour fcst: -0: 2:20080:d=2021032312:DIRC:surface:102 hour fcst: -0: 3:61173:d=2021032312:UOGRD:surface:102 hour fcst: -0: 4:80609:d=2021032312:VOGRD:surface:102 hour fcst: -0: 5:100901:d=2021032312:WIND:surface:102 hour fcst: -0: 6:131054:d=2021032312:WDIR:surface:102 hour fcst: -0: 7:171056:d=2021032312:UGRD:surface:102 hour fcst: -0: 8:201117:d=2021032312:VGRD:surface:102 hour fcst: -0: 9:230902:d=2021032312:ICEC:surface:102 hour fcst: -0: 10:243537:d=2021032312:HTSGW:surface:102 hour fcst: -0: 11:268956:d=2021032312:IMWF:surface:102 hour fcst: -0: 12:295524:d=2021032312:MWSPER:surface:102 hour fcst: -0: 13:322275:d=2021032312:PERPW:surface:102 hour fcst: -0: 14:349992:d=2021032312:WWSDIR:surface:102 hour fcst: -0: 15:389791:d=2021032312:DIRPW:surface:102 hour fcst: -0: 16:430067:d=2021032312:WVHGT:surface:102 hour fcst: -0: 17:455034:d=2021032312:SWELL:1 in sequence:102 hour fcst: -0: 18:476161:d=2021032312:SWELL:2 in sequence:102 hour fcst: -0: 19:492174:d=2021032312:SWELL:3 in sequence:102 hour fcst: -0: 20:505860:d=2021032312:WVPER:surface:102 hour fcst: -0: 21:532878:d=2021032312:SWPER:1 in sequence:102 hour fcst: -0: 22:555459:d=2021032312:SWPER:2 in sequence:102 hour fcst: -0: 23:573302:d=2021032312:SWPER:3 in sequence:102 hour fcst: -0: 24:587827:d=2021032312:WVDIR:surface:102 hour fcst: -0: 25:625972:d=2021032312:SWDIR:1 in sequence:102 hour fcst: -0: 26:658088:d=2021032312:SWDIR:2 in sequence:102 hour fcst: -0: 27:680620:d=2021032312:SWDIR:3 in sequence:102 hour fcst: -0: + wave_grib2_sbs.sh[104]err=0 -0: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.atlocn.0p16.f102.grib2 -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f102.grib2 ]] -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f102.grib2.idx ]] -0: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.atlocn.0p16.f102.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f102.grib2 -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f102.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f102.grib2 = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f102.grib2 ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f102.grib2 -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f102.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f102.grib2.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f102.grib2.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f102.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f102.grib2 -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.atlocn.0p16.f102.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f102.grib2.idx -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f102.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f102.grib2.idx = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f102.grib2.idx ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f102.grib2.idx -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f102.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f102.grib2.idx.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f102.grib2.idx.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f102.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f102.grib2.idx -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.atlocn.0p16.f102.grib2 and gfs.wave.t12z.atlocn.0p16.f102.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_at_10m to COM' -0: INFO: Copied gfs.wave.t12z.atlocn.0p16.f102.grib2 and gfs.wave.t12z.atlocn.0p16.f102.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_at_10m to COM -0: + wave_grib2_sbs.sh[130][[ at_10m == '' ]] -0: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -0: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.atlocn.0p16.f102.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -0: INFO: gfs.wave.t12z.atlocn.0p16.f102.grib2 is global.0p50 or SENDDBN is NO, no alert sent -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + cmdfile.6[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032718 3600. 9999 -5: + cmdfile.6[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grid_interp_sbs.sh[25]grdID=gnh_10m -5: + wave_grid_interp_sbs.sh[26]valid_time=2021032718 -5: + wave_grid_interp_sbs.sh[27]dt=3600. -5: + wave_grid_interp_sbs.sh[28]nst=9999 -5: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010 -5: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.uglo_100km ./out_grd.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mod_def.uglo_100km ./mod_def.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mod_def.gnh_10m ./mod_def.gnh_10m -5: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m'\''' -5: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m' -5: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[51]weights_found=1 -5: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 180000' -5: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 180000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gnh_10m/g ww3_gint.inp.tmpl -5: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -5: $ Input file for interpolation of uglo_100km to gnh_10m -5: $------------------------------------------------ -5: $ Start Time 3600. NSteps -5: 20210327 180000 3600. 9999 -5: $ Total number of grids -5: 2 -5: $ Grid extensions -5: 'uglo_100km' -5: 'gnh_10m' -5: $ -5: 0 -5: $ -5: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[70]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.306705 ']' -5: ++ prep_step[4]echo gfs_ww3_gint.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -5: INFO: Executing 'gfs_ww3_gint.x' -5: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[73]cat grid_interp.gnh_10m.out -5: -5: *** WAVEWATCH III Grid interpolation *** -5: =============================================== -5: -5: Comment character is '$' -5: -5: Time Information : -5: --------------------------------------------- -5: Starting Time : 2021/03/27 18:00:00 UTC -5: Interval (in sec) : 3600.00 -5: Number of requests : 9999 -5: --------------------------------------------- -5: Number of grids (including output grid) = 2 -5: -5: -5: Extension for grid 1 is --> uglo_100km -5: -5: Grid Particulars are : -5: Dimensions = 45166 1 -5: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Extension for grid 2 is --> gnh_10m -5: -5: Grid Particulars are : -5: Dimensions = 2160 406 -5: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -5: -5: -5: Preparing interpolation weights for output grid -5: Total number of wet points for interpolation 571209 -5: -5: -5: Variable: Grid Interpolation Map Units: 0.100E+01 -5: -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 406 | | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: 355 | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 304 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 253 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 202 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 151 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 100 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 49 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: -5: -5: Interpolating fields .... -5: -5: Output group 1 -5: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -5: Output group 2 -5: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -5: Output group 3 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 4 -5: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -5: Output group 5 -5: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -5: Output group 6 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 7 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 8 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 9 -5: Output variables skipped -5: Output group 10 -5: Output variables skipped -5: ------------------------------------------------ -5: 1Current vel. -5: 1Wind speed -5: 1Ice concentration -5: 2Wave height -5: 2Mean wave period(+2) -5: 2Mean wave period(+1) -5: 2Peak frequency -5: 2Mean wave dir. a1b1 -5: 2Peak direction -5: 4Part. wave height -5: 4Part. peak period -5: 4Part. mean direction -5: 5Charnock parameter -5: ------------------------------------------------ -5: OUTPUT TIME : 2021/03/27 18:00:00 UTC -5: -5: End of file reached -5: -5: -5: *** End of Grid interpolation Routine *** -5: =============================================== -5: -5: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -5: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -5: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_gnh_10m/out_grd.gnh_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.gnh_10m -5: + cmdfile.6[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032718 102 global 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + cmdfile.6[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib2_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grib2_sbs.sh[30]grdID=gnh_10m -5: + wave_grib2_sbs.sh[31]GRIDNR=255 -5: + wave_grib2_sbs.sh[32]MODNR=11 -5: + wave_grib2_sbs.sh[33]valid_time=2021032718 -5: + wave_grib2_sbs.sh[34]fhr=102 -5: + wave_grib2_sbs.sh[35]grid_region=global -5: + wave_grib2_sbs.sh[36]grid_res=0p16 -5: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010 -5: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_gnh_10m -5: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_gnh_10m -5: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_gnh_10m -5: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_gnh_10m -5: ++ wave_grib2_sbs.sh[47]printf %03i 102 -5: + wave_grib2_sbs.sh[47]FH3=102 -5: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p16 -5: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p16.f102.grib2 -5: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f102.grib2 ]] -5: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mod_def.gnh_10m ./mod_def.ww3 -5: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.gnh_10m ./out_grd.ww3 -5: + wave_grib2_sbs.sh[73]ngrib=1 -5: + wave_grib2_sbs.sh[74]dtgrib=3600 -5: + wave_grib2_sbs.sh[75]tstart='20210327 180000' -5: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 180000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gnh_10m.inp.tmpl -5: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -5: $ WAVEWATCH-III gridded output input file -5: $ ---------------------------------------- -5: 20210327 180000 3600 1 -5: N -5: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -5: $ -5: 20210327 180000 7 11 255 0 0 -5: $ -5: $ end of input file -5: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[88]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.306705 ']' -5: ++ prep_step[4]echo gfs_ww3_grib.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[90]export err=0 -5: + wave_grib2_sbs.sh[90]err=0 -5: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[95]cat grib2_global_102.out -5: -5: *** WAVEWATCH III GRIB output postp. *** -5: ============================================== -5: -5: Comment character is '$' -5: -5: Grid name : GFSv16-wave N Hemisphere 1/6 d -5: -5: LINEIN: -5: 20210327 180000 3600 1 -5: -5: 20210327180000 3600 1 -5: GEN_PRO -99999 -5: -5: Output time data : -5: ----------------------------------------------------- -5: First time : 2021/03/27 18:00:00 UTC -5: Interval : 01:00:00 -5: Number of requests : 1 -5: Fields : Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: Requested output fields not yet available: -5: ----------------------------------------------------- -5: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -5: -5: Successfully requested output fields : -5: ----------------------------------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: -5: Additional GRIB parameters : -5: ----------------------------------------------------- -5: Run time : 2021/03/27 18:00:00 UTC -5: GRIB center ID : 7 -5: GRIB gen. proc. ID : 11 -5: GRIB grid ID : 255 -5: GRIB GDS parameter : 0 -5: Fields in file : -5: -------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: CHOSEN GRID TYPE: : LLRECTILINEAR -5: -5: -5: -5: Generating file -5: ----------------------------------------------------- -5: Data for 2021/03/27 18:00:00 UTC 0H forecast. -5: -5: End of program -5: ========================================= -5: WAVEWATCH III GRIB output -5: -5: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -5: + wave_grib2_sbs.sh[102][[ 102 -gt 0 ]] -5: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '102 hour fcst' -grib gfs.wave.t12z.global.0p16.f102.grib2 -5: 1:0:d=2021032312:SPC:surface:102 hour fcst: -5: 2:237972:d=2021032312:DIRC:surface:102 hour fcst: -5: 3:874844:d=2021032312:UOGRD:surface:102 hour fcst: -5: 4:1108886:d=2021032312:VOGRD:surface:102 hour fcst: -5: 5:1348199:d=2021032312:WIND:surface:102 hour fcst: -5: 6:1757450:d=2021032312:WDIR:surface:102 hour fcst: -5: 7:2393836:d=2021032312:UGRD:surface:102 hour fcst: -5: 8:2794843:d=2021032312:VGRD:surface:102 hour fcst: -5: 9:3197697:d=2021032312:ICEC:surface:102 hour fcst: -5: 10:3312205:d=2021032312:HTSGW:surface:102 hour fcst: -5: 11:3601081:d=2021032312:IMWF:surface:102 hour fcst: -5: 12:3914077:d=2021032312:MWSPER:surface:102 hour fcst: -5: 13:4229378:d=2021032312:PERPW:surface:102 hour fcst: -5: 14:4565818:d=2021032312:WWSDIR:surface:102 hour fcst: -5: 15:5145976:d=2021032312:DIRPW:surface:102 hour fcst: -5: 16:5749966:d=2021032312:WVHGT:surface:102 hour fcst: -5: 17:6026662:d=2021032312:SWELL:1 in sequence:102 hour fcst: -5: 18:6320619:d=2021032312:SWELL:2 in sequence:102 hour fcst: -5: 19:6520572:d=2021032312:SWELL:3 in sequence:102 hour fcst: -5: 20:6649175:d=2021032312:WVPER:surface:102 hour fcst: -5: 21:6958731:d=2021032312:SWPER:1 in sequence:102 hour fcst: -5: 22:7291674:d=2021032312:SWPER:2 in sequence:102 hour fcst: -5: 23:7537084:d=2021032312:SWPER:3 in sequence:102 hour fcst: -5: 24:7682605:d=2021032312:WVDIR:surface:102 hour fcst: -5: 25:8174780:d=2021032312:SWDIR:1 in sequence:102 hour fcst: -5: 26:8775128:d=2021032312:SWDIR:2 in sequence:102 hour fcst: -5: 27:9163038:d=2021032312:SWDIR:3 in sequence:102 hour fcst: -5: + wave_grib2_sbs.sh[104]err=0 -5: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p16.f102.grib2 -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f102.grib2 ]] -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f102.grib2.idx ]] -5: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p16.f102.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f102.grib2 -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f102.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f102.grib2 = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f102.grib2 ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f102.grib2 -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f102.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f102.grib2.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f102.grib2.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f102.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f102.grib2 -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p16.f102.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f102.grib2.idx -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f102.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f102.grib2.idx = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f102.grib2.idx ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f102.grib2.idx -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f102.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f102.grib2.idx.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f102.grib2.idx.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f102.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f102.grib2.idx -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p16.f102.grib2 and gfs.wave.t12z.global.0p16.f102.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_gnh_10m to COM' -5: INFO: Copied gfs.wave.t12z.global.0p16.f102.grib2 and gfs.wave.t12z.global.0p16.f102.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_gnh_10m to COM -5: + wave_grib2_sbs.sh[130][[ gnh_10m == '' ]] -5: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -5: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p16.f102.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -5: INFO: gfs.wave.t12z.global.0p16.f102.grib2 is global.0p50 or SENDDBN is NO, no alert sent -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + cmdfile.4[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032718 3600. 9999 -3: + cmdfile.4[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grid_interp_sbs.sh[25]grdID=glo_30m -3: + wave_grid_interp_sbs.sh[26]valid_time=2021032718 -3: + wave_grid_interp_sbs.sh[27]dt=3600. -3: + wave_grid_interp_sbs.sh[28]nst=9999 -3: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010 -3: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.uglo_100km ./out_grd.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mod_def.uglo_100km ./mod_def.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mod_def.glo_30m ./mod_def.glo_30m -3: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ]] -3: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m'\''' -3: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m' -3: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[51]weights_found=1 -3: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 180000' -3: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 180000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/glo_30m/g ww3_gint.inp.tmpl -3: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -3: $ Input file for interpolation of uglo_100km to glo_30m -3: $------------------------------------------------ -3: $ Start Time 3600. NSteps -3: 20210327 180000 3600. 9999 -3: $ Total number of grids -3: 2 -3: $ Grid extensions -3: 'uglo_100km' -3: 'glo_30m' -3: $ -3: 0 -3: $ -3: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[70]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.306705 ']' -3: ++ prep_step[4]echo gfs_ww3_gint.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -3: INFO: Executing 'gfs_ww3_gint.x' -3: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[73]cat grid_interp.glo_30m.out -3: -3: *** WAVEWATCH III Grid interpolation *** -3: =============================================== -3: -3: Comment character is '$' -3: -3: Time Information : -3: --------------------------------------------- -3: Starting Time : 2021/03/27 18:00:00 UTC -3: Interval (in sec) : 3600.00 -3: Number of requests : 9999 -3: --------------------------------------------- -3: Number of grids (including output grid) = 2 -3: -3: -3: Extension for grid 1 is --> uglo_100km -3: -3: Grid Particulars are : -3: Dimensions = 45166 1 -3: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Extension for grid 2 is --> glo_30m -3: -3: Grid Particulars are : -3: Dimensions = 720 336 -3: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -3: -3: -3: Preparing interpolation weights for output grid -3: Total number of wet points for interpolation 167619 -3: -3: -3: Variable: Grid Interpolation Map Units: 0.100E+01 -3: -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 336 | | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 291 | 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 | -3: 246 | 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 201 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 156 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 111 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 66 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 21 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 | -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: -3: -3: Interpolating fields .... -3: -3: Output group 1 -3: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -3: Output group 2 -3: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -3: Output group 3 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 4 -3: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -3: Output group 5 -3: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -3: Output group 6 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 7 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 8 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 9 -3: Output variables skipped -3: Output group 10 -3: Output variables skipped -3: ------------------------------------------------ -3: 1Current vel. -3: 1Wind speed -3: 1Ice concentration -3: 2Wave height -3: 2Mean wave period(+2) -3: 2Mean wave period(+1) -3: 2Peak frequency -3: 2Mean wave dir. a1b1 -3: 2Peak direction -3: 4Part. wave height -3: 4Part. peak period -3: 4Part. mean direction -3: 5Charnock parameter -3: ------------------------------------------------ -3: OUTPUT TIME : 2021/03/27 18:00:00 UTC -3: -3: End of file reached -3: -3: -3: *** End of Grid interpolation Routine *** -3: =============================================== -3: -3: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -3: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -3: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_glo_30m/out_grd.glo_30m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.glo_30m -3: + cmdfile.4[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032718 102 global 0p50 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + cmdfile.4[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib2_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grib2_sbs.sh[30]grdID=glo_30m -3: + wave_grib2_sbs.sh[31]GRIDNR=255 -3: + wave_grib2_sbs.sh[32]MODNR=11 -3: + wave_grib2_sbs.sh[33]valid_time=2021032718 -3: + wave_grib2_sbs.sh[34]fhr=102 -3: + wave_grib2_sbs.sh[35]grid_region=global -3: + wave_grib2_sbs.sh[36]grid_res=0p50 -3: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010 -3: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_glo_30m -3: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_glo_30m -3: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_glo_30m -3: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_glo_30m -3: ++ wave_grib2_sbs.sh[47]printf %03i 102 -3: + wave_grib2_sbs.sh[47]FH3=102 -3: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p50 -3: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p50.f102.grib2 -3: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f102.grib2 ]] -3: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mod_def.glo_30m ./mod_def.ww3 -3: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.glo_30m ./out_grd.ww3 -3: + wave_grib2_sbs.sh[73]ngrib=1 -3: + wave_grib2_sbs.sh[74]dtgrib=3600 -3: + wave_grib2_sbs.sh[75]tstart='20210327 180000' -3: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 180000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.glo_30m.inp.tmpl -3: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -3: $ WAVEWATCH-III gridded output input file -3: $ ---------------------------------------- -3: 20210327 180000 3600 1 -3: N -3: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -3: $ -3: 20210327 180000 7 11 255 0 0 -3: $ -3: $ end of input file -3: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[88]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.306705 ']' -3: ++ prep_step[4]echo gfs_ww3_grib.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[90]export err=0 -3: + wave_grib2_sbs.sh[90]err=0 -3: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[95]cat grib2_global_102.out -3: -3: *** WAVEWATCH III GRIB output postp. *** -3: ============================================== -3: -3: Comment character is '$' -3: -3: Grid name : Global 30 min wave grid -3: -3: LINEIN: -3: 20210327 180000 3600 1 -3: -3: 20210327180000 3600 1 -3: GEN_PRO -99999 -3: -3: Output time data : -3: ----------------------------------------------------- -3: First time : 2021/03/27 18:00:00 UTC -3: Interval : 01:00:00 -3: Number of requests : 1 -3: Fields : Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: Requested output fields not yet available: -3: ----------------------------------------------------- -3: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -3: -3: Successfully requested output fields : -3: ----------------------------------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: -3: Additional GRIB parameters : -3: ----------------------------------------------------- -3: Run time : 2021/03/27 18:00:00 UTC -3: GRIB center ID : 7 -3: GRIB gen. proc. ID : 11 -3: GRIB grid ID : 255 -3: GRIB GDS parameter : 0 -3: Fields in file : -3: -------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: CHOSEN GRID TYPE: : LLRECTILINEAR -3: -3: -3: -3: Generating file -3: ----------------------------------------------------- -3: Data for 2021/03/27 18:00:00 UTC 0H forecast. -3: -3: End of program -3: ========================================= -3: WAVEWATCH III GRIB output -3: -3: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -3: + wave_grib2_sbs.sh[102][[ 102 -gt 0 ]] -3: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '102 hour fcst' -grib gfs.wave.t12z.global.0p50.f102.grib2 -3: 1:0:d=2021032312:SPC:surface:102 hour fcst: -3: 2:77828:d=2021032312:DIRC:surface:102 hour fcst: -3: 3:322319:d=2021032312:UOGRD:surface:102 hour fcst: -3: 4:397668:d=2021032312:VOGRD:surface:102 hour fcst: -3: 5:477101:d=2021032312:WIND:surface:102 hour fcst: -3: 6:635692:d=2021032312:WDIR:surface:102 hour fcst: -3: 7:871825:d=2021032312:UGRD:surface:102 hour fcst: -3: 8:1026720:d=2021032312:VGRD:surface:102 hour fcst: -3: 9:1183301:d=2021032312:ICEC:surface:102 hour fcst: -3: 10:1223485:d=2021032312:HTSGW:surface:102 hour fcst: -3: 11:1322224:d=2021032312:IMWF:surface:102 hour fcst: -3: 12:1432275:d=2021032312:MWSPER:surface:102 hour fcst: -3: 13:1543209:d=2021032312:PERPW:surface:102 hour fcst: -3: 14:1660536:d=2021032312:WWSDIR:surface:102 hour fcst: -3: 15:1861146:d=2021032312:DIRPW:surface:102 hour fcst: -3: 16:2067109:d=2021032312:WVHGT:surface:102 hour fcst: -3: 17:2162407:d=2021032312:SWELL:1 in sequence:102 hour fcst: -3: 18:2267286:d=2021032312:SWELL:2 in sequence:102 hour fcst: -3: 19:2340903:d=2021032312:SWELL:3 in sequence:102 hour fcst: -3: 20:2385167:d=2021032312:WVPER:surface:102 hour fcst: -3: 21:2491873:d=2021032312:SWPER:1 in sequence:102 hour fcst: -3: 22:2610230:d=2021032312:SWPER:2 in sequence:102 hour fcst: -3: 23:2703197:d=2021032312:SWPER:3 in sequence:102 hour fcst: -3: 24:2756029:d=2021032312:WVDIR:surface:102 hour fcst: -3: 25:2917319:d=2021032312:SWDIR:1 in sequence:102 hour fcst: -3: 26:3120970:d=2021032312:SWDIR:2 in sequence:102 hour fcst: -3: 27:3267827:d=2021032312:SWDIR:3 in sequence:102 hour fcst: -3: + wave_grib2_sbs.sh[104]err=0 -3: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p50.f102.grib2 -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f102.grib2 ]] -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f102.grib2.idx ]] -3: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p50.f102.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f102.grib2 -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f102.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f102.grib2 = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f102.grib2 ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f102.grib2 -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f102.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f102.grib2.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f102.grib2.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f102.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f102.grib2 -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p50.f102.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f102.grib2.idx -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f102.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f102.grib2.idx = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f102.grib2.idx ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f102.grib2.idx -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f102.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f102.grib2.idx.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f102.grib2.idx.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f102.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f102.grib2.idx -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p50.f102.grib2 and gfs.wave.t12z.global.0p50.f102.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_glo_30m to COM' -3: INFO: Copied gfs.wave.t12z.global.0p50.f102.grib2 and gfs.wave.t12z.global.0p50.f102.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_glo_30m to COM -3: + wave_grib2_sbs.sh[130][[ glo_30m == '' ]] -3: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -3: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p50.f102.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -3: INFO: gfs.wave.t12z.global.0p50.f102.grib2 is global.0p50 or SENDDBN is NO, no alert sent -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + cmdfile.5[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032718 3600. 9999 -4: + cmdfile.5[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grid_interp_sbs.sh[25]grdID=aoc_9km -4: + wave_grid_interp_sbs.sh[26]valid_time=2021032718 -4: + wave_grid_interp_sbs.sh[27]dt=3600. -4: + wave_grid_interp_sbs.sh[28]nst=9999 -4: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010 -4: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.uglo_100km ./out_grd.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mod_def.uglo_100km ./mod_def.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mod_def.aoc_9km ./mod_def.aoc_9km -4: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km'\''' -4: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km' -4: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[51]weights_found=1 -4: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 180000' -4: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 180000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/aoc_9km/g ww3_gint.inp.tmpl -4: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -4: $ Input file for interpolation of uglo_100km to aoc_9km -4: $------------------------------------------------ -4: $ Start Time 3600. NSteps -4: 20210327 180000 3600. 9999 -4: $ Total number of grids -4: 2 -4: $ Grid extensions -4: 'uglo_100km' -4: 'aoc_9km' -4: $ -4: 0 -4: $ -4: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[70]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.306705 ']' -4: ++ prep_step[4]echo gfs_ww3_gint.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -4: INFO: Executing 'gfs_ww3_gint.x' -4: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[73]cat grid_interp.aoc_9km.out -4: -4: *** WAVEWATCH III Grid interpolation *** -4: =============================================== -4: -4: Comment character is '$' -4: -4: Time Information : -4: --------------------------------------------- -4: Starting Time : 2021/03/27 18:00:00 UTC -4: Interval (in sec) : 3600.00 -4: Number of requests : 9999 -4: --------------------------------------------- -4: Number of grids (including output grid) = 2 -4: -4: -4: Extension for grid 1 is --> uglo_100km -4: -4: Grid Particulars are : -4: Dimensions = 45166 1 -4: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Extension for grid 2 is --> aoc_9km -4: -4: Grid Particulars are : -4: Dimensions = 1006 1006 -4: Grid Type = 2 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -4: -4: -4: Preparing interpolation weights for output grid -4: Total number of wet points for interpolation 360052 -4: -4: -4: Variable: Grid Interpolation Map Units: 0.100E+01 -4: -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: *** | | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: 880 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 | -4: 754 | 0 0 0 | -4: | 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 628 | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 | -4: 502 | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 376 | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 250 | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 124 | 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 | -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: -4: -4: Interpolating fields .... -4: -4: Output group 1 -4: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -4: Output group 2 -4: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -4: Output group 3 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 4 -4: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -4: Output group 5 -4: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -4: Output group 6 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 7 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 8 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 9 -4: Output variables skipped -4: Output group 10 -4: Output variables skipped -4: ------------------------------------------------ -4: 1Current vel. -4: 1Wind speed -4: 1Ice concentration -4: 2Wave height -4: 2Mean wave period(+2) -4: 2Mean wave period(+1) -4: 2Peak frequency -4: 2Mean wave dir. a1b1 -4: 2Peak direction -4: 4Part. wave height -4: 4Part. peak period -4: 4Part. mean direction -4: 5Charnock parameter -4: ------------------------------------------------ -4: OUTPUT TIME : 2021/03/27 18:00:00 UTC -4: -4: End of file reached -4: -4: -4: *** End of Grid interpolation Routine *** -4: =============================================== -4: -4: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -4: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -4: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_aoc_9km/out_grd.aoc_9km /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.aoc_9km -4: + cmdfile.5[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032718 102 arctic 9km 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + cmdfile.5[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib2_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grib2_sbs.sh[30]grdID=aoc_9km -4: + wave_grib2_sbs.sh[31]GRIDNR=255 -4: + wave_grib2_sbs.sh[32]MODNR=11 -4: + wave_grib2_sbs.sh[33]valid_time=2021032718 -4: + wave_grib2_sbs.sh[34]fhr=102 -4: + wave_grib2_sbs.sh[35]grid_region=arctic -4: + wave_grib2_sbs.sh[36]grid_res=9km -4: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010 -4: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_aoc_9km -4: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_aoc_9km -4: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_aoc_9km -4: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_aoc_9km -4: ++ wave_grib2_sbs.sh[47]printf %03i 102 -4: + wave_grib2_sbs.sh[47]FH3=102 -4: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_arctic_9km -4: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.arctic.9km.f102.grib2 -4: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f102.grib2 ]] -4: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mod_def.aoc_9km ./mod_def.ww3 -4: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.aoc_9km ./out_grd.ww3 -4: + wave_grib2_sbs.sh[73]ngrib=1 -4: + wave_grib2_sbs.sh[74]dtgrib=3600 -4: + wave_grib2_sbs.sh[75]tstart='20210327 180000' -4: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 180000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.aoc_9km.inp.tmpl -4: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -4: $ WAVEWATCH-III gridded output input file -4: $ ---------------------------------------- -4: 20210327 180000 3600 1 -4: N -4: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -4: $ -4: 20210327 180000 7 11 255 0 20 -4: $ -4: 70 0 9.0 9.0 64 -4: $ 60 0 8.64919046313 8.64919046313 64 -4: $ end of input file -4: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[88]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.306705 ']' -4: ++ prep_step[4]echo gfs_ww3_grib.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[90]export err=0 -4: + wave_grib2_sbs.sh[90]err=0 -4: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[95]cat grib2_arctic_102.out -4: -4: *** WAVEWATCH III GRIB output postp. *** -4: ============================================== -4: -4: Comment character is '$' -4: -4: Grid name : Arctic Ocean PolarStereo 9km -4: -4: LINEIN: -4: 20210327 180000 3600 1 -4: -4: 20210327180000 3600 1 -4: GEN_PRO -99999 -4: -4: Output time data : -4: ----------------------------------------------------- -4: First time : 2021/03/27 18:00:00 UTC -4: Interval : 01:00:00 -4: Number of requests : 1 -4: Fields : Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: Requested output fields not yet available: -4: ----------------------------------------------------- -4: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -4: -4: Successfully requested output fields : -4: ----------------------------------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: -4: Additional GRIB parameters : -4: ----------------------------------------------------- -4: Run time : 2021/03/27 18:00:00 UTC -4: GRIB center ID : 7 -4: GRIB gen. proc. ID : 11 -4: GRIB grid ID : 255 -4: GRIB GDS parameter : 0 -4: Fields in file : -4: -------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: CHOSEN GRID TYPE: : POLARSTEREO -4: -4: -4: -4: Generating file -4: ----------------------------------------------------- -4: Data for 2021/03/27 18:00:00 UTC 0H forecast. -4: -4: End of program -4: ========================================= -4: WAVEWATCH III GRIB output -4: -4: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -4: + wave_grib2_sbs.sh[102][[ 102 -gt 0 ]] -4: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '102 hour fcst' -grib gfs.wave.t12z.arctic.9km.f102.grib2 -4: 1:0:d=2021032312:SPC:surface:102 hour fcst: -4: 2:190733:d=2021032312:DIRC:surface:102 hour fcst: -4: 3:657728:d=2021032312:UOGRD:surface:102 hour fcst: -4: 4:844559:d=2021032312:VOGRD:surface:102 hour fcst: -4: 5:1035402:d=2021032312:WIND:surface:102 hour fcst: -4: 6:1338040:d=2021032312:WDIR:surface:102 hour fcst: -4: 7:1770226:d=2021032312:UGRD:surface:102 hour fcst: -4: 8:2068325:d=2021032312:VGRD:surface:102 hour fcst: -4: 9:2371084:d=2021032312:ICEC:surface:102 hour fcst: -4: 10:2540628:d=2021032312:HTSGW:surface:102 hour fcst: -4: 11:2755635:d=2021032312:IMWF:surface:102 hour fcst: -4: 12:2974502:d=2021032312:MWSPER:surface:102 hour fcst: -4: 13:3195159:d=2021032312:PERPW:surface:102 hour fcst: -4: 14:3425259:d=2021032312:WWSDIR:surface:102 hour fcst: -4: 15:3734751:d=2021032312:DIRPW:surface:102 hour fcst: -4: 16:4049950:d=2021032312:WVHGT:surface:102 hour fcst: -4: 17:4262468:d=2021032312:SWELL:1 in sequence:102 hour fcst: -4: 18:4474312:d=2021032312:SWELL:2 in sequence:102 hour fcst: -4: 19:4637445:d=2021032312:SWELL:3 in sequence:102 hour fcst: -4: 20:4771783:d=2021032312:WVPER:surface:102 hour fcst: -4: 21:4994319:d=2021032312:SWPER:1 in sequence:102 hour fcst: -4: 22:5215132:d=2021032312:SWPER:2 in sequence:102 hour fcst: -4: 23:5389763:d=2021032312:SWPER:3 in sequence:102 hour fcst: -4: 24:5527997:d=2021032312:WVDIR:surface:102 hour fcst: -4: 25:5816616:d=2021032312:SWDIR:1 in sequence:102 hour fcst: -4: 26:6128053:d=2021032312:SWDIR:2 in sequence:102 hour fcst: -4: 27:6344088:d=2021032312:SWDIR:3 in sequence:102 hour fcst: -4: + wave_grib2_sbs.sh[104]err=0 -4: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.arctic.9km.f102.grib2 -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f102.grib2 ]] -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f102.grib2.idx ]] -4: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.arctic.9km.f102.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f102.grib2 -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f102.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f102.grib2 = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f102.grib2 ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f102.grib2 -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f102.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f102.grib2.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f102.grib2.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f102.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f102.grib2 -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.arctic.9km.f102.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f102.grib2.idx -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f102.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f102.grib2.idx = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f102.grib2.idx ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f102.grib2.idx -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f102.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f102.grib2.idx.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f102.grib2.idx.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f102.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f102.grib2.idx -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.arctic.9km.f102.grib2 and gfs.wave.t12z.arctic.9km.f102.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_aoc_9km to COM' -4: INFO: Copied gfs.wave.t12z.arctic.9km.f102.grib2 and gfs.wave.t12z.arctic.9km.f102.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_aoc_9km to COM -4: + wave_grib2_sbs.sh[130][[ aoc_9km == '' ]] -4: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -4: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.arctic.9km.f102.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -4: INFO: gfs.wave.t12z.arctic.9km.f102.grib2 is global.0p50 or SENDDBN is NO, no alert sent -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + cmdfile.3[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032718 3600. 9999 -2: + cmdfile.3[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grid_interp_sbs.sh[25]grdID=wc_10m -2: + wave_grid_interp_sbs.sh[26]valid_time=2021032718 -2: + wave_grid_interp_sbs.sh[27]dt=3600. -2: + wave_grid_interp_sbs.sh[28]nst=9999 -2: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010 -2: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.uglo_100km ./out_grd.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mod_def.uglo_100km ./mod_def.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mod_def.wc_10m ./mod_def.wc_10m -2: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ]] -2: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m'\''' -2: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m' -2: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[51]weights_found=1 -2: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 180000' -2: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 180000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/wc_10m/g ww3_gint.inp.tmpl -2: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -2: $ Input file for interpolation of uglo_100km to wc_10m -2: $------------------------------------------------ -2: $ Start Time 3600. NSteps -2: 20210327 180000 3600. 9999 -2: $ Total number of grids -2: 2 -2: $ Grid extensions -2: 'uglo_100km' -2: 'wc_10m' -2: $ -2: 0 -2: $ -2: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[70]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.306705 ']' -2: ++ prep_step[4]echo gfs_ww3_gint.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -2: INFO: Executing 'gfs_ww3_gint.x' -2: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[73]cat grid_interp.wc_10m.out -2: -2: *** WAVEWATCH III Grid interpolation *** -2: =============================================== -2: -2: Comment character is '$' -2: -2: Time Information : -2: --------------------------------------------- -2: Starting Time : 2021/03/27 18:00:00 UTC -2: Interval (in sec) : 3600.00 -2: Number of requests : 9999 -2: --------------------------------------------- -2: Number of grids (including output grid) = 2 -2: -2: -2: Extension for grid 1 is --> uglo_100km -2: -2: Grid Particulars are : -2: Dimensions = 45166 1 -2: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Extension for grid 2 is --> wc_10m -2: -2: Grid Particulars are : -2: Dimensions = 241 151 -2: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -2: -2: -2: Preparing interpolation weights for output grid -2: Total number of wet points for interpolation 11044 -2: -2: -2: Variable: Grid Interpolation Map Units: 0.100E+01 -2: -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: +---------------------------------------------------------------------------------------------------------------+ -2: 151 | 0 0 0 0 | -2: | 0 0 0 0 0 | -2: | 0 0 0 0 0 | -2: 130 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 109 | 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 88 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 | -2: 67 | 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 | -2: 46 | 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: 25 | 0 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 0 | -2: | | -2: 4 | | -2: +---------------------------------------------------------------------------------------------------------------+ -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: -2: -2: Interpolating fields .... -2: -2: Output group 1 -2: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -2: Output group 2 -2: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -2: Output group 3 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 4 -2: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -2: Output group 5 -2: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -2: Output group 6 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 7 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 8 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 9 -2: Output variables skipped -2: Output group 10 -2: Output variables skipped -2: ------------------------------------------------ -2: 1Current vel. -2: 1Wind speed -2: 1Ice concentration -2: 2Wave height -2: 2Mean wave period(+2) -2: 2Mean wave period(+1) -2: 2Peak frequency -2: 2Mean wave dir. a1b1 -2: 2Peak direction -2: 4Part. wave height -2: 4Part. peak period -2: 4Part. mean direction -2: 5Charnock parameter -2: ------------------------------------------------ -2: OUTPUT TIME : 2021/03/27 18:00:00 UTC -2: -2: End of file reached -2: -2: -2: *** End of Grid interpolation Routine *** -2: =============================================== -2: -2: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -2: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -2: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grid_interp_wc_10m/out_grd.wc_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.wc_10m -2: + cmdfile.3[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032718 102 wcoast 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + cmdfile.3[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib2_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grib2_sbs.sh[30]grdID=wc_10m -2: + wave_grib2_sbs.sh[31]GRIDNR=255 -2: + wave_grib2_sbs.sh[32]MODNR=11 -2: + wave_grib2_sbs.sh[33]valid_time=2021032718 -2: + wave_grib2_sbs.sh[34]fhr=102 -2: + wave_grib2_sbs.sh[35]grid_region=wcoast -2: + wave_grib2_sbs.sh[36]grid_res=0p16 -2: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010 -2: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_wc_10m -2: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_wc_10m -2: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_wc_10m -2: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_wc_10m -2: ++ wave_grib2_sbs.sh[47]printf %03i 102 -2: + wave_grib2_sbs.sh[47]FH3=102 -2: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_wcoast_0p16 -2: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.wcoast.0p16.f102.grib2 -2: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f102.grib2 ]] -2: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/mod_def.wc_10m ./mod_def.ww3 -2: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/out_grd.wc_10m ./out_grd.ww3 -2: + wave_grib2_sbs.sh[73]ngrib=1 -2: + wave_grib2_sbs.sh[74]dtgrib=3600 -2: + wave_grib2_sbs.sh[75]tstart='20210327 180000' -2: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 180000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.wc_10m.inp.tmpl -2: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -2: $ WAVEWATCH-III gridded output input file -2: $ ---------------------------------------- -2: 20210327 180000 3600 1 -2: N -2: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -2: $ -2: 20210327 180000 7 11 255 0 0 -2: $ -2: $ end of input file -2: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[88]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.306705 ']' -2: ++ prep_step[4]echo gfs_ww3_grib.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[90]export err=0 -2: + wave_grib2_sbs.sh[90]err=0 -2: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[95]cat grib2_wcoast_102.out -2: -2: *** WAVEWATCH III GRIB output postp. *** -2: ============================================== -2: -2: Comment character is '$' -2: -2: Grid name : West Coast 10 min wave grid -2: -2: LINEIN: -2: 20210327 180000 3600 1 -2: -2: 20210327180000 3600 1 -2: GEN_PRO -99999 -2: -2: Output time data : -2: ----------------------------------------------------- -2: First time : 2021/03/27 18:00:00 UTC -2: Interval : 01:00:00 -2: Number of requests : 1 -2: Fields : Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: Requested output fields not yet available: -2: ----------------------------------------------------- -2: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -2: -2: Successfully requested output fields : -2: ----------------------------------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: -2: Additional GRIB parameters : -2: ----------------------------------------------------- -2: Run time : 2021/03/27 18:00:00 UTC -2: GRIB center ID : 7 -2: GRIB gen. proc. ID : 11 -2: GRIB grid ID : 255 -2: GRIB GDS parameter : 0 -2: Fields in file : -2: -------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: CHOSEN GRID TYPE: : LLRECTILINEAR -2: -2: -2: -2: Generating file -2: ----------------------------------------------------- -2: Data for 2021/03/27 18:00:00 UTC 0H forecast. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: -2: End of program -2: ========================================= -2: WAVEWATCH III GRIB output -2: -2: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -2: + wave_grib2_sbs.sh[102][[ 102 -gt 0 ]] -2: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '102 hour fcst' -grib gfs.wave.t12z.wcoast.0p16.f102.grib2 -2: 1:0:d=2021032312:SPC:surface:102 hour fcst: -2: 2:7249:d=2021032312:DIRC:surface:102 hour fcst: -2: 3:24961:d=2021032312:UOGRD:surface:102 hour fcst: -2: 4:32242:d=2021032312:VOGRD:surface:102 hour fcst: -2: 5:39785:d=2021032312:WIND:surface:102 hour fcst: -2: 6:51723:d=2021032312:WDIR:surface:102 hour fcst: -2: 7:68925:d=2021032312:UGRD:surface:102 hour fcst: -2: 8:80065:d=2021032312:VGRD:surface:102 hour fcst: -2: 9:92057:d=2021032312:ICEC:surface:102 hour fcst: -2: 10:96787:d=2021032312:HTSGW:surface:102 hour fcst: -2: 11:106308:d=2021032312:IMWF:surface:102 hour fcst: -2: 12:116023:d=2021032312:MWSPER:surface:102 hour fcst: -2: 13:125777:d=2021032312:PERPW:surface:102 hour fcst: -2: 14:135814:d=2021032312:WWSDIR:surface:102 hour fcst: -2: 15:151346:d=2021032312:DIRPW:surface:102 hour fcst: -2: 16:167348:d=2021032312:WVHGT:surface:102 hour fcst: -2: 17:174781:d=2021032312:SWELL:1 in sequence:102 hour fcst: -2: 18:183149:d=2021032312:SWELL:2 in sequence:102 hour fcst: -2: 19:188369:d=2021032312:SWELL:3 in sequence:102 hour fcst: -2: 20:193099:d=2021032312:WVPER:surface:102 hour fcst: -2: 21:200994:d=2021032312:SWPER:1 in sequence:102 hour fcst: -2: 22:209385:d=2021032312:SWPER:2 in sequence:102 hour fcst: -2: 23:214845:d=2021032312:SWPER:3 in sequence:102 hour fcst: -2: 24:219575:d=2021032312:WVDIR:surface:102 hour fcst: -2: 25:229826:d=2021032312:SWDIR:1 in sequence:102 hour fcst: -2: 26:242036:d=2021032312:SWDIR:2 in sequence:102 hour fcst: -2: 27:247836:d=2021032312:SWDIR:3 in sequence:102 hour fcst: -2: + wave_grib2_sbs.sh[104]err=0 -2: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.wcoast.0p16.f102.grib2 -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f102.grib2 ]] -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f102.grib2.idx ]] -2: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.wcoast.0p16.f102.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f102.grib2 -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f102.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f102.grib2 = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f102.grib2 ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f102.grib2 -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f102.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f102.grib2.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f102.grib2.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f102.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f102.grib2 -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.wcoast.0p16.f102.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f102.grib2.idx -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f102.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f102.grib2.idx = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f102.grib2.idx ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f102.grib2.idx -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f102.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f102.grib2.idx.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f102.grib2.idx.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f102.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f102.grib2.idx -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.wcoast.0p16.f102.grib2 and gfs.wave.t12z.wcoast.0p16.f102.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_wc_10m to COM' -2: INFO: Copied gfs.wave.t12z.wcoast.0p16.f102.grib2 and gfs.wave.t12z.wcoast.0p16.f102.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010/grib_wc_10m to COM -2: + wave_grib2_sbs.sh[130][[ wc_10m == '' ]] -2: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -2: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.wcoast.0p16.f102.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -2: INFO: gfs.wave.t12z.wcoast.0p16.f102.grib2 is global.0p50 or SENDDBN is NO, no alert sent -+ run_mpmd.sh[113]exit 0 -+ run_mpmd.sh[1]postamble run_mpmd.sh 1753758279 0 -+ preamble.sh[62]set +x -End run_mpmd.sh at 03:04:50 with error code 0 (time elapsed: 00:00:11) -+ exgfs_wave_post_gridded_sbs.sh[122]true -+ exgfs_wave_post_gridded_sbs.sh[123]export err=0 -+ exgfs_wave_post_gridded_sbs.sh[123]err=0 -+ exgfs_wave_post_gridded_sbs.sh[124][[ 0 -ne 0 ]] -+ exgfs_wave_post_gridded_sbs.sh[130]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ exgfs_wave_post_gridded_sbs.sh[131]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ exgfs_wave_post_gridded_sbs.sh[132]gribchk=gfs.wave.t12z.gsouth.0p25.f102.grib2 -+ exgfs_wave_post_gridded_sbs.sh[133][[ ! -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f102.grib2 ]] -+ exgfs_wave_post_gridded_sbs.sh[138]exit 0 -+ JGLOBAL_WAVE_POST_SBS[28]true -+ JGLOBAL_WAVE_POST_SBS[29]export err=0 -+ JGLOBAL_WAVE_POST_SBS[29]err=0 -+ JGLOBAL_WAVE_POST_SBS[30][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[37]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312 -+ JGLOBAL_WAVE_POST_SBS[38][[ NO != \Y\E\S ]] -+ JGLOBAL_WAVE_POST_SBS[39]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f102.306010 -+ JGLOBAL_WAVE_POST_SBS[42]exit 0 -+ JGLOBAL_WAVE_POST_SBS[1]postamble /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS 1753758267 0 -+ preamble.sh[62]set +x -End /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at 03:04:50 with error code 0 (time elapsed: 00:00:23) -Begin /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at Tue Jul 29 03:04:50 UTC 2025 -++ jjob_header.sh[46]OPTIND=1 -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[50]env_job=wavepostsbs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[49]read -ra configs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[61]shift 4 -++ jjob_header.sh[63][[ -z wavepostsbs ]] -++ jjob_header.sh[71]export DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010 -++ jjob_header.sh[71]DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010 -++ jjob_header.sh[72][[ YES == \Y\E\S ]] -++ jjob_header.sh[73]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010 -++ jjob_header.sh[75]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010 -++ jjob_header.sh[76]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010 -++ jjob_header.sh[85]export pid=309696 -++ jjob_header.sh[85]pid=309696 -++ jjob_header.sh[86]export pgmout=OUTPUT.309696 -++ jjob_header.sh[86]pgmout=OUTPUT.309696 -++ jjob_header.sh[87]export pgmerr=errfile -++ jjob_header.sh[87]pgmerr=errfile -++ jjob_header.sh[90]export pgm= -++ jjob_header.sh[90]pgm= -++ jjob_header.sh[96]export cycle=t12z -++ jjob_header.sh[96]cycle=t12z -++ jjob_header.sh[97]setpdy.sh -+ setpdy.sh[20]'[' /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010 == /home/mterry ']' -+ setpdy.sh[25][[ ! t12z =~ t??z ]] -+ setpdy.sh[30]case $# in -+ setpdy.sh[31]dates_before_PDY=7 -+ setpdy.sh[32]dates_after_PDY=7 -+ setpdy.sh[50]COMDATEROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+ setpdy.sh[53]'[' -z 20210323 ']' -+ setpdy.sh[57]sed 's/[0-9]\{8\}/20210323/' /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z -sed: can't read /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z: No such file or directory -++ jjob_header.sh[97]true -++ jjob_header.sh[98]source ./PDY -/work2/noaa/global/mterry/global-workflow_forked/ush/jjob_header.sh: line 98: ./PDY: No such file or directory -++ jjob_header.sh[98]true -++ jjob_header.sh[104]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[104]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.base -+++ config.base[6]echo 'BEGIN: config.base' -BEGIN: config.base -+++ config.base[9]export machine=HERCULES -+++ config.base[9]machine=HERCULES -+++ config.base[12]export RUN_ENVIR=emc -+++ config.base[12]RUN_ENVIR=emc -+++ config.base[15]export ACCOUNT=fv3-cpu -+++ config.base[15]ACCOUNT=fv3-cpu -+++ config.base[16]export QUEUE=batch -+++ config.base[16]QUEUE=batch -+++ config.base[17]export QUEUE_SERVICE=batch -+++ config.base[17]QUEUE_SERVICE=batch -+++ config.base[18]export QUEUE_DTN=batch -+++ config.base[18]QUEUE_DTN=batch -+++ config.base[19]export PARTITION_BATCH=hercules -+++ config.base[19]PARTITION_BATCH=hercules -+++ config.base[20]export PARTITION_SERVICE=service -+++ config.base[20]PARTITION_SERVICE=service -+++ config.base[21]export PARTITION_DTN= -+++ config.base[21]PARTITION_DTN= -+++ config.base[22]export RESERVATION= -+++ config.base[22]RESERVATION= -+++ config.base[23]export CLUSTERS= -+++ config.base[23]CLUSTERS= -+++ config.base[24]export CLUSTERS_SERVICE= -+++ config.base[24]CLUSTERS_SERVICE= -+++ config.base[25]export CLUSTERS_DTN= -+++ config.base[25]CLUSTERS_DTN= -+++ config.base[28]export HPSS_PROJECT=emc-global -+++ config.base[28]HPSS_PROJECT=emc-global -+++ config.base[31]export HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[31]HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[32]export EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[32]EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[33]export FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[33]FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[34]export PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[34]PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[35]export SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[35]SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[36]export USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[36]USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[38]export FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[38]FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[39]export FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[39]FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[40]export FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[40]FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[41]export FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[41]FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[42]export FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[42]FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[43]export FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[43]FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[44]export FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[44]FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[45]export FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[45]FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[50]export PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[50]PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[51]export COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[51]COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[52]export COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[52]COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[53]export DMPDIR=/work/noaa/rstprod/dump -+++ config.base[53]DMPDIR=/work/noaa/rstprod/dump -+++ config.base[57]export COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[57]COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[58]export COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[58]COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[59]export COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[59]COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[62]export HOMEDIR=/work2/noaa/global/mterry -+++ config.base[62]HOMEDIR=/work2/noaa/global/mterry -+++ config.base[63]export STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[63]STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]export PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[65]export NOSCRUB=/work2/noaa/global/mterry -+++ config.base[65]NOSCRUB=/work2/noaa/global/mterry -+++ config.base[68]export BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[68]BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[71]export BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[71]BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[74]export DO_PREP_SFC=NO -+++ config.base[74]DO_PREP_SFC=NO -+++ config.base[77]export DO_GOES=NO -+++ config.base[77]DO_GOES=NO -+++ config.base[78]export DO_BUFRSND=NO -+++ config.base[78]DO_BUFRSND=NO -+++ config.base[79]export DO_GEMPAK=NO -+++ config.base[79]DO_GEMPAK=NO -+++ config.base[80]export DO_AWIPS=NO -+++ config.base[80]DO_AWIPS=NO -+++ config.base[81]export DO_NPOESS=NO -+++ config.base[81]DO_NPOESS=NO -+++ config.base[82]export DO_TRACKER=YES -+++ config.base[82]DO_TRACKER=YES -+++ config.base[83]export DO_GENESIS=YES -+++ config.base[83]DO_GENESIS=YES -+++ config.base[84]export DO_GENESIS_FSU=NO -+++ config.base[84]DO_GENESIS_FSU=NO -+++ config.base[85]export DO_VERFOZN=YES -+++ config.base[85]DO_VERFOZN=YES -+++ config.base[86]export DO_VERFRAD=YES -+++ config.base[86]DO_VERFRAD=YES -+++ config.base[87]export DO_VMINMON=YES -+++ config.base[87]DO_VMINMON=YES -+++ config.base[88]export DO_ANLSTAT=NO -+++ config.base[88]DO_ANLSTAT=NO -+++ config.base[91]export MODE=forecast-only -+++ config.base[91]MODE=forecast-only -+++ config.base[92]export DO_TEST_MODE=YES -+++ config.base[92]DO_TEST_MODE=YES -+++ config.base[101]export FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[101]FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[102]export HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[102]HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[103]export HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[103]HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[106]export NMV=/bin/mv -+++ config.base[106]NMV=/bin/mv -+++ config.base[107]export 'NLN=/bin/ln -sf' -+++ config.base[107]NLN='/bin/ln -sf' -+++ config.base[108]export VERBOSE=YES -+++ config.base[108]VERBOSE=YES -+++ config.base[109]export KEEPDATA=NO -+++ config.base[109]KEEPDATA=NO -+++ config.base[110]export DEBUG_POSTSCRIPT=NO -+++ config.base[110]DEBUG_POSTSCRIPT=NO -+++ config.base[111]export CHGRP_RSTPROD=YES -+++ config.base[111]CHGRP_RSTPROD=YES -+++ config.base[112]export 'CHGRP_CMD=chgrp rstprod' -+++ config.base[112]CHGRP_CMD='chgrp rstprod' -+++ config.base[113]export NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[113]NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[114]export NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[114]NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[117]export BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[117]BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[120]export SDATE=2021032312 -+++ config.base[120]SDATE=2021032312 -+++ config.base[121]export EDATE=2021032312 -+++ config.base[121]EDATE=2021032312 -+++ config.base[122]export EXP_WARM_START=.false. -+++ config.base[122]EXP_WARM_START=.false. -+++ config.base[123]export assim_freq=6 -+++ config.base[123]assim_freq=6 -+++ config.base[124]export PSLOT=C48_S2SW -+++ config.base[124]PSLOT=C48_S2SW -+++ config.base[125]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[125]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[126]export ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[126]ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[127]export DUMP_SUFFIX= -+++ config.base[127]DUMP_SUFFIX= -+++ config.base[128][[ 2021032312 -ge 2019092100 ]] -+++ config.base[128][[ 2021032312 -le 2019110700 ]] -+++ config.base[131]export ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[131]ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[132]export ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[132]ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[133]export FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[133]FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[136]export envir=prod -+++ config.base[136]envir=prod -+++ config.base[137]export NET=gfs -+++ config.base[137]NET=gfs -+++ config.base[138]export RUN=gfs -+++ config.base[138]RUN=gfs -+++ config.base[141]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.com -++++ config.com[4]echo 'BEGIN: config.com' -BEGIN: config.com -++++ config.com[38][[ emc == \n\c\o ]] -++++ config.com[43]COM_OBSPROC_TMPL='${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos' -++++ config.com[44]COM_RTOFS_TMPL='${DMPDIR}' -++++ config.com[45]COM_TCVITAL_TMPL='${DMPDIR}/${RUN}.${YMD}/${HH}/atmos' -++++ config.com[47]declare -rx 'COM_OBS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/obs' -++++ config.com[48]declare -rx COM_OBSPROC_TMPL COM_RTOFS_TMPL -++++ config.com[50]COM_BASE='${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}' -++++ config.com[52]declare -rx 'COM_TOP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}' -++++ config.com[54]declare -rx 'COM_CONF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf' -++++ config.com[55]declare -rx 'COM_OBS_JEDI=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi' -++++ config.com[57]declare -rx 'COM_ATMOS_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input' -++++ config.com[58]declare -rx 'COM_ATMOS_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart' -++++ config.com[59]declare -rx 'COM_ATMOS_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos' -++++ config.com[60]declare -rx 'COM_SNOW_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow' -++++ config.com[61]declare -rx 'COM_SNOW_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon' -++++ config.com[62]declare -rx 'COM_ATMOS_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history' -++++ config.com[63]declare -rx 'COM_ATMOS_MASTER_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master' -++++ config.com[64]declare -rx 'COM_ATMOS_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2' -++++ config.com[65]declare -rx 'COM_ATMOS_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}' -++++ config.com[66]declare -rx 'COM_ATMOS_BUFR_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr' -++++ config.com[67]declare -rx 'COM_ATMOS_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}' -++++ config.com[68]declare -rx 'COM_ATMOS_GENESIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital' -++++ config.com[69]declare -rx 'COM_ATMOS_TRACK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks' -++++ config.com[70]declare -rx 'COM_ATMOS_GOES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim' -++++ config.com[71]declare -rx 'COM_ATMOS_IMAGERY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery' -++++ config.com[72]declare -rx 'COM_ATMOS_OZNMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon' -++++ config.com[73]declare -rx 'COM_ATMOS_RADMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon' -++++ config.com[74]declare -rx 'COM_ATMOS_MINMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon' -++++ config.com[75]declare -rx 'COM_ATMOS_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon' -++++ config.com[76]declare -rx 'COM_ATMOS_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo' -++++ config.com[78]declare -rx 'COM_WAVE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart' -++++ config.com[79]declare -rx 'COM_WAVE_PREP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep' -++++ config.com[80]declare -rx 'COM_WAVE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history' -++++ config.com[81]declare -rx 'COM_WAVE_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded' -++++ config.com[82]declare -rx 'COM_WAVE_GRID_RES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}' -++++ config.com[83]declare -rx 'COM_WAVE_STATION_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station' -++++ config.com[84]declare -rx 'COM_WAVE_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak' -++++ config.com[85]declare -rx 'COM_WAVE_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo' -++++ config.com[87]declare -rx 'COM_OCEAN_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history' -++++ config.com[88]declare -rx 'COM_OCEAN_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart' -++++ config.com[89]declare -rx 'COM_OCEAN_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input' -++++ config.com[90]declare -rx 'COM_OCEAN_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean' -++++ config.com[91]declare -rx 'COM_OCEAN_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon' -++++ config.com[92]declare -rx 'COM_OCEAN_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf' -++++ config.com[93]declare -rx 'COM_OCEAN_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean' -++++ config.com[94]declare -rx 'COM_OCEAN_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf' -++++ config.com[95]declare -rx 'COM_OCEAN_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2' -++++ config.com[96]declare -rx 'COM_OCEAN_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}' -++++ config.com[98]declare -rx 'COM_ICE_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice' -++++ config.com[99]declare -rx 'COM_ICE_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf' -++++ config.com[100]declare -rx 'COM_ICE_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon' -++++ config.com[101]declare -rx 'COM_ICE_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice' -++++ config.com[102]declare -rx 'COM_ICE_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input' -++++ config.com[103]declare -rx 'COM_ICE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history' -++++ config.com[104]declare -rx 'COM_ICE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart' -++++ config.com[105]declare -rx 'COM_ICE_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf' -++++ config.com[106]declare -rx 'COM_ICE_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2' -++++ config.com[107]declare -rx 'COM_ICE_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}' -++++ config.com[109]declare -rx 'COM_CHEM_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history' -++++ config.com[110]declare -rx 'COM_CHEM_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem' -++++ config.com[111]declare -rx 'COM_CHEM_BMAT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix' -++++ config.com[112]declare -rx 'COM_CHEM_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon' -++++ config.com[114]declare -rx 'COM_MED_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart' -+++ config.base[143]export LOGSCRIPT= -+++ config.base[143]LOGSCRIPT= -+++ config.base[145]export 'REDOUT=1>' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wave -+++ config.wave[6]echo 'BEGIN: config.wave' -BEGIN: config.wave -+++ config.wave[13]export RUNRSTwave=gdas -+++ config.wave[13]RUNRSTwave=gdas -+++ config.wave[16]export MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[16]MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[19]case "${waveGRD}" in -+++ config.wave[64]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[64]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[65]export wavepostGRD= -+++ config.wave[65]wavepostGRD= -+++ config.wave[66]export waveuoutpGRD=uglo_100km -+++ config.wave[66]waveuoutpGRD=uglo_100km -+++ config.wave[75]export WAVEWND_DID= -+++ config.wave[75]WAVEWND_DID= -+++ config.wave[76]export WAVEWND_FID= -+++ config.wave[76]WAVEWND_FID= -+++ config.wave[79][[ gfs == \g\f\s ]] -+++ config.wave[80]export FHMAX_WAV=120 -+++ config.wave[80]FHMAX_WAV=120 -+++ config.wave[82]export WAVHINDH=0 -+++ config.wave[82]WAVHINDH=0 -+++ config.wave[83]export FHMAX_WAV_IBP=180 -+++ config.wave[83]FHMAX_WAV_IBP=180 -+++ config.wave[84](( FHMAX_WAV < FHMAX_WAV_IBP )) -+++ config.wave[84]export FHMAX_WAV_IBP=120 -+++ config.wave[84]FHMAX_WAV_IBP=120 -+++ config.wave[87]export DTFLD_WAV=3600 -+++ config.wave[87]DTFLD_WAV=3600 -+++ config.wave[88]export DTPNT_WAV=3600 -+++ config.wave[88]DTPNT_WAV=3600 -+++ config.wave[89]export FHINCP_WAV=1 -+++ config.wave[89]FHINCP_WAV=1 -+++ config.wave[92]export 'OUTPARS_WAV=WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[92]OUTPARS_WAV='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[95][[ gfs == \g\d\a\s ]] -+++ config.wave[99](( INTERVAL_GFS > 0 )) -+++ config.wave[100]export WAVHCYC=6 -+++ config.wave[100]WAVHCYC=6 -+++ config.wave[101]export FHMAX_WAV_CUR=192 -+++ config.wave[101]FHMAX_WAV_CUR=192 -+++ config.wave[109]export RSTTYPE_WAV=T -+++ config.wave[109]RSTTYPE_WAV=T -+++ config.wave[110][[ gfs != gfs ]] -+++ config.wave[115]rst_dt_gfs=43200 -+++ config.wave[116][[ 43200 -gt 0 ]] -+++ config.wave[117]export DT_1_RST_WAV=0 -+++ config.wave[117]DT_1_RST_WAV=0 -+++ config.wave[120]export DT_2_RST_WAV=43200 -+++ config.wave[120]DT_2_RST_WAV=43200 -+++ config.wave[126]export RSTIOFF_WAV=0 -+++ config.wave[126]RSTIOFF_WAV=0 -+++ config.wave[131]export RUNMEM=-1 -+++ config.wave[131]RUNMEM=-1 -+++ config.wave[134](( RUNMEM == -1 )) -+++ config.wave[136]export waveMEMB= -+++ config.wave[136]waveMEMB= -+++ config.wave[143]export WW3ATMINP=CPL -+++ config.wave[143]WW3ATMINP=CPL -+++ config.wave[144][[ YES == \Y\E\S ]] -+++ config.wave[145]export WW3ICEINP=CPL -+++ config.wave[145]WW3ICEINP=CPL -+++ config.wave[146]export WAVEICE_FID= -+++ config.wave[146]WAVEICE_FID= -+++ config.wave[152][[ YES == \Y\E\S ]] -+++ config.wave[153]export WW3CURINP=CPL -+++ config.wave[153]WW3CURINP=CPL -+++ config.wave[154]export WAVECUR_FID= -+++ config.wave[154]WAVECUR_FID= -+++ config.wave[161]export WW3ATMIENS=F -+++ config.wave[161]WW3ATMIENS=F -+++ config.wave[162]export WW3ICEIENS=F -+++ config.wave[162]WW3ICEIENS=F -+++ config.wave[163]export WW3CURIENS=F -+++ config.wave[163]WW3CURIENS=F -+++ config.wave[165]export GOFILETYPE=1 -+++ config.wave[165]GOFILETYPE=1 -+++ config.wave[166]export POFILETYPE=1 -+++ config.wave[166]POFILETYPE=1 -+++ config.wave[170]export FUNIPNT=T -+++ config.wave[170]FUNIPNT=T -+++ config.wave[172]export IOSRV=1 -+++ config.wave[172]IOSRV=1 -+++ config.wave[174]export FPNTPROC=T -+++ config.wave[174]FPNTPROC=T -+++ config.wave[176]export FGRDPROC=F -+++ config.wave[176]FGRDPROC=F -+++ config.wave[178]export FLAGMASKCOMP=F -+++ config.wave[178]FLAGMASKCOMP=F -+++ config.wave[180]export FLAGMASKOUT=F -+++ config.wave[180]FLAGMASKOUT=F -+++ config.wave[182]echo 'END: config.wave' -END: config.wave -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wavepostsbs -+++ config.wavepostsbs[6]echo 'BEGIN: config.wavepostsbs' -BEGIN: config.wavepostsbs -+++ config.wavepostsbs[9]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources wavepostsbs -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=wavepostsbs -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[156]ntasks=8 -++++ config.resources[157]threads_per_task=1 -++++ config.resources[158]tasks_per_node=80 -++++ config.resources[159]NTASKS=8 -++++ config.resources[160]memory=20GB -++++ config.resources[161]walltime=00:15:00 -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export NTASKS -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export memory -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.wavepostsbs[12]export MAX_TASKS=25 -+++ config.wavepostsbs[12]MAX_TASKS=25 -+++ config.wavepostsbs[15]export WAV_SUBGRBSRC= -+++ config.wavepostsbs[15]WAV_SUBGRBSRC= -+++ config.wavepostsbs[16]export WAV_SUBGRB= -+++ config.wavepostsbs[16]WAV_SUBGRB= -+++ config.wavepostsbs[19]export DOFLD_WAV=YES -+++ config.wavepostsbs[19]DOFLD_WAV=YES -+++ config.wavepostsbs[20]export DOPNT_WAV=YES -+++ config.wavepostsbs[20]DOPNT_WAV=YES -+++ config.wavepostsbs[21]export DOGRB_WAV=YES -+++ config.wavepostsbs[21]DOGRB_WAV=YES -+++ config.wavepostsbs[22][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+++ config.wavepostsbs[23]export DOGRI_WAV=YES -+++ config.wavepostsbs[23]DOGRI_WAV=YES -+++ config.wavepostsbs[27]export DOSPC_WAV=YES -+++ config.wavepostsbs[27]DOSPC_WAV=YES -+++ config.wavepostsbs[28]export DOBLL_WAV=YES -+++ config.wavepostsbs[28]DOBLL_WAV=YES -+++ config.wavepostsbs[30]echo 'END: config.wavepostsbs' -END: config.wavepostsbs -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env wavepostsbs -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=wavepostsbs -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 20971520 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 8 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[34]max_threads_per_task=1 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 1 ]] -+++ HERCULES.env[40][[ 1 -gt 1 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 8' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[63]export USE_CFP=YES -+++ HERCULES.env[63]USE_CFP=YES -+++ HERCULES.env[64][[ wavepostsbs == \w\a\v\e\p\r\e\p ]] -+++ HERCULES.env[65]export 'wavempexec=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[65]wavempexec='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[66]export 'wave_mpmd=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[66]wave_mpmd='--multi-prog --output=mpmd.%j.%t.out' -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[4]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ JGLOBAL_WAVE_POST_SBS[7]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[7]HH=12 -+ JGLOBAL_WAVE_POST_SBS[7]declare_from_tmpl -rx COMIN_WAVE_PREP:COM_WAVE_PREP_TMPL COMIN_WAVE_HISTORY:COM_WAVE_HISTORY_TMPL COMOUT_WAVE_GRID:COM_WAVE_GRID_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_WAVE_PREP=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep -declare_from_tmpl :: COMIN_WAVE_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history -declare_from_tmpl :: COMOUT_WAVE_GRID=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded -+ JGLOBAL_WAVE_POST_SBS[12][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n '' ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_atlocn_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_atlocn_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_atlocn_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_epacif_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_epacif_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_epacif_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_wcoast_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_wcoast_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_wcoast_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p50 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p50:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p50=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p50 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=arctic.9km -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_arctic_9km:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_arctic_9km=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_arctic_9km -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_gsouth_0p25:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_gsouth_0p25=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 ]] -+ JGLOBAL_WAVE_POST_SBS[28]/work2/noaa/global/mterry/global-workflow_forked/scripts/exgfs_wave_post_gridded_sbs.sh -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ exgfs_wave_post_gridded_sbs.sh[24]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ exgfs_wave_post_gridded_sbs.sh[26]DOGRI_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[27]DOGRB_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[29]export waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[29]waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[30]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[30]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[31]export wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[31]wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[33]cat - INFO: Grid information: - INFO: Native wave grids: uglo_100km - INFO: Interpolated grids: at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m - INFO: Post-process grids: -++ exgfs_wave_post_gridded_sbs.sh[40]printf %03i 105 -+ exgfs_wave_post_gridded_sbs.sh[40]fhr3=105 -++ exgfs_wave_post_gridded_sbs.sh[41]date -u -d '20210323 12 + 105 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[41]valid_time=2021032721 -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[49]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f105.bin ./out_grd.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f105.bin ./out_grd.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[52][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[62]cat - INFO: Summary: - INFO: Grid interp: DOGRI_WAV="YES" - INFO: Grib files: DOGRB_WAV="YES" - INFO: Fields to be included in grib files: - INFO: OUTPARS_WAV="WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA" -+ exgfs_wave_post_gridded_sbs.sh[70][[ YES == \N\O ]] -+ exgfs_wave_post_gridded_sbs.sh[76]rm -f 'cmdfile.*' cmdfile -+ exgfs_wave_post_gridded_sbs.sh[77]count=0 -+ exgfs_wave_post_gridded_sbs.sh[80][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[81]dt_int=3600. -+ exgfs_wave_post_gridded_sbs.sh[82]n_int=9999 -++ exgfs_wave_post_gridded_sbs.sh[83]date -u -d '20210327 21 - 0 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[83]ymdh_int=2021032721 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=1 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032721 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032721 105 atlocn 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib2_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib2_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=2 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032721 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032721 105 epacif 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib2_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib2_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=3 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032721 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032721 105 wcoast 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib2_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib2_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=4 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032721 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032721 105 global 0p50 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib2_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib2_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=5 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032721 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032721 105 arctic 9km '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib2_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib2_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=6 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032721 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032721 105 global 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib2_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib2_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=7 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032721 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032721 105 gsouth 0p25 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib2_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib2_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[100][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[113][[ 8 -lt 7 ]] -+ exgfs_wave_post_gridded_sbs.sh[121]echo 'INFO: Running MPMD job with 7 commands' -INFO: Running MPMD job with 7 commands -+ exgfs_wave_post_gridded_sbs.sh[122]/work2/noaa/global/mterry/global-workflow_forked/ush/run_mpmd.sh /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/cmdfile -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ run_mpmd.sh[31]source /work2/noaa/global/mterry/global-workflow_forked/ush/preamble.sh -++ preamble.sh[20]set +x -Begin run_mpmd.sh at Tue Jul 29 03:05:02 UTC 2025 -+ run_mpmd.sh[33]cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/cmdfile -+ run_mpmd.sh[36][[ YES != \Y\E\S ]] -+ run_mpmd.sh[46]export OMP_NUM_THREADS=1 -+ run_mpmd.sh[46]OMP_NUM_THREADS=1 -++ run_mpmd.sh[49]wc -l -+ run_mpmd.sh[49]nprocs=7 -+ run_mpmd.sh[52]mpmd_cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mpmd_cmdfile -+ run_mpmd.sh[53][[ -s /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mpmd_cmdfile ]] -+ run_mpmd.sh[55]cat - INFO: Executing MPMD job, STDOUT redirected for each process separately - INFO: On failure, logs for each job will be available in /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mpmd.proc_num.out - INFO: The proc_num corresponds to the line in '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mpmd_cmdfile' -+ run_mpmd.sh[61][[ srun -l --export=ALL --hint=nomultithread =~ ^srun.* ]] -+ run_mpmd.sh[65]nm=0 -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '0 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/cmdfile.1' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '1 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/cmdfile.2' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '2 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/cmdfile.3' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '3 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/cmdfile.4' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '4 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/cmdfile.5' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '5 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/cmdfile.6' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '6 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/cmdfile.7' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[72]set +e -+ run_mpmd.sh[74]srun -l --export=ALL --hint=nomultithread --multi-prog --output=mpmd.%j.%t.out -n 7 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mpmd_cmdfile -+ run_mpmd.sh[75]err=0 -+ run_mpmd.sh[76]set_strict -+ preamble.sh[35][[ YES == \Y\E\S ]] -+ preamble.sh[37]set -eu -+ run_mpmd.sh[103][[ 0 -eq 0 ]] -+ run_mpmd.sh[104]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mpmd_cmdfile -++ run_mpmd.sh[105]find . -name 'mpmd.*.out' -+ run_mpmd.sh[105]out_files='./mpmd.5951799.1.out -./mpmd.5951799.6.out -./mpmd.5951799.0.out -./mpmd.5951799.5.out -./mpmd.5951799.3.out -./mpmd.5951799.4.out -./mpmd.5951799.2.out' -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951799.1.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951799.1.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951799.6.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951799.6.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951799.0.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951799.0.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951799.5.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951799.5.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951799.3.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951799.3.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951799.4.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951799.4.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951799.2.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951799.2.out -+ run_mpmd.sh[110]cat mpmd.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + cmdfile.2[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032721 3600. 9999 -1: + cmdfile.2[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grid_interp_sbs.sh[25]grdID=ep_10m -1: + wave_grid_interp_sbs.sh[26]valid_time=2021032721 -1: + wave_grid_interp_sbs.sh[27]dt=3600. -1: + wave_grid_interp_sbs.sh[28]nst=9999 -1: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010 -1: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.uglo_100km ./out_grd.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mod_def.uglo_100km ./mod_def.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mod_def.ep_10m ./mod_def.ep_10m -1: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ]] -1: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m'\''' -1: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m' -1: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[51]weights_found=1 -1: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 210000' -1: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 210000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/ep_10m/g ww3_gint.inp.tmpl -1: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -1: $ Input file for interpolation of uglo_100km to ep_10m -1: $------------------------------------------------ -1: $ Start Time 3600. NSteps -1: 20210327 210000 3600. 9999 -1: $ Total number of grids -1: 2 -1: $ Grid extensions -1: 'uglo_100km' -1: 'ep_10m' -1: $ -1: 0 -1: $ -1: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[70]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.309696 ']' -1: ++ prep_step[4]echo gfs_ww3_gint.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -1: INFO: Executing 'gfs_ww3_gint.x' -1: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[73]cat grid_interp.ep_10m.out -1: -1: *** WAVEWATCH III Grid interpolation *** -1: =============================================== -1: -1: Comment character is '$' -1: -1: Time Information : -1: --------------------------------------------- -1: Starting Time : 2021/03/27 21:00:00 UTC -1: Interval (in sec) : 3600.00 -1: Number of requests : 9999 -1: --------------------------------------------- -1: Number of grids (including output grid) = 2 -1: -1: -1: Extension for grid 1 is --> uglo_100km -1: -1: Grid Particulars are : -1: Dimensions = 45166 1 -1: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Extension for grid 2 is --> ep_10m -1: -1: Grid Particulars are : -1: Dimensions = 511 301 -1: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -1: -1: -1: Preparing interpolation weights for output grid -1: Total number of wet points for interpolation 7439 -1: -1: -1: Variable: Grid Interpolation Map Units: 0.100E+01 -1: -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 301 | | -1: | | -1: | 0 0 | -1: 262 | 0 0 0 0 | -1: | 0 0 0 0 | -1: | 0 0 0 0 | -1: 223 | 0 0 0 | -1: | 0 | -1: | 0 | -1: 184 | | -1: | 0 0 0 | -1: | 0 0 0 0 | -1: 145 | 0 | -1: | | -1: | | -1: 106 | | -1: | | -1: | | -1: 67 | | -1: | 0 | -1: | | -1: 28 | 0 | -1: | | -1: | | -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: -1: -1: Interpolating fields .... -1: -1: Output group 1 -1: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -1: Output group 2 -1: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -1: Output group 3 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 4 -1: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -1: Output group 5 -1: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -1: Output group 6 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 7 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 8 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 9 -1: Output variables skipped -1: Output group 10 -1: Output variables skipped -1: ------------------------------------------------ -1: 1Current vel. -1: 1Wind speed -1: 1Ice concentration -1: 2Wave height -1: 2Mean wave period(+2) -1: 2Mean wave period(+1) -1: 2Peak frequency -1: 2Mean wave dir. a1b1 -1: 2Peak direction -1: 4Part. wave height -1: 4Part. peak period -1: 4Part. mean direction -1: 5Charnock parameter -1: ------------------------------------------------ -1: OUTPUT TIME : 2021/03/27 21:00:00 UTC -1: -1: End of file reached -1: -1: -1: *** End of Grid interpolation Routine *** -1: =============================================== -1: -1: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -1: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -1: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_ep_10m/out_grd.ep_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.ep_10m -1: + cmdfile.2[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032721 105 epacif 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + cmdfile.2[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib2_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grib2_sbs.sh[30]grdID=ep_10m -1: + wave_grib2_sbs.sh[31]GRIDNR=255 -1: + wave_grib2_sbs.sh[32]MODNR=11 -1: + wave_grib2_sbs.sh[33]valid_time=2021032721 -1: + wave_grib2_sbs.sh[34]fhr=105 -1: + wave_grib2_sbs.sh[35]grid_region=epacif -1: + wave_grib2_sbs.sh[36]grid_res=0p16 -1: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010 -1: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_ep_10m -1: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_ep_10m -1: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_ep_10m -1: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_ep_10m -1: ++ wave_grib2_sbs.sh[47]printf %03i 105 -1: + wave_grib2_sbs.sh[47]FH3=105 -1: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_epacif_0p16 -1: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.epacif.0p16.f105.grib2 -1: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f105.grib2 ]] -1: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mod_def.ep_10m ./mod_def.ww3 -1: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.ep_10m ./out_grd.ww3 -1: + wave_grib2_sbs.sh[73]ngrib=1 -1: + wave_grib2_sbs.sh[74]dtgrib=3600 -1: + wave_grib2_sbs.sh[75]tstart='20210327 210000' -1: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 210000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.ep_10m.inp.tmpl -1: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -1: $ WAVEWATCH-III gridded output input file -1: $ ---------------------------------------- -1: 20210327 210000 3600 1 -1: N -1: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -1: $ -1: 20210327 210000 7 11 255 0 0 -1: $ -1: $ end of input file -1: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[88]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.309696 ']' -1: ++ prep_step[4]echo gfs_ww3_grib.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[90]export err=0 -1: + wave_grib2_sbs.sh[90]err=0 -1: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[95]cat grib2_epacif_105.out -1: -1: *** WAVEWATCH III GRIB output postp. *** -1: ============================================== -1: -1: Comment character is '$' -1: -1: Grid name : East Pacific 10 min wave grid -1: -1: LINEIN: -1: 20210327 210000 3600 1 -1: -1: 20210327210000 3600 1 -1: GEN_PRO -99999 -1: -1: Output time data : -1: ----------------------------------------------------- -1: First time : 2021/03/27 21:00:00 UTC -1: Interval : 01:00:00 -1: Number of requests : 1 -1: Fields : Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: Requested output fields not yet available: -1: ----------------------------------------------------- -1: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -1: -1: Successfully requested output fields : -1: ----------------------------------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: -1: Additional GRIB parameters : -1: ----------------------------------------------------- -1: Run time : 2021/03/27 21:00:00 UTC -1: GRIB center ID : 7 -1: GRIB gen. proc. ID : 11 -1: GRIB grid ID : 255 -1: GRIB GDS parameter : 0 -1: Fields in file : -1: -------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: CHOSEN GRID TYPE: : LLRECTILINEAR -1: -1: -1: -1: Generating file -1: ----------------------------------------------------- -1: Data for 2021/03/27 21:00:00 UTC 0H forecast. -1: Warning: bitmask off everywhere. -1: Pretend one point in jpcpack to avoid crash. -1: Warning: bitmask off everywhere. -1: Pretend one point in jpcpack to avoid crash. -1: Warning: bitmask off everywhere. -1: Pretend one point in jpcpack to avoid crash. -1: -1: End of program -1: ========================================= -1: WAVEWATCH III GRIB output -1: -1: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -1: + wave_grib2_sbs.sh[102][[ 105 -gt 0 ]] -1: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '105 hour fcst' -grib gfs.wave.t12z.epacif.0p16.f105.grib2 -1: 1:0:d=2021032312:SPC:surface:105 hour fcst: -1: 2:21981:d=2021032312:DIRC:surface:105 hour fcst: -1: 3:50616:d=2021032312:UOGRD:surface:105 hour fcst: -1: 4:72430:d=2021032312:VOGRD:surface:105 hour fcst: -1: 5:94489:d=2021032312:WIND:surface:105 hour fcst: -1: 6:119003:d=2021032312:WDIR:surface:105 hour fcst: -1: 7:146093:d=2021032312:UGRD:surface:105 hour fcst: -1: 8:170675:d=2021032312:VGRD:surface:105 hour fcst: -1: 9:195099:d=2021032312:ICEC:surface:105 hour fcst: -1: 10:214507:d=2021032312:HTSGW:surface:105 hour fcst: -1: 11:236854:d=2021032312:IMWF:surface:105 hour fcst: -1: 12:259851:d=2021032312:MWSPER:surface:105 hour fcst: -1: 13:282823:d=2021032312:PERPW:surface:105 hour fcst: -1: 14:305794:d=2021032312:WWSDIR:surface:105 hour fcst: -1: 15:331827:d=2021032312:DIRPW:surface:105 hour fcst: -1: 16:358111:d=2021032312:WVHGT:surface:105 hour fcst: -1: 17:380374:d=2021032312:SWELL:1 in sequence:105 hour fcst: -1: 18:403360:d=2021032312:SWELL:2 in sequence:105 hour fcst: -1: 19:423930:d=2021032312:SWELL:3 in sequence:105 hour fcst: -1: 20:443338:d=2021032312:WVPER:surface:105 hour fcst: -1: 21:465550:d=2021032312:SWPER:1 in sequence:105 hour fcst: -1: 22:488706:d=2021032312:SWPER:2 in sequence:105 hour fcst: -1: 23:509170:d=2021032312:SWPER:3 in sequence:105 hour fcst: -1: 24:528578:d=2021032312:WVDIR:surface:105 hour fcst: -1: 25:553453:d=2021032312:SWDIR:1 in sequence:105 hour fcst: -1: 26:580748:d=2021032312:SWDIR:2 in sequence:105 hour fcst: -1: 27:602285:d=2021032312:SWDIR:3 in sequence:105 hour fcst: -1: + wave_grib2_sbs.sh[104]err=0 -1: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.epacif.0p16.f105.grib2 -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f105.grib2 ]] -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f105.grib2.idx ]] -1: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.epacif.0p16.f105.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f105.grib2 -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f105.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f105.grib2 = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f105.grib2 ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f105.grib2 -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f105.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f105.grib2.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f105.grib2.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f105.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f105.grib2 -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.epacif.0p16.f105.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f105.grib2.idx -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f105.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f105.grib2.idx = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f105.grib2.idx ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f105.grib2.idx -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f105.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f105.grib2.idx.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f105.grib2.idx.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f105.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f105.grib2.idx -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.epacif.0p16.f105.grib2 and gfs.wave.t12z.epacif.0p16.f105.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_ep_10m to COM' -1: INFO: Copied gfs.wave.t12z.epacif.0p16.f105.grib2 and gfs.wave.t12z.epacif.0p16.f105.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_ep_10m to COM -1: + wave_grib2_sbs.sh[130][[ ep_10m == '' ]] -1: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -1: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.epacif.0p16.f105.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -1: INFO: gfs.wave.t12z.epacif.0p16.f105.grib2 is global.0p50 or SENDDBN is NO, no alert sent -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + cmdfile.7[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032721 3600. 9999 -6: + cmdfile.7[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grid_interp_sbs.sh[25]grdID=gsh_15m -6: + wave_grid_interp_sbs.sh[26]valid_time=2021032721 -6: + wave_grid_interp_sbs.sh[27]dt=3600. -6: + wave_grid_interp_sbs.sh[28]nst=9999 -6: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010 -6: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.uglo_100km ./out_grd.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mod_def.uglo_100km ./mod_def.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mod_def.gsh_15m ./mod_def.gsh_15m -6: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m'\''' -6: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m' -6: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[51]weights_found=1 -6: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 210000' -6: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 210000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gsh_15m/g ww3_gint.inp.tmpl -6: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -6: $ Input file for interpolation of uglo_100km to gsh_15m -6: $------------------------------------------------ -6: $ Start Time 3600. NSteps -6: 20210327 210000 3600. 9999 -6: $ Total number of grids -6: 2 -6: $ Grid extensions -6: 'uglo_100km' -6: 'gsh_15m' -6: $ -6: 0 -6: $ -6: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[70]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.309696 ']' -6: ++ prep_step[4]echo gfs_ww3_gint.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -6: INFO: Executing 'gfs_ww3_gint.x' -6: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[73]cat grid_interp.gsh_15m.out -6: -6: *** WAVEWATCH III Grid interpolation *** -6: =============================================== -6: -6: Comment character is '$' -6: -6: Time Information : -6: --------------------------------------------- -6: Starting Time : 2021/03/27 21:00:00 UTC -6: Interval (in sec) : 3600.00 -6: Number of requests : 9999 -6: --------------------------------------------- -6: Number of grids (including output grid) = 2 -6: -6: -6: Extension for grid 1 is --> uglo_100km -6: -6: Grid Particulars are : -6: Dimensions = 45166 1 -6: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Extension for grid 2 is --> gsh_15m -6: -6: Grid Particulars are : -6: Dimensions = 1440 277 -6: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -6: -6: -6: Preparing interpolation weights for output grid -6: Total number of wet points for interpolation 317192 -6: -6: -6: Variable: Grid Interpolation Map Units: 0.100E+01 -6: -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 277 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 241 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 205 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 169 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 133 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 97 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 61 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 | -6: 25 | 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 | -6: | | -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: -6: -6: Interpolating fields .... -6: -6: Output group 1 -6: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -6: Output group 2 -6: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -6: Output group 3 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 4 -6: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -6: Output group 5 -6: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -6: Output group 6 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 7 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 8 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 9 -6: Output variables skipped -6: Output group 10 -6: Output variables skipped -6: ------------------------------------------------ -6: 1Current vel. -6: 1Wind speed -6: 1Ice concentration -6: 2Wave height -6: 2Mean wave period(+2) -6: 2Mean wave period(+1) -6: 2Peak frequency -6: 2Mean wave dir. a1b1 -6: 2Peak direction -6: 4Part. wave height -6: 4Part. peak period -6: 4Part. mean direction -6: 5Charnock parameter -6: ------------------------------------------------ -6: OUTPUT TIME : 2021/03/27 21:00:00 UTC -6: -6: End of file reached -6: -6: -6: *** End of Grid interpolation Routine *** -6: =============================================== -6: -6: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -6: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -6: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_gsh_15m/out_grd.gsh_15m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.gsh_15m -6: + cmdfile.7[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032721 105 gsouth 0p25 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + cmdfile.7[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib2_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grib2_sbs.sh[30]grdID=gsh_15m -6: + wave_grib2_sbs.sh[31]GRIDNR=255 -6: + wave_grib2_sbs.sh[32]MODNR=11 -6: + wave_grib2_sbs.sh[33]valid_time=2021032721 -6: + wave_grib2_sbs.sh[34]fhr=105 -6: + wave_grib2_sbs.sh[35]grid_region=gsouth -6: + wave_grib2_sbs.sh[36]grid_res=0p25 -6: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010 -6: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_gsh_15m -6: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_gsh_15m -6: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_gsh_15m -6: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_gsh_15m -6: ++ wave_grib2_sbs.sh[47]printf %03i 105 -6: + wave_grib2_sbs.sh[47]FH3=105 -6: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -6: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.gsouth.0p25.f105.grib2 -6: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f105.grib2 ]] -6: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mod_def.gsh_15m ./mod_def.ww3 -6: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.gsh_15m ./out_grd.ww3 -6: + wave_grib2_sbs.sh[73]ngrib=1 -6: + wave_grib2_sbs.sh[74]dtgrib=3600 -6: + wave_grib2_sbs.sh[75]tstart='20210327 210000' -6: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 210000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gsh_15m.inp.tmpl -6: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -6: $ WAVEWATCH-III gridded output input file -6: $ ---------------------------------------- -6: 20210327 210000 3600 1 -6: N -6: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -6: $ -6: 20210327 210000 7 11 255 0 0 -6: $ -6: $ end of input file -6: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[88]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.309696 ']' -6: ++ prep_step[4]echo gfs_ww3_grib.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[90]export err=0 -6: + wave_grib2_sbs.sh[90]err=0 -6: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[95]cat grib2_gsouth_105.out -6: -6: *** WAVEWATCH III GRIB output postp. *** -6: ============================================== -6: -6: Comment character is '$' -6: -6: Grid name : GFSv16-wave S Hemisphere 1/4 d -6: -6: LINEIN: -6: 20210327 210000 3600 1 -6: -6: 20210327210000 3600 1 -6: GEN_PRO -99999 -6: -6: Output time data : -6: ----------------------------------------------------- -6: First time : 2021/03/27 21:00:00 UTC -6: Interval : 01:00:00 -6: Number of requests : 1 -6: Fields : Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: Requested output fields not yet available: -6: ----------------------------------------------------- -6: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -6: -6: Successfully requested output fields : -6: ----------------------------------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: -6: Additional GRIB parameters : -6: ----------------------------------------------------- -6: Run time : 2021/03/27 21:00:00 UTC -6: GRIB center ID : 7 -6: GRIB gen. proc. ID : 11 -6: GRIB grid ID : 255 -6: GRIB GDS parameter : 0 -6: Fields in file : -6: -------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: CHOSEN GRID TYPE: : LLRECTILINEAR -6: -6: -6: -6: Generating file -6: ----------------------------------------------------- -6: Data for 2021/03/27 21:00:00 UTC 0H forecast. -6: -6: End of program -6: ========================================= -6: WAVEWATCH III GRIB output -6: -6: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -6: + wave_grib2_sbs.sh[102][[ 105 -gt 0 ]] -6: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '105 hour fcst' -grib gfs.wave.t12z.gsouth.0p25.f105.grib2 -6: 1:0:d=2021032312:SPC:surface:105 hour fcst: -6: 2:118992:d=2021032312:DIRC:surface:105 hour fcst: -6: 3:506839:d=2021032312:UOGRD:surface:105 hour fcst: -6: 4:621105:d=2021032312:VOGRD:surface:105 hour fcst: -6: 5:743057:d=2021032312:WIND:surface:105 hour fcst: -6: 6:986990:d=2021032312:WDIR:surface:105 hour fcst: -6: 7:1367361:d=2021032312:UGRD:surface:105 hour fcst: -6: 8:1605014:d=2021032312:VGRD:surface:105 hour fcst: -6: 9:1843764:d=2021032312:ICEC:surface:105 hour fcst: -6: 10:1904881:d=2021032312:HTSGW:surface:105 hour fcst: -6: 11:2059874:d=2021032312:IMWF:surface:105 hour fcst: -6: 12:2231947:d=2021032312:MWSPER:surface:105 hour fcst: -6: 13:2404687:d=2021032312:PERPW:surface:105 hour fcst: -6: 14:2588753:d=2021032312:WWSDIR:surface:105 hour fcst: -6: 15:2919907:d=2021032312:DIRPW:surface:105 hour fcst: -6: 16:3255695:d=2021032312:WVHGT:surface:105 hour fcst: -6: 17:3410413:d=2021032312:SWELL:1 in sequence:105 hour fcst: -6: 18:3584103:d=2021032312:SWELL:2 in sequence:105 hour fcst: -6: 19:3717173:d=2021032312:SWELL:3 in sequence:105 hour fcst: -6: 20:3799993:d=2021032312:WVPER:surface:105 hour fcst: -6: 21:3974363:d=2021032312:SWPER:1 in sequence:105 hour fcst: -6: 22:4170587:d=2021032312:SWPER:2 in sequence:105 hour fcst: -6: 23:4341791:d=2021032312:SWPER:3 in sequence:105 hour fcst: -6: 24:4444178:d=2021032312:WVDIR:surface:105 hour fcst: -6: 25:4714662:d=2021032312:SWDIR:1 in sequence:105 hour fcst: -6: 26:5065367:d=2021032312:SWDIR:2 in sequence:105 hour fcst: -6: 27:5355642:d=2021032312:SWDIR:3 in sequence:105 hour fcst: -6: + wave_grib2_sbs.sh[104]err=0 -6: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.gsouth.0p25.f105.grib2 -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f105.grib2 ]] -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f105.grib2.idx ]] -6: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.gsouth.0p25.f105.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f105.grib2 -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f105.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f105.grib2 = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f105.grib2 ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f105.grib2 -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f105.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f105.grib2.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f105.grib2.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f105.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f105.grib2 -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.gsouth.0p25.f105.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f105.grib2.idx -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f105.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f105.grib2.idx = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f105.grib2.idx ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f105.grib2.idx -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f105.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f105.grib2.idx.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f105.grib2.idx.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f105.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f105.grib2.idx -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.gsouth.0p25.f105.grib2 and gfs.wave.t12z.gsouth.0p25.f105.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_gsh_15m to COM' -6: INFO: Copied gfs.wave.t12z.gsouth.0p25.f105.grib2 and gfs.wave.t12z.gsouth.0p25.f105.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_gsh_15m to COM -6: + wave_grib2_sbs.sh[130][[ gsh_15m == '' ]] -6: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -6: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.gsouth.0p25.f105.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -6: INFO: gfs.wave.t12z.gsouth.0p25.f105.grib2 is global.0p50 or SENDDBN is NO, no alert sent -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + cmdfile.1[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032721 3600. 9999 -0: + cmdfile.1[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grid_interp_sbs.sh[25]grdID=at_10m -0: + wave_grid_interp_sbs.sh[26]valid_time=2021032721 -0: + wave_grid_interp_sbs.sh[27]dt=3600. -0: + wave_grid_interp_sbs.sh[28]nst=9999 -0: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010 -0: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.uglo_100km ./out_grd.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mod_def.uglo_100km ./mod_def.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mod_def.at_10m ./mod_def.at_10m -0: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ]] -0: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m'\''' -0: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m' -0: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[51]weights_found=1 -0: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 210000' -0: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 210000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/at_10m/g ww3_gint.inp.tmpl -0: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -0: $ Input file for interpolation of uglo_100km to at_10m -0: $------------------------------------------------ -0: $ Start Time 3600. NSteps -0: 20210327 210000 3600. 9999 -0: $ Total number of grids -0: 2 -0: $ Grid extensions -0: 'uglo_100km' -0: 'at_10m' -0: $ -0: 0 -0: $ -0: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[70]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.309696 ']' -0: ++ prep_step[4]echo gfs_ww3_gint.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -0: INFO: Executing 'gfs_ww3_gint.x' -0: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[73]cat grid_interp.at_10m.out -0: -0: *** WAVEWATCH III Grid interpolation *** -0: =============================================== -0: -0: Comment character is '$' -0: -0: Time Information : -0: --------------------------------------------- -0: Starting Time : 2021/03/27 21:00:00 UTC -0: Interval (in sec) : 3600.00 -0: Number of requests : 9999 -0: --------------------------------------------- -0: Number of grids (including output grid) = 2 -0: -0: -0: Extension for grid 1 is --> uglo_100km -0: -0: Grid Particulars are : -0: Dimensions = 45166 1 -0: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Extension for grid 2 is --> at_10m -0: -0: Grid Particulars are : -0: Dimensions = 301 331 -0: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -0: -0: -0: Preparing interpolation weights for output grid -0: Total number of wet points for interpolation 29591 -0: -0: -0: Variable: Grid Interpolation Map Units: 0.100E+01 -0: -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 331 | | -0: | | -0: | | -0: 289 | | -0: | | -0: | 0 0 0 | -0: 247 | 0 0 0 0 | -0: | 0 0 0 0 0 | -0: | 0 0 0 0 | -0: 205 | 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 163 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 121 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 79 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 | -0: | 0 0 | -0: 37 | 0 | -0: | | -0: | | -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: -0: -0: Interpolating fields .... -0: -0: Output group 1 -0: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -0: Output group 2 -0: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -0: Output group 3 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 4 -0: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -0: Output group 5 -0: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -0: Output group 6 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 7 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 8 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 9 -0: Output variables skipped -0: Output group 10 -0: Output variables skipped -0: ------------------------------------------------ -0: 1Current vel. -0: 1Wind speed -0: 1Ice concentration -0: 2Wave height -0: 2Mean wave period(+2) -0: 2Mean wave period(+1) -0: 2Peak frequency -0: 2Mean wave dir. a1b1 -0: 2Peak direction -0: 4Part. wave height -0: 4Part. peak period -0: 4Part. mean direction -0: 5Charnock parameter -0: ------------------------------------------------ -0: OUTPUT TIME : 2021/03/27 21:00:00 UTC -0: -0: End of file reached -0: -0: -0: *** End of Grid interpolation Routine *** -0: =============================================== -0: -0: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -0: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -0: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_at_10m/out_grd.at_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.at_10m -0: + cmdfile.1[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032721 105 atlocn 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + cmdfile.1[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib2_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grib2_sbs.sh[30]grdID=at_10m -0: + wave_grib2_sbs.sh[31]GRIDNR=255 -0: + wave_grib2_sbs.sh[32]MODNR=11 -0: + wave_grib2_sbs.sh[33]valid_time=2021032721 -0: + wave_grib2_sbs.sh[34]fhr=105 -0: + wave_grib2_sbs.sh[35]grid_region=atlocn -0: + wave_grib2_sbs.sh[36]grid_res=0p16 -0: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010 -0: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_at_10m -0: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_at_10m -0: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_at_10m -0: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_at_10m -0: ++ wave_grib2_sbs.sh[47]printf %03i 105 -0: + wave_grib2_sbs.sh[47]FH3=105 -0: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_atlocn_0p16 -0: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.atlocn.0p16.f105.grib2 -0: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f105.grib2 ]] -0: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mod_def.at_10m ./mod_def.ww3 -0: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.at_10m ./out_grd.ww3 -0: + wave_grib2_sbs.sh[73]ngrib=1 -0: + wave_grib2_sbs.sh[74]dtgrib=3600 -0: + wave_grib2_sbs.sh[75]tstart='20210327 210000' -0: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 210000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.at_10m.inp.tmpl -0: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -0: $ WAVEWATCH-III gridded output input file -0: $ ---------------------------------------- -0: 20210327 210000 3600 1 -0: N -0: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -0: $ -0: 20210327 210000 7 11 255 0 0 -0: $ -0: $ end of input file -0: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[88]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.309696 ']' -0: ++ prep_step[4]echo gfs_ww3_grib.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[90]export err=0 -0: + wave_grib2_sbs.sh[90]err=0 -0: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[95]cat grib2_atlocn_105.out -0: -0: *** WAVEWATCH III GRIB output postp. *** -0: ============================================== -0: -0: Comment character is '$' -0: -0: Grid name : NW Atlantic 10 min wave grid -0: -0: LINEIN: -0: 20210327 210000 3600 1 -0: -0: 20210327210000 3600 1 -0: GEN_PRO -99999 -0: -0: Output time data : -0: ----------------------------------------------------- -0: First time : 2021/03/27 21:00:00 UTC -0: Interval : 01:00:00 -0: Number of requests : 1 -0: Fields : Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: Requested output fields not yet available: -0: ----------------------------------------------------- -0: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -0: -0: Successfully requested output fields : -0: ----------------------------------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: -0: Additional GRIB parameters : -0: ----------------------------------------------------- -0: Run time : 2021/03/27 21:00:00 UTC -0: GRIB center ID : 7 -0: GRIB gen. proc. ID : 11 -0: GRIB grid ID : 255 -0: GRIB GDS parameter : 0 -0: Fields in file : -0: -------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: CHOSEN GRID TYPE: : LLRECTILINEAR -0: -0: -0: -0: Generating file -0: ----------------------------------------------------- -0: Data for 2021/03/27 21:00:00 UTC 0H forecast. -0: -0: End of program -0: ========================================= -0: WAVEWATCH III GRIB output -0: -0: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -0: + wave_grib2_sbs.sh[102][[ 105 -gt 0 ]] -0: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '105 hour fcst' -grib gfs.wave.t12z.atlocn.0p16.f105.grib2 -0: 1:0:d=2021032312:SPC:surface:105 hour fcst: -0: 2:20191:d=2021032312:DIRC:surface:105 hour fcst: -0: 3:60263:d=2021032312:UOGRD:surface:105 hour fcst: -0: 4:79947:d=2021032312:VOGRD:surface:105 hour fcst: -0: 5:100349:d=2021032312:WIND:surface:105 hour fcst: -0: 6:130491:d=2021032312:WDIR:surface:105 hour fcst: -0: 7:170510:d=2021032312:UGRD:surface:105 hour fcst: -0: 8:200785:d=2021032312:VGRD:surface:105 hour fcst: -0: 9:230132:d=2021032312:ICEC:surface:105 hour fcst: -0: 10:242767:d=2021032312:HTSGW:surface:105 hour fcst: -0: 11:268213:d=2021032312:IMWF:surface:105 hour fcst: -0: 12:294757:d=2021032312:MWSPER:surface:105 hour fcst: -0: 13:321590:d=2021032312:PERPW:surface:105 hour fcst: -0: 14:349330:d=2021032312:WWSDIR:surface:105 hour fcst: -0: 15:389044:d=2021032312:DIRPW:surface:105 hour fcst: -0: 16:429311:d=2021032312:WVHGT:surface:105 hour fcst: -0: 17:454212:d=2021032312:SWELL:1 in sequence:105 hour fcst: -0: 18:475265:d=2021032312:SWELL:2 in sequence:105 hour fcst: -0: 19:491620:d=2021032312:SWELL:3 in sequence:105 hour fcst: -0: 20:505484:d=2021032312:WVPER:surface:105 hour fcst: -0: 21:532252:d=2021032312:SWPER:1 in sequence:105 hour fcst: -0: 22:554699:d=2021032312:SWPER:2 in sequence:105 hour fcst: -0: 23:573107:d=2021032312:SWPER:3 in sequence:105 hour fcst: -0: 24:587901:d=2021032312:WVDIR:surface:105 hour fcst: -0: 25:625509:d=2021032312:SWDIR:1 in sequence:105 hour fcst: -0: 26:657291:d=2021032312:SWDIR:2 in sequence:105 hour fcst: -0: 27:680978:d=2021032312:SWDIR:3 in sequence:105 hour fcst: -0: + wave_grib2_sbs.sh[104]err=0 -0: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.atlocn.0p16.f105.grib2 -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f105.grib2 ]] -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f105.grib2.idx ]] -0: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.atlocn.0p16.f105.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f105.grib2 -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f105.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f105.grib2 = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f105.grib2 ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f105.grib2 -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f105.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f105.grib2.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f105.grib2.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f105.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f105.grib2 -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.atlocn.0p16.f105.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f105.grib2.idx -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f105.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f105.grib2.idx = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f105.grib2.idx ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f105.grib2.idx -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f105.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f105.grib2.idx.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f105.grib2.idx.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f105.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f105.grib2.idx -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.atlocn.0p16.f105.grib2 and gfs.wave.t12z.atlocn.0p16.f105.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_at_10m to COM' -0: INFO: Copied gfs.wave.t12z.atlocn.0p16.f105.grib2 and gfs.wave.t12z.atlocn.0p16.f105.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_at_10m to COM -0: + wave_grib2_sbs.sh[130][[ at_10m == '' ]] -0: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -0: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.atlocn.0p16.f105.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -0: INFO: gfs.wave.t12z.atlocn.0p16.f105.grib2 is global.0p50 or SENDDBN is NO, no alert sent -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + cmdfile.6[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032721 3600. 9999 -5: + cmdfile.6[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grid_interp_sbs.sh[25]grdID=gnh_10m -5: + wave_grid_interp_sbs.sh[26]valid_time=2021032721 -5: + wave_grid_interp_sbs.sh[27]dt=3600. -5: + wave_grid_interp_sbs.sh[28]nst=9999 -5: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010 -5: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.uglo_100km ./out_grd.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mod_def.uglo_100km ./mod_def.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mod_def.gnh_10m ./mod_def.gnh_10m -5: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m'\''' -5: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m' -5: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[51]weights_found=1 -5: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 210000' -5: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 210000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gnh_10m/g ww3_gint.inp.tmpl -5: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -5: $ Input file for interpolation of uglo_100km to gnh_10m -5: $------------------------------------------------ -5: $ Start Time 3600. NSteps -5: 20210327 210000 3600. 9999 -5: $ Total number of grids -5: 2 -5: $ Grid extensions -5: 'uglo_100km' -5: 'gnh_10m' -5: $ -5: 0 -5: $ -5: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[70]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.309696 ']' -5: ++ prep_step[4]echo gfs_ww3_gint.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -5: INFO: Executing 'gfs_ww3_gint.x' -5: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[73]cat grid_interp.gnh_10m.out -5: -5: *** WAVEWATCH III Grid interpolation *** -5: =============================================== -5: -5: Comment character is '$' -5: -5: Time Information : -5: --------------------------------------------- -5: Starting Time : 2021/03/27 21:00:00 UTC -5: Interval (in sec) : 3600.00 -5: Number of requests : 9999 -5: --------------------------------------------- -5: Number of grids (including output grid) = 2 -5: -5: -5: Extension for grid 1 is --> uglo_100km -5: -5: Grid Particulars are : -5: Dimensions = 45166 1 -5: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Extension for grid 2 is --> gnh_10m -5: -5: Grid Particulars are : -5: Dimensions = 2160 406 -5: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -5: -5: -5: Preparing interpolation weights for output grid -5: Total number of wet points for interpolation 571209 -5: -5: -5: Variable: Grid Interpolation Map Units: 0.100E+01 -5: -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 406 | | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: 355 | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 304 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 253 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 202 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 151 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 100 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 49 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: -5: -5: Interpolating fields .... -5: -5: Output group 1 -5: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -5: Output group 2 -5: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -5: Output group 3 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 4 -5: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -5: Output group 5 -5: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -5: Output group 6 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 7 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 8 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 9 -5: Output variables skipped -5: Output group 10 -5: Output variables skipped -5: ------------------------------------------------ -5: 1Current vel. -5: 1Wind speed -5: 1Ice concentration -5: 2Wave height -5: 2Mean wave period(+2) -5: 2Mean wave period(+1) -5: 2Peak frequency -5: 2Mean wave dir. a1b1 -5: 2Peak direction -5: 4Part. wave height -5: 4Part. peak period -5: 4Part. mean direction -5: 5Charnock parameter -5: ------------------------------------------------ -5: OUTPUT TIME : 2021/03/27 21:00:00 UTC -5: -5: End of file reached -5: -5: -5: *** End of Grid interpolation Routine *** -5: =============================================== -5: -5: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -5: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -5: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_gnh_10m/out_grd.gnh_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.gnh_10m -5: + cmdfile.6[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032721 105 global 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + cmdfile.6[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib2_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grib2_sbs.sh[30]grdID=gnh_10m -5: + wave_grib2_sbs.sh[31]GRIDNR=255 -5: + wave_grib2_sbs.sh[32]MODNR=11 -5: + wave_grib2_sbs.sh[33]valid_time=2021032721 -5: + wave_grib2_sbs.sh[34]fhr=105 -5: + wave_grib2_sbs.sh[35]grid_region=global -5: + wave_grib2_sbs.sh[36]grid_res=0p16 -5: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010 -5: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_gnh_10m -5: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_gnh_10m -5: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_gnh_10m -5: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_gnh_10m -5: ++ wave_grib2_sbs.sh[47]printf %03i 105 -5: + wave_grib2_sbs.sh[47]FH3=105 -5: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p16 -5: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p16.f105.grib2 -5: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f105.grib2 ]] -5: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mod_def.gnh_10m ./mod_def.ww3 -5: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.gnh_10m ./out_grd.ww3 -5: + wave_grib2_sbs.sh[73]ngrib=1 -5: + wave_grib2_sbs.sh[74]dtgrib=3600 -5: + wave_grib2_sbs.sh[75]tstart='20210327 210000' -5: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 210000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gnh_10m.inp.tmpl -5: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -5: $ WAVEWATCH-III gridded output input file -5: $ ---------------------------------------- -5: 20210327 210000 3600 1 -5: N -5: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -5: $ -5: 20210327 210000 7 11 255 0 0 -5: $ -5: $ end of input file -5: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[88]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.309696 ']' -5: ++ prep_step[4]echo gfs_ww3_grib.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[90]export err=0 -5: + wave_grib2_sbs.sh[90]err=0 -5: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[95]cat grib2_global_105.out -5: -5: *** WAVEWATCH III GRIB output postp. *** -5: ============================================== -5: -5: Comment character is '$' -5: -5: Grid name : GFSv16-wave N Hemisphere 1/6 d -5: -5: LINEIN: -5: 20210327 210000 3600 1 -5: -5: 20210327210000 3600 1 -5: GEN_PRO -99999 -5: -5: Output time data : -5: ----------------------------------------------------- -5: First time : 2021/03/27 21:00:00 UTC -5: Interval : 01:00:00 -5: Number of requests : 1 -5: Fields : Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: Requested output fields not yet available: -5: ----------------------------------------------------- -5: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -5: -5: Successfully requested output fields : -5: ----------------------------------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: -5: Additional GRIB parameters : -5: ----------------------------------------------------- -5: Run time : 2021/03/27 21:00:00 UTC -5: GRIB center ID : 7 -5: GRIB gen. proc. ID : 11 -5: GRIB grid ID : 255 -5: GRIB GDS parameter : 0 -5: Fields in file : -5: -------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: CHOSEN GRID TYPE: : LLRECTILINEAR -5: -5: -5: -5: Generating file -5: ----------------------------------------------------- -5: Data for 2021/03/27 21:00:00 UTC 0H forecast. -5: -5: End of program -5: ========================================= -5: WAVEWATCH III GRIB output -5: -5: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -5: + wave_grib2_sbs.sh[102][[ 105 -gt 0 ]] -5: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '105 hour fcst' -grib gfs.wave.t12z.global.0p16.f105.grib2 -5: 1:0:d=2021032312:SPC:surface:105 hour fcst: -5: 2:238888:d=2021032312:DIRC:surface:105 hour fcst: -5: 3:873333:d=2021032312:UOGRD:surface:105 hour fcst: -5: 4:1108534:d=2021032312:VOGRD:surface:105 hour fcst: -5: 5:1348207:d=2021032312:WIND:surface:105 hour fcst: -5: 6:1757306:d=2021032312:WDIR:surface:105 hour fcst: -5: 7:2395442:d=2021032312:UGRD:surface:105 hour fcst: -5: 8:2797278:d=2021032312:VGRD:surface:105 hour fcst: -5: 9:3201727:d=2021032312:ICEC:surface:105 hour fcst: -5: 10:3316196:d=2021032312:HTSGW:surface:105 hour fcst: -5: 11:3604485:d=2021032312:IMWF:surface:105 hour fcst: -5: 12:3917062:d=2021032312:MWSPER:surface:105 hour fcst: -5: 13:4231974:d=2021032312:PERPW:surface:105 hour fcst: -5: 14:4568384:d=2021032312:WWSDIR:surface:105 hour fcst: -5: 15:5148019:d=2021032312:DIRPW:surface:105 hour fcst: -5: 16:5752136:d=2021032312:WVHGT:surface:105 hour fcst: -5: 17:6029061:d=2021032312:SWELL:1 in sequence:105 hour fcst: -5: 18:6325269:d=2021032312:SWELL:2 in sequence:105 hour fcst: -5: 19:6527804:d=2021032312:SWELL:3 in sequence:105 hour fcst: -5: 20:6657982:d=2021032312:WVPER:surface:105 hour fcst: -5: 21:6967525:d=2021032312:SWPER:1 in sequence:105 hour fcst: -5: 22:7303079:d=2021032312:SWPER:2 in sequence:105 hour fcst: -5: 23:7552194:d=2021032312:SWPER:3 in sequence:105 hour fcst: -5: 24:7699677:d=2021032312:WVDIR:surface:105 hour fcst: -5: 25:8191823:d=2021032312:SWDIR:1 in sequence:105 hour fcst: -5: 26:8797053:d=2021032312:SWDIR:2 in sequence:105 hour fcst: -5: 27:9191554:d=2021032312:SWDIR:3 in sequence:105 hour fcst: -5: + wave_grib2_sbs.sh[104]err=0 -5: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p16.f105.grib2 -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f105.grib2 ]] -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f105.grib2.idx ]] -5: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p16.f105.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f105.grib2 -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f105.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f105.grib2 = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f105.grib2 ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f105.grib2 -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f105.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f105.grib2.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f105.grib2.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f105.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f105.grib2 -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p16.f105.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f105.grib2.idx -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f105.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f105.grib2.idx = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f105.grib2.idx ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f105.grib2.idx -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f105.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f105.grib2.idx.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f105.grib2.idx.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f105.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f105.grib2.idx -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p16.f105.grib2 and gfs.wave.t12z.global.0p16.f105.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_gnh_10m to COM' -5: INFO: Copied gfs.wave.t12z.global.0p16.f105.grib2 and gfs.wave.t12z.global.0p16.f105.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_gnh_10m to COM -5: + wave_grib2_sbs.sh[130][[ gnh_10m == '' ]] -5: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -5: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p16.f105.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -5: INFO: gfs.wave.t12z.global.0p16.f105.grib2 is global.0p50 or SENDDBN is NO, no alert sent -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + cmdfile.4[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032721 3600. 9999 -3: + cmdfile.4[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grid_interp_sbs.sh[25]grdID=glo_30m -3: + wave_grid_interp_sbs.sh[26]valid_time=2021032721 -3: + wave_grid_interp_sbs.sh[27]dt=3600. -3: + wave_grid_interp_sbs.sh[28]nst=9999 -3: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010 -3: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.uglo_100km ./out_grd.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mod_def.uglo_100km ./mod_def.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mod_def.glo_30m ./mod_def.glo_30m -3: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ]] -3: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m'\''' -3: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m' -3: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[51]weights_found=1 -3: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 210000' -3: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 210000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/glo_30m/g ww3_gint.inp.tmpl -3: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -3: $ Input file for interpolation of uglo_100km to glo_30m -3: $------------------------------------------------ -3: $ Start Time 3600. NSteps -3: 20210327 210000 3600. 9999 -3: $ Total number of grids -3: 2 -3: $ Grid extensions -3: 'uglo_100km' -3: 'glo_30m' -3: $ -3: 0 -3: $ -3: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[70]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.309696 ']' -3: ++ prep_step[4]echo gfs_ww3_gint.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -3: INFO: Executing 'gfs_ww3_gint.x' -3: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[73]cat grid_interp.glo_30m.out -3: -3: *** WAVEWATCH III Grid interpolation *** -3: =============================================== -3: -3: Comment character is '$' -3: -3: Time Information : -3: --------------------------------------------- -3: Starting Time : 2021/03/27 21:00:00 UTC -3: Interval (in sec) : 3600.00 -3: Number of requests : 9999 -3: --------------------------------------------- -3: Number of grids (including output grid) = 2 -3: -3: -3: Extension for grid 1 is --> uglo_100km -3: -3: Grid Particulars are : -3: Dimensions = 45166 1 -3: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Extension for grid 2 is --> glo_30m -3: -3: Grid Particulars are : -3: Dimensions = 720 336 -3: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -3: -3: -3: Preparing interpolation weights for output grid -3: Total number of wet points for interpolation 167619 -3: -3: -3: Variable: Grid Interpolation Map Units: 0.100E+01 -3: -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 336 | | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 291 | 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 | -3: 246 | 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 201 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 156 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 111 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 66 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 21 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 | -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: -3: -3: Interpolating fields .... -3: -3: Output group 1 -3: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -3: Output group 2 -3: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -3: Output group 3 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 4 -3: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -3: Output group 5 -3: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -3: Output group 6 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 7 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 8 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 9 -3: Output variables skipped -3: Output group 10 -3: Output variables skipped -3: ------------------------------------------------ -3: 1Current vel. -3: 1Wind speed -3: 1Ice concentration -3: 2Wave height -3: 2Mean wave period(+2) -3: 2Mean wave period(+1) -3: 2Peak frequency -3: 2Mean wave dir. a1b1 -3: 2Peak direction -3: 4Part. wave height -3: 4Part. peak period -3: 4Part. mean direction -3: 5Charnock parameter -3: ------------------------------------------------ -3: OUTPUT TIME : 2021/03/27 21:00:00 UTC -3: -3: End of file reached -3: -3: -3: *** End of Grid interpolation Routine *** -3: =============================================== -3: -3: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -3: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -3: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_glo_30m/out_grd.glo_30m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.glo_30m -3: + cmdfile.4[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032721 105 global 0p50 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + cmdfile.4[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib2_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grib2_sbs.sh[30]grdID=glo_30m -3: + wave_grib2_sbs.sh[31]GRIDNR=255 -3: + wave_grib2_sbs.sh[32]MODNR=11 -3: + wave_grib2_sbs.sh[33]valid_time=2021032721 -3: + wave_grib2_sbs.sh[34]fhr=105 -3: + wave_grib2_sbs.sh[35]grid_region=global -3: + wave_grib2_sbs.sh[36]grid_res=0p50 -3: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010 -3: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_glo_30m -3: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_glo_30m -3: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_glo_30m -3: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_glo_30m -3: ++ wave_grib2_sbs.sh[47]printf %03i 105 -3: + wave_grib2_sbs.sh[47]FH3=105 -3: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p50 -3: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p50.f105.grib2 -3: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f105.grib2 ]] -3: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mod_def.glo_30m ./mod_def.ww3 -3: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.glo_30m ./out_grd.ww3 -3: + wave_grib2_sbs.sh[73]ngrib=1 -3: + wave_grib2_sbs.sh[74]dtgrib=3600 -3: + wave_grib2_sbs.sh[75]tstart='20210327 210000' -3: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 210000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.glo_30m.inp.tmpl -3: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -3: $ WAVEWATCH-III gridded output input file -3: $ ---------------------------------------- -3: 20210327 210000 3600 1 -3: N -3: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -3: $ -3: 20210327 210000 7 11 255 0 0 -3: $ -3: $ end of input file -3: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[88]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.309696 ']' -3: ++ prep_step[4]echo gfs_ww3_grib.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[90]export err=0 -3: + wave_grib2_sbs.sh[90]err=0 -3: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[95]cat grib2_global_105.out -3: -3: *** WAVEWATCH III GRIB output postp. *** -3: ============================================== -3: -3: Comment character is '$' -3: -3: Grid name : Global 30 min wave grid -3: -3: LINEIN: -3: 20210327 210000 3600 1 -3: -3: 20210327210000 3600 1 -3: GEN_PRO -99999 -3: -3: Output time data : -3: ----------------------------------------------------- -3: First time : 2021/03/27 21:00:00 UTC -3: Interval : 01:00:00 -3: Number of requests : 1 -3: Fields : Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: Requested output fields not yet available: -3: ----------------------------------------------------- -3: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -3: -3: Successfully requested output fields : -3: ----------------------------------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: -3: Additional GRIB parameters : -3: ----------------------------------------------------- -3: Run time : 2021/03/27 21:00:00 UTC -3: GRIB center ID : 7 -3: GRIB gen. proc. ID : 11 -3: GRIB grid ID : 255 -3: GRIB GDS parameter : 0 -3: Fields in file : -3: -------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: CHOSEN GRID TYPE: : LLRECTILINEAR -3: -3: -3: -3: Generating file -3: ----------------------------------------------------- -3: Data for 2021/03/27 21:00:00 UTC 0H forecast. -3: -3: End of program -3: ========================================= -3: WAVEWATCH III GRIB output -3: -3: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -3: + wave_grib2_sbs.sh[102][[ 105 -gt 0 ]] -3: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '105 hour fcst' -grib gfs.wave.t12z.global.0p50.f105.grib2 -3: 1:0:d=2021032312:SPC:surface:105 hour fcst: -3: 2:78092:d=2021032312:DIRC:surface:105 hour fcst: -3: 3:321733:d=2021032312:UOGRD:surface:105 hour fcst: -3: 4:397307:d=2021032312:VOGRD:surface:105 hour fcst: -3: 5:477082:d=2021032312:WIND:surface:105 hour fcst: -3: 6:635801:d=2021032312:WDIR:surface:105 hour fcst: -3: 7:872275:d=2021032312:UGRD:surface:105 hour fcst: -3: 8:1027487:d=2021032312:VGRD:surface:105 hour fcst: -3: 9:1184264:d=2021032312:ICEC:surface:105 hour fcst: -3: 10:1224401:d=2021032312:HTSGW:surface:105 hour fcst: -3: 11:1323025:d=2021032312:IMWF:surface:105 hour fcst: -3: 12:1433116:d=2021032312:MWSPER:surface:105 hour fcst: -3: 13:1544223:d=2021032312:PERPW:surface:105 hour fcst: -3: 14:1661563:d=2021032312:WWSDIR:surface:105 hour fcst: -3: 15:1862166:d=2021032312:DIRPW:surface:105 hour fcst: -3: 16:2068263:d=2021032312:WVHGT:surface:105 hour fcst: -3: 17:2163449:d=2021032312:SWELL:1 in sequence:105 hour fcst: -3: 18:2268552:d=2021032312:SWELL:2 in sequence:105 hour fcst: -3: 19:2342686:d=2021032312:SWELL:3 in sequence:105 hour fcst: -3: 20:2387509:d=2021032312:WVPER:surface:105 hour fcst: -3: 21:2494316:d=2021032312:SWPER:1 in sequence:105 hour fcst: -3: 22:2613103:d=2021032312:SWPER:2 in sequence:105 hour fcst: -3: 23:2706584:d=2021032312:SWPER:3 in sequence:105 hour fcst: -3: 24:2760128:d=2021032312:WVDIR:surface:105 hour fcst: -3: 25:2921678:d=2021032312:SWDIR:1 in sequence:105 hour fcst: -3: 26:3125886:d=2021032312:SWDIR:2 in sequence:105 hour fcst: -3: 27:3273652:d=2021032312:SWDIR:3 in sequence:105 hour fcst: -3: + wave_grib2_sbs.sh[104]err=0 -3: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p50.f105.grib2 -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f105.grib2 ]] -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f105.grib2.idx ]] -3: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p50.f105.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f105.grib2 -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f105.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f105.grib2 = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f105.grib2 ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f105.grib2 -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f105.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f105.grib2.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f105.grib2.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f105.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f105.grib2 -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p50.f105.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f105.grib2.idx -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f105.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f105.grib2.idx = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f105.grib2.idx ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f105.grib2.idx -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f105.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f105.grib2.idx.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f105.grib2.idx.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f105.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f105.grib2.idx -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p50.f105.grib2 and gfs.wave.t12z.global.0p50.f105.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_glo_30m to COM' -3: INFO: Copied gfs.wave.t12z.global.0p50.f105.grib2 and gfs.wave.t12z.global.0p50.f105.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_glo_30m to COM -3: + wave_grib2_sbs.sh[130][[ glo_30m == '' ]] -3: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -3: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p50.f105.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -3: INFO: gfs.wave.t12z.global.0p50.f105.grib2 is global.0p50 or SENDDBN is NO, no alert sent -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + cmdfile.5[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032721 3600. 9999 -4: + cmdfile.5[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grid_interp_sbs.sh[25]grdID=aoc_9km -4: + wave_grid_interp_sbs.sh[26]valid_time=2021032721 -4: + wave_grid_interp_sbs.sh[27]dt=3600. -4: + wave_grid_interp_sbs.sh[28]nst=9999 -4: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010 -4: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.uglo_100km ./out_grd.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mod_def.uglo_100km ./mod_def.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mod_def.aoc_9km ./mod_def.aoc_9km -4: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km'\''' -4: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km' -4: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[51]weights_found=1 -4: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 210000' -4: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 210000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/aoc_9km/g ww3_gint.inp.tmpl -4: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -4: $ Input file for interpolation of uglo_100km to aoc_9km -4: $------------------------------------------------ -4: $ Start Time 3600. NSteps -4: 20210327 210000 3600. 9999 -4: $ Total number of grids -4: 2 -4: $ Grid extensions -4: 'uglo_100km' -4: 'aoc_9km' -4: $ -4: 0 -4: $ -4: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[70]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.309696 ']' -4: ++ prep_step[4]echo gfs_ww3_gint.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -4: INFO: Executing 'gfs_ww3_gint.x' -4: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[73]cat grid_interp.aoc_9km.out -4: -4: *** WAVEWATCH III Grid interpolation *** -4: =============================================== -4: -4: Comment character is '$' -4: -4: Time Information : -4: --------------------------------------------- -4: Starting Time : 2021/03/27 21:00:00 UTC -4: Interval (in sec) : 3600.00 -4: Number of requests : 9999 -4: --------------------------------------------- -4: Number of grids (including output grid) = 2 -4: -4: -4: Extension for grid 1 is --> uglo_100km -4: -4: Grid Particulars are : -4: Dimensions = 45166 1 -4: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Extension for grid 2 is --> aoc_9km -4: -4: Grid Particulars are : -4: Dimensions = 1006 1006 -4: Grid Type = 2 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -4: -4: -4: Preparing interpolation weights for output grid -4: Total number of wet points for interpolation 360052 -4: -4: -4: Variable: Grid Interpolation Map Units: 0.100E+01 -4: -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: *** | | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: 880 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 | -4: 754 | 0 0 0 | -4: | 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 628 | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 | -4: 502 | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 376 | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 250 | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 124 | 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 | -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: -4: -4: Interpolating fields .... -4: -4: Output group 1 -4: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -4: Output group 2 -4: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -4: Output group 3 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 4 -4: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -4: Output group 5 -4: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -4: Output group 6 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 7 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 8 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 9 -4: Output variables skipped -4: Output group 10 -4: Output variables skipped -4: ------------------------------------------------ -4: 1Current vel. -4: 1Wind speed -4: 1Ice concentration -4: 2Wave height -4: 2Mean wave period(+2) -4: 2Mean wave period(+1) -4: 2Peak frequency -4: 2Mean wave dir. a1b1 -4: 2Peak direction -4: 4Part. wave height -4: 4Part. peak period -4: 4Part. mean direction -4: 5Charnock parameter -4: ------------------------------------------------ -4: OUTPUT TIME : 2021/03/27 21:00:00 UTC -4: -4: End of file reached -4: -4: -4: *** End of Grid interpolation Routine *** -4: =============================================== -4: -4: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -4: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -4: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_aoc_9km/out_grd.aoc_9km /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.aoc_9km -4: + cmdfile.5[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032721 105 arctic 9km 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + cmdfile.5[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib2_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grib2_sbs.sh[30]grdID=aoc_9km -4: + wave_grib2_sbs.sh[31]GRIDNR=255 -4: + wave_grib2_sbs.sh[32]MODNR=11 -4: + wave_grib2_sbs.sh[33]valid_time=2021032721 -4: + wave_grib2_sbs.sh[34]fhr=105 -4: + wave_grib2_sbs.sh[35]grid_region=arctic -4: + wave_grib2_sbs.sh[36]grid_res=9km -4: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010 -4: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_aoc_9km -4: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_aoc_9km -4: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_aoc_9km -4: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_aoc_9km -4: ++ wave_grib2_sbs.sh[47]printf %03i 105 -4: + wave_grib2_sbs.sh[47]FH3=105 -4: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_arctic_9km -4: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.arctic.9km.f105.grib2 -4: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f105.grib2 ]] -4: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mod_def.aoc_9km ./mod_def.ww3 -4: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.aoc_9km ./out_grd.ww3 -4: + wave_grib2_sbs.sh[73]ngrib=1 -4: + wave_grib2_sbs.sh[74]dtgrib=3600 -4: + wave_grib2_sbs.sh[75]tstart='20210327 210000' -4: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 210000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.aoc_9km.inp.tmpl -4: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -4: $ WAVEWATCH-III gridded output input file -4: $ ---------------------------------------- -4: 20210327 210000 3600 1 -4: N -4: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -4: $ -4: 20210327 210000 7 11 255 0 20 -4: $ -4: 70 0 9.0 9.0 64 -4: $ 60 0 8.64919046313 8.64919046313 64 -4: $ end of input file -4: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[88]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.309696 ']' -4: ++ prep_step[4]echo gfs_ww3_grib.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[90]export err=0 -4: + wave_grib2_sbs.sh[90]err=0 -4: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[95]cat grib2_arctic_105.out -4: -4: *** WAVEWATCH III GRIB output postp. *** -4: ============================================== -4: -4: Comment character is '$' -4: -4: Grid name : Arctic Ocean PolarStereo 9km -4: -4: LINEIN: -4: 20210327 210000 3600 1 -4: -4: 20210327210000 3600 1 -4: GEN_PRO -99999 -4: -4: Output time data : -4: ----------------------------------------------------- -4: First time : 2021/03/27 21:00:00 UTC -4: Interval : 01:00:00 -4: Number of requests : 1 -4: Fields : Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: Requested output fields not yet available: -4: ----------------------------------------------------- -4: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -4: -4: Successfully requested output fields : -4: ----------------------------------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: -4: Additional GRIB parameters : -4: ----------------------------------------------------- -4: Run time : 2021/03/27 21:00:00 UTC -4: GRIB center ID : 7 -4: GRIB gen. proc. ID : 11 -4: GRIB grid ID : 255 -4: GRIB GDS parameter : 0 -4: Fields in file : -4: -------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: CHOSEN GRID TYPE: : POLARSTEREO -4: -4: -4: -4: Generating file -4: ----------------------------------------------------- -4: Data for 2021/03/27 21:00:00 UTC 0H forecast. -4: -4: End of program -4: ========================================= -4: WAVEWATCH III GRIB output -4: -4: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -4: + wave_grib2_sbs.sh[102][[ 105 -gt 0 ]] -4: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '105 hour fcst' -grib gfs.wave.t12z.arctic.9km.f105.grib2 -4: 1:0:d=2021032312:SPC:surface:105 hour fcst: -4: 2:191047:d=2021032312:DIRC:surface:105 hour fcst: -4: 3:653244:d=2021032312:UOGRD:surface:105 hour fcst: -4: 4:839245:d=2021032312:VOGRD:surface:105 hour fcst: -4: 5:1030460:d=2021032312:WIND:surface:105 hour fcst: -4: 6:1333919:d=2021032312:WDIR:surface:105 hour fcst: -4: 7:1765670:d=2021032312:UGRD:surface:105 hour fcst: -4: 8:2063856:d=2021032312:VGRD:surface:105 hour fcst: -4: 9:2367142:d=2021032312:ICEC:surface:105 hour fcst: -4: 10:2536398:d=2021032312:HTSGW:surface:105 hour fcst: -4: 11:2751691:d=2021032312:IMWF:surface:105 hour fcst: -4: 12:2970966:d=2021032312:MWSPER:surface:105 hour fcst: -4: 13:3191825:d=2021032312:PERPW:surface:105 hour fcst: -4: 14:3422526:d=2021032312:WWSDIR:surface:105 hour fcst: -4: 15:3733186:d=2021032312:DIRPW:surface:105 hour fcst: -4: 16:4051817:d=2021032312:WVHGT:surface:105 hour fcst: -4: 17:4266173:d=2021032312:SWELL:1 in sequence:105 hour fcst: -4: 18:4476201:d=2021032312:SWELL:2 in sequence:105 hour fcst: -4: 19:4640351:d=2021032312:SWELL:3 in sequence:105 hour fcst: -4: 20:4775054:d=2021032312:WVPER:surface:105 hour fcst: -4: 21:4999710:d=2021032312:SWPER:1 in sequence:105 hour fcst: -4: 22:5219991:d=2021032312:SWPER:2 in sequence:105 hour fcst: -4: 23:5397719:d=2021032312:SWPER:3 in sequence:105 hour fcst: -4: 24:5536508:d=2021032312:WVDIR:surface:105 hour fcst: -4: 25:5826979:d=2021032312:SWDIR:1 in sequence:105 hour fcst: -4: 26:6134794:d=2021032312:SWDIR:2 in sequence:105 hour fcst: -4: 27:6354264:d=2021032312:SWDIR:3 in sequence:105 hour fcst: -4: + wave_grib2_sbs.sh[104]err=0 -4: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.arctic.9km.f105.grib2 -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f105.grib2 ]] -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f105.grib2.idx ]] -4: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.arctic.9km.f105.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f105.grib2 -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f105.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f105.grib2 = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f105.grib2 ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f105.grib2 -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f105.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f105.grib2.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f105.grib2.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f105.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f105.grib2 -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.arctic.9km.f105.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f105.grib2.idx -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f105.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f105.grib2.idx = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f105.grib2.idx ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f105.grib2.idx -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f105.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f105.grib2.idx.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f105.grib2.idx.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f105.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f105.grib2.idx -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.arctic.9km.f105.grib2 and gfs.wave.t12z.arctic.9km.f105.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_aoc_9km to COM' -4: INFO: Copied gfs.wave.t12z.arctic.9km.f105.grib2 and gfs.wave.t12z.arctic.9km.f105.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_aoc_9km to COM -4: + wave_grib2_sbs.sh[130][[ aoc_9km == '' ]] -4: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -4: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.arctic.9km.f105.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -4: INFO: gfs.wave.t12z.arctic.9km.f105.grib2 is global.0p50 or SENDDBN is NO, no alert sent -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + cmdfile.3[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032721 3600. 9999 -2: + cmdfile.3[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grid_interp_sbs.sh[25]grdID=wc_10m -2: + wave_grid_interp_sbs.sh[26]valid_time=2021032721 -2: + wave_grid_interp_sbs.sh[27]dt=3600. -2: + wave_grid_interp_sbs.sh[28]nst=9999 -2: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010 -2: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.uglo_100km ./out_grd.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mod_def.uglo_100km ./mod_def.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mod_def.wc_10m ./mod_def.wc_10m -2: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ]] -2: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m'\''' -2: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m' -2: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[51]weights_found=1 -2: + wave_grid_interp_sbs.sh[59]ymdhms='20210327 210000' -2: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210327 210000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/wc_10m/g ww3_gint.inp.tmpl -2: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -2: $ Input file for interpolation of uglo_100km to wc_10m -2: $------------------------------------------------ -2: $ Start Time 3600. NSteps -2: 20210327 210000 3600. 9999 -2: $ Total number of grids -2: 2 -2: $ Grid extensions -2: 'uglo_100km' -2: 'wc_10m' -2: $ -2: 0 -2: $ -2: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[70]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.309696 ']' -2: ++ prep_step[4]echo gfs_ww3_gint.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -2: INFO: Executing 'gfs_ww3_gint.x' -2: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[73]cat grid_interp.wc_10m.out -2: -2: *** WAVEWATCH III Grid interpolation *** -2: =============================================== -2: -2: Comment character is '$' -2: -2: Time Information : -2: --------------------------------------------- -2: Starting Time : 2021/03/27 21:00:00 UTC -2: Interval (in sec) : 3600.00 -2: Number of requests : 9999 -2: --------------------------------------------- -2: Number of grids (including output grid) = 2 -2: -2: -2: Extension for grid 1 is --> uglo_100km -2: -2: Grid Particulars are : -2: Dimensions = 45166 1 -2: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Extension for grid 2 is --> wc_10m -2: -2: Grid Particulars are : -2: Dimensions = 241 151 -2: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -2: -2: -2: Preparing interpolation weights for output grid -2: Total number of wet points for interpolation 11044 -2: -2: -2: Variable: Grid Interpolation Map Units: 0.100E+01 -2: -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: +---------------------------------------------------------------------------------------------------------------+ -2: 151 | 0 0 0 0 | -2: | 0 0 0 0 0 | -2: | 0 0 0 0 0 | -2: 130 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 109 | 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 88 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 | -2: 67 | 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 | -2: 46 | 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: 25 | 0 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 0 | -2: | | -2: 4 | | -2: +---------------------------------------------------------------------------------------------------------------+ -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: -2: -2: Interpolating fields .... -2: -2: Output group 1 -2: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -2: Output group 2 -2: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -2: Output group 3 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 4 -2: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -2: Output group 5 -2: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -2: Output group 6 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 7 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 8 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 9 -2: Output variables skipped -2: Output group 10 -2: Output variables skipped -2: ------------------------------------------------ -2: 1Current vel. -2: 1Wind speed -2: 1Ice concentration -2: 2Wave height -2: 2Mean wave period(+2) -2: 2Mean wave period(+1) -2: 2Peak frequency -2: 2Mean wave dir. a1b1 -2: 2Peak direction -2: 4Part. wave height -2: 4Part. peak period -2: 4Part. mean direction -2: 5Charnock parameter -2: ------------------------------------------------ -2: OUTPUT TIME : 2021/03/27 21:00:00 UTC -2: -2: End of file reached -2: -2: -2: *** End of Grid interpolation Routine *** -2: =============================================== -2: -2: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -2: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -2: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grid_interp_wc_10m/out_grd.wc_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.wc_10m -2: + cmdfile.3[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032721 105 wcoast 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + cmdfile.3[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib2_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grib2_sbs.sh[30]grdID=wc_10m -2: + wave_grib2_sbs.sh[31]GRIDNR=255 -2: + wave_grib2_sbs.sh[32]MODNR=11 -2: + wave_grib2_sbs.sh[33]valid_time=2021032721 -2: + wave_grib2_sbs.sh[34]fhr=105 -2: + wave_grib2_sbs.sh[35]grid_region=wcoast -2: + wave_grib2_sbs.sh[36]grid_res=0p16 -2: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010 -2: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_wc_10m -2: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_wc_10m -2: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_wc_10m -2: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_wc_10m -2: ++ wave_grib2_sbs.sh[47]printf %03i 105 -2: + wave_grib2_sbs.sh[47]FH3=105 -2: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_wcoast_0p16 -2: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.wcoast.0p16.f105.grib2 -2: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f105.grib2 ]] -2: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/mod_def.wc_10m ./mod_def.ww3 -2: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/out_grd.wc_10m ./out_grd.ww3 -2: + wave_grib2_sbs.sh[73]ngrib=1 -2: + wave_grib2_sbs.sh[74]dtgrib=3600 -2: + wave_grib2_sbs.sh[75]tstart='20210327 210000' -2: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210327 210000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.wc_10m.inp.tmpl -2: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -2: $ WAVEWATCH-III gridded output input file -2: $ ---------------------------------------- -2: 20210327 210000 3600 1 -2: N -2: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -2: $ -2: 20210327 210000 7 11 255 0 0 -2: $ -2: $ end of input file -2: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[88]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.309696 ']' -2: ++ prep_step[4]echo gfs_ww3_grib.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[90]export err=0 -2: + wave_grib2_sbs.sh[90]err=0 -2: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[95]cat grib2_wcoast_105.out -2: -2: *** WAVEWATCH III GRIB output postp. *** -2: ============================================== -2: -2: Comment character is '$' -2: -2: Grid name : West Coast 10 min wave grid -2: -2: LINEIN: -2: 20210327 210000 3600 1 -2: -2: 20210327210000 3600 1 -2: GEN_PRO -99999 -2: -2: Output time data : -2: ----------------------------------------------------- -2: First time : 2021/03/27 21:00:00 UTC -2: Interval : 01:00:00 -2: Number of requests : 1 -2: Fields : Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: Requested output fields not yet available: -2: ----------------------------------------------------- -2: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -2: -2: Successfully requested output fields : -2: ----------------------------------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: -2: Additional GRIB parameters : -2: ----------------------------------------------------- -2: Run time : 2021/03/27 21:00:00 UTC -2: GRIB center ID : 7 -2: GRIB gen. proc. ID : 11 -2: GRIB grid ID : 255 -2: GRIB GDS parameter : 0 -2: Fields in file : -2: -------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: CHOSEN GRID TYPE: : LLRECTILINEAR -2: -2: -2: -2: Generating file -2: ----------------------------------------------------- -2: Data for 2021/03/27 21:00:00 UTC 0H forecast. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: Warning: bitmask off everywhere. -2: Pretend one point in jpcpack to avoid crash. -2: -2: End of program -2: ========================================= -2: WAVEWATCH III GRIB output -2: -2: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -2: + wave_grib2_sbs.sh[102][[ 105 -gt 0 ]] -2: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '105 hour fcst' -grib gfs.wave.t12z.wcoast.0p16.f105.grib2 -2: 1:0:d=2021032312:SPC:surface:105 hour fcst: -2: 2:7280:d=2021032312:DIRC:surface:105 hour fcst: -2: 3:24945:d=2021032312:UOGRD:surface:105 hour fcst: -2: 4:32299:d=2021032312:VOGRD:surface:105 hour fcst: -2: 5:39835:d=2021032312:WIND:surface:105 hour fcst: -2: 6:51853:d=2021032312:WDIR:surface:105 hour fcst: -2: 7:69212:d=2021032312:UGRD:surface:105 hour fcst: -2: 8:80436:d=2021032312:VGRD:surface:105 hour fcst: -2: 9:92388:d=2021032312:ICEC:surface:105 hour fcst: -2: 10:97118:d=2021032312:HTSGW:surface:105 hour fcst: -2: 11:106643:d=2021032312:IMWF:surface:105 hour fcst: -2: 12:116555:d=2021032312:MWSPER:surface:105 hour fcst: -2: 13:126441:d=2021032312:PERPW:surface:105 hour fcst: -2: 14:136418:d=2021032312:WWSDIR:surface:105 hour fcst: -2: 15:152016:d=2021032312:DIRPW:surface:105 hour fcst: -2: 16:167828:d=2021032312:WVHGT:surface:105 hour fcst: -2: 17:175171:d=2021032312:SWELL:1 in sequence:105 hour fcst: -2: 18:183824:d=2021032312:SWELL:2 in sequence:105 hour fcst: -2: 19:188964:d=2021032312:SWELL:3 in sequence:105 hour fcst: -2: 20:193694:d=2021032312:WVPER:surface:105 hour fcst: -2: 21:201497:d=2021032312:SWPER:1 in sequence:105 hour fcst: -2: 22:210124:d=2021032312:SWPER:2 in sequence:105 hour fcst: -2: 23:215523:d=2021032312:SWPER:3 in sequence:105 hour fcst: -2: 24:220253:d=2021032312:WVDIR:surface:105 hour fcst: -2: 25:230225:d=2021032312:SWDIR:1 in sequence:105 hour fcst: -2: 26:243295:d=2021032312:SWDIR:2 in sequence:105 hour fcst: -2: 27:248972:d=2021032312:SWDIR:3 in sequence:105 hour fcst: -2: + wave_grib2_sbs.sh[104]err=0 -2: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.wcoast.0p16.f105.grib2 -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f105.grib2 ]] -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f105.grib2.idx ]] -2: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.wcoast.0p16.f105.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f105.grib2 -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f105.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f105.grib2 = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f105.grib2 ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f105.grib2 -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f105.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f105.grib2.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f105.grib2.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f105.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f105.grib2 -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.wcoast.0p16.f105.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f105.grib2.idx -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f105.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f105.grib2.idx = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f105.grib2.idx ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f105.grib2.idx -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f105.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f105.grib2.idx.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f105.grib2.idx.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f105.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f105.grib2.idx -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.wcoast.0p16.f105.grib2 and gfs.wave.t12z.wcoast.0p16.f105.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_wc_10m to COM' -2: INFO: Copied gfs.wave.t12z.wcoast.0p16.f105.grib2 and gfs.wave.t12z.wcoast.0p16.f105.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010/grib_wc_10m to COM -2: + wave_grib2_sbs.sh[130][[ wc_10m == '' ]] -2: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -2: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.wcoast.0p16.f105.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -2: INFO: gfs.wave.t12z.wcoast.0p16.f105.grib2 is global.0p50 or SENDDBN is NO, no alert sent -+ run_mpmd.sh[113]exit 0 -+ run_mpmd.sh[1]postamble run_mpmd.sh 1753758302 0 -+ preamble.sh[62]set +x -End run_mpmd.sh at 03:05:13 with error code 0 (time elapsed: 00:00:11) -+ exgfs_wave_post_gridded_sbs.sh[122]true -+ exgfs_wave_post_gridded_sbs.sh[123]export err=0 -+ exgfs_wave_post_gridded_sbs.sh[123]err=0 -+ exgfs_wave_post_gridded_sbs.sh[124][[ 0 -ne 0 ]] -+ exgfs_wave_post_gridded_sbs.sh[130]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ exgfs_wave_post_gridded_sbs.sh[131]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ exgfs_wave_post_gridded_sbs.sh[132]gribchk=gfs.wave.t12z.gsouth.0p25.f105.grib2 -+ exgfs_wave_post_gridded_sbs.sh[133][[ ! -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f105.grib2 ]] -+ exgfs_wave_post_gridded_sbs.sh[138]exit 0 -+ JGLOBAL_WAVE_POST_SBS[28]true -+ JGLOBAL_WAVE_POST_SBS[29]export err=0 -+ JGLOBAL_WAVE_POST_SBS[29]err=0 -+ JGLOBAL_WAVE_POST_SBS[30][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[37]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312 -+ JGLOBAL_WAVE_POST_SBS[38][[ NO != \Y\E\S ]] -+ JGLOBAL_WAVE_POST_SBS[39]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f105.306010 -+ JGLOBAL_WAVE_POST_SBS[42]exit 0 -+ JGLOBAL_WAVE_POST_SBS[1]postamble /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS 1753758290 0 -+ preamble.sh[62]set +x -End /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at 03:05:13 with error code 0 (time elapsed: 00:00:23) -Begin /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at Tue Jul 29 03:05:13 UTC 2025 -++ jjob_header.sh[46]OPTIND=1 -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[50]env_job=wavepostsbs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[48]case "${option}" in -++ jjob_header.sh[49]read -ra configs -++ jjob_header.sh[47]getopts c:e: option -++ jjob_header.sh[61]shift 4 -++ jjob_header.sh[63][[ -z wavepostsbs ]] -++ jjob_header.sh[71]export DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010 -++ jjob_header.sh[71]DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010 -++ jjob_header.sh[72][[ YES == \Y\E\S ]] -++ jjob_header.sh[73]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010 -++ jjob_header.sh[75]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010 -++ jjob_header.sh[76]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010 -++ jjob_header.sh[85]export pid=312689 -++ jjob_header.sh[85]pid=312689 -++ jjob_header.sh[86]export pgmout=OUTPUT.312689 -++ jjob_header.sh[86]pgmout=OUTPUT.312689 -++ jjob_header.sh[87]export pgmerr=errfile -++ jjob_header.sh[87]pgmerr=errfile -++ jjob_header.sh[90]export pgm= -++ jjob_header.sh[90]pgm= -++ jjob_header.sh[96]export cycle=t12z -++ jjob_header.sh[96]cycle=t12z -++ jjob_header.sh[97]setpdy.sh -+ setpdy.sh[20]'[' /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010 == /home/mterry ']' -+ setpdy.sh[25][[ ! t12z =~ t??z ]] -+ setpdy.sh[30]case $# in -+ setpdy.sh[31]dates_before_PDY=7 -+ setpdy.sh[32]dates_after_PDY=7 -+ setpdy.sh[50]COMDATEROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+ setpdy.sh[53]'[' -z 20210323 ']' -+ setpdy.sh[57]sed 's/[0-9]\{8\}/20210323/' /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z -sed: can't read /work2/noaa/global/mterry/RUNTESTS/COMROOT/date/t12z: No such file or directory -++ jjob_header.sh[97]true -++ jjob_header.sh[98]source ./PDY -/work2/noaa/global/mterry/global-workflow_forked/ush/jjob_header.sh: line 98: ./PDY: No such file or directory -++ jjob_header.sh[98]true -++ jjob_header.sh[104]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[104]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.base -+++ config.base[6]echo 'BEGIN: config.base' -BEGIN: config.base -+++ config.base[9]export machine=HERCULES -+++ config.base[9]machine=HERCULES -+++ config.base[12]export RUN_ENVIR=emc -+++ config.base[12]RUN_ENVIR=emc -+++ config.base[15]export ACCOUNT=fv3-cpu -+++ config.base[15]ACCOUNT=fv3-cpu -+++ config.base[16]export QUEUE=batch -+++ config.base[16]QUEUE=batch -+++ config.base[17]export QUEUE_SERVICE=batch -+++ config.base[17]QUEUE_SERVICE=batch -+++ config.base[18]export QUEUE_DTN=batch -+++ config.base[18]QUEUE_DTN=batch -+++ config.base[19]export PARTITION_BATCH=hercules -+++ config.base[19]PARTITION_BATCH=hercules -+++ config.base[20]export PARTITION_SERVICE=service -+++ config.base[20]PARTITION_SERVICE=service -+++ config.base[21]export PARTITION_DTN= -+++ config.base[21]PARTITION_DTN= -+++ config.base[22]export RESERVATION= -+++ config.base[22]RESERVATION= -+++ config.base[23]export CLUSTERS= -+++ config.base[23]CLUSTERS= -+++ config.base[24]export CLUSTERS_SERVICE= -+++ config.base[24]CLUSTERS_SERVICE= -+++ config.base[25]export CLUSTERS_DTN= -+++ config.base[25]CLUSTERS_DTN= -+++ config.base[28]export HPSS_PROJECT=emc-global -+++ config.base[28]HPSS_PROJECT=emc-global -+++ config.base[31]export HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[31]HOMEgfs=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[32]export EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[32]EXECgfs=/work2/noaa/global/mterry/global-workflow_forked/exec -+++ config.base[33]export FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[33]FIXgfs=/work2/noaa/global/mterry/global-workflow_forked/fix -+++ config.base[34]export PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[34]PARMgfs=/work2/noaa/global/mterry/global-workflow_forked/parm -+++ config.base[35]export SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[35]SCRgfs=/work2/noaa/global/mterry/global-workflow_forked/scripts -+++ config.base[36]export USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[36]USHgfs=/work2/noaa/global/mterry/global-workflow_forked/ush -+++ config.base[38]export FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[38]FIXam=/work2/noaa/global/mterry/global-workflow_forked/fix/am -+++ config.base[39]export FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[39]FIXaer=/work2/noaa/global/mterry/global-workflow_forked/fix/aer -+++ config.base[40]export FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[40]FIXcpl=/work2/noaa/global/mterry/global-workflow_forked/fix/cpl -+++ config.base[41]export FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[41]FIXlut=/work2/noaa/global/mterry/global-workflow_forked/fix/lut -+++ config.base[42]export FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[42]FIXcice=/work2/noaa/global/mterry/global-workflow_forked/fix/cice -+++ config.base[43]export FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[43]FIXmom=/work2/noaa/global/mterry/global-workflow_forked/fix/mom6 -+++ config.base[44]export FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[44]FIXreg2grb2=/work2/noaa/global/mterry/global-workflow_forked/fix/reg2grb2 -+++ config.base[45]export FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[45]FIXgdas=/work2/noaa/global/mterry/global-workflow_forked/fix/gdas -+++ config.base[50]export PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[50]PACKAGEROOT=/work2/noaa/global/role-global/nwpara -+++ config.base[51]export COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[51]COMROOT=/work2/noaa/global/mterry/RUNTESTS/COMROOT -+++ config.base[52]export COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[52]COMINsyn=/work2/noaa/global/role-global/com/gfs/prod/syndat -+++ config.base[53]export DMPDIR=/work/noaa/rstprod/dump -+++ config.base[53]DMPDIR=/work/noaa/rstprod/dump -+++ config.base[57]export COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[57]COMINecmwf=/work2/noaa/global/role-global/data/external_gempak/ecmwf -+++ config.base[58]export COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[58]COMINnam=/work2/noaa/global/role-global/data/external_gempak/nam -+++ config.base[59]export COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[59]COMINukmet=/work2/noaa/global/role-global/data/external_gempak/ukmet -+++ config.base[62]export HOMEDIR=/work2/noaa/global/mterry -+++ config.base[62]HOMEDIR=/work2/noaa/global/mterry -+++ config.base[63]export STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[63]STMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]export PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[64]PTMP=/work2/noaa/stmp/mterry/HERCULES -+++ config.base[65]export NOSCRUB=/work2/noaa/global/mterry -+++ config.base[65]NOSCRUB=/work2/noaa/global/mterry -+++ config.base[68]export BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[68]BASE_GIT=/work2/noaa/global/role-global/git -+++ config.base[71]export BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[71]BASE_DATA=/work2/noaa/global/role-global/data -+++ config.base[74]export DO_PREP_SFC=NO -+++ config.base[74]DO_PREP_SFC=NO -+++ config.base[77]export DO_GOES=NO -+++ config.base[77]DO_GOES=NO -+++ config.base[78]export DO_BUFRSND=NO -+++ config.base[78]DO_BUFRSND=NO -+++ config.base[79]export DO_GEMPAK=NO -+++ config.base[79]DO_GEMPAK=NO -+++ config.base[80]export DO_AWIPS=NO -+++ config.base[80]DO_AWIPS=NO -+++ config.base[81]export DO_NPOESS=NO -+++ config.base[81]DO_NPOESS=NO -+++ config.base[82]export DO_TRACKER=YES -+++ config.base[82]DO_TRACKER=YES -+++ config.base[83]export DO_GENESIS=YES -+++ config.base[83]DO_GENESIS=YES -+++ config.base[84]export DO_GENESIS_FSU=NO -+++ config.base[84]DO_GENESIS_FSU=NO -+++ config.base[85]export DO_VERFOZN=YES -+++ config.base[85]DO_VERFOZN=YES -+++ config.base[86]export DO_VERFRAD=YES -+++ config.base[86]DO_VERFRAD=YES -+++ config.base[87]export DO_VMINMON=YES -+++ config.base[87]DO_VMINMON=YES -+++ config.base[88]export DO_ANLSTAT=NO -+++ config.base[88]DO_ANLSTAT=NO -+++ config.base[91]export MODE=forecast-only -+++ config.base[91]MODE=forecast-only -+++ config.base[92]export DO_TEST_MODE=YES -+++ config.base[92]DO_TEST_MODE=YES -+++ config.base[101]export FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[101]FIXgsi=/work2/noaa/global/mterry/global-workflow_forked/fix/gsi -+++ config.base[102]export HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[102]HOMEpost=/work2/noaa/global/mterry/global-workflow_forked -+++ config.base[103]export HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[103]HOMEobsproc=/work2/noaa/global/role-global/git/obsproc/v -+++ config.base[106]export NMV=/bin/mv -+++ config.base[106]NMV=/bin/mv -+++ config.base[107]export 'NLN=/bin/ln -sf' -+++ config.base[107]NLN='/bin/ln -sf' -+++ config.base[108]export VERBOSE=YES -+++ config.base[108]VERBOSE=YES -+++ config.base[109]export KEEPDATA=NO -+++ config.base[109]KEEPDATA=NO -+++ config.base[110]export DEBUG_POSTSCRIPT=NO -+++ config.base[110]DEBUG_POSTSCRIPT=NO -+++ config.base[111]export CHGRP_RSTPROD=YES -+++ config.base[111]CHGRP_RSTPROD=YES -+++ config.base[112]export 'CHGRP_CMD=chgrp rstprod' -+++ config.base[112]CHGRP_CMD='chgrp rstprod' -+++ config.base[113]export NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[113]NCDUMP=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/netcdf-c-4.9.2-tslbcfy/bin/ncdump -+++ config.base[114]export NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[114]NCLEN=/work2/noaa/global/mterry/global-workflow_forked/ush/getncdimlen -+++ config.base[117]export BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[117]BASE_ENV=/work2/noaa/global/mterry/global-workflow_forked/env -+++ config.base[120]export SDATE=2021032312 -+++ config.base[120]SDATE=2021032312 -+++ config.base[121]export EDATE=2021032312 -+++ config.base[121]EDATE=2021032312 -+++ config.base[122]export EXP_WARM_START=.false. -+++ config.base[122]EXP_WARM_START=.false. -+++ config.base[123]export assim_freq=6 -+++ config.base[123]assim_freq=6 -+++ config.base[124]export PSLOT=C48_S2SW -+++ config.base[124]PSLOT=C48_S2SW -+++ config.base[125]export EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[125]EXPDIR=/work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW -+++ config.base[126]export ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[126]ROTDIR=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW -+++ config.base[127]export DUMP_SUFFIX= -+++ config.base[127]DUMP_SUFFIX= -+++ config.base[128][[ 2021032312 -ge 2019092100 ]] -+++ config.base[128][[ 2021032312 -le 2019110700 ]] -+++ config.base[131]export ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[131]ARCDIR=/work2/noaa/global/mterry/archive/C48_S2SW -+++ config.base[132]export ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[132]ATARDIR=/NCEPDEV/emc-global/1year/mterry/HERCULES/scratch/C48_S2SW -+++ config.base[133]export FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[133]FETCHDIR=/NCEPDEV/emc-global/1year/David.Grumm/test_data -+++ config.base[136]export envir=prod -+++ config.base[136]envir=prod -+++ config.base[137]export NET=gfs -+++ config.base[137]NET=gfs -+++ config.base[138]export RUN=gfs -+++ config.base[138]RUN=gfs -+++ config.base[141]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.com -++++ config.com[4]echo 'BEGIN: config.com' -BEGIN: config.com -++++ config.com[38][[ emc == \n\c\o ]] -++++ config.com[43]COM_OBSPROC_TMPL='${DMPDIR}/${RUN}${DUMP_SUFFIX}.${YMD}/${HH}/atmos' -++++ config.com[44]COM_RTOFS_TMPL='${DMPDIR}' -++++ config.com[45]COM_TCVITAL_TMPL='${DMPDIR}/${RUN}.${YMD}/${HH}/atmos' -++++ config.com[47]declare -rx 'COM_OBS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/obs' -++++ config.com[48]declare -rx COM_OBSPROC_TMPL COM_RTOFS_TMPL -++++ config.com[50]COM_BASE='${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}' -++++ config.com[52]declare -rx 'COM_TOP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}' -++++ config.com[54]declare -rx 'COM_CONF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/conf' -++++ config.com[55]declare -rx 'COM_OBS_JEDI=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/obs_jedi' -++++ config.com[57]declare -rx 'COM_ATMOS_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/input' -++++ config.com[58]declare -rx 'COM_ATMOS_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/restart' -++++ config.com[59]declare -rx 'COM_ATMOS_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/atmos' -++++ config.com[60]declare -rx 'COM_SNOW_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/snow' -++++ config.com[61]declare -rx 'COM_SNOW_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/snow/anlmon' -++++ config.com[62]declare -rx 'COM_ATMOS_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/history' -++++ config.com[63]declare -rx 'COM_ATMOS_MASTER_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/atmos/master' -++++ config.com[64]declare -rx 'COM_ATMOS_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2' -++++ config.com[65]declare -rx 'COM_ATMOS_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/grib2/${GRID}' -++++ config.com[66]declare -rx 'COM_ATMOS_BUFR_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/bufr' -++++ config.com[67]declare -rx 'COM_ATMOS_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/gempak/${GRID}' -++++ config.com[68]declare -rx 'COM_ATMOS_GENESIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/genesis_vital' -++++ config.com[69]declare -rx 'COM_ATMOS_TRACK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/cyclone/tracks' -++++ config.com[70]declare -rx 'COM_ATMOS_GOES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/goes_sim' -++++ config.com[71]declare -rx 'COM_ATMOS_IMAGERY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/imagery' -++++ config.com[72]declare -rx 'COM_ATMOS_OZNMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/oznmon' -++++ config.com[73]declare -rx 'COM_ATMOS_RADMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/radmon' -++++ config.com[74]declare -rx 'COM_ATMOS_MINMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/minmon' -++++ config.com[75]declare -rx 'COM_ATMOS_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/anlmon' -++++ config.com[76]declare -rx 'COM_ATMOS_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/atmos/wmo' -++++ config.com[78]declare -rx 'COM_WAVE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/restart' -++++ config.com[79]declare -rx 'COM_WAVE_PREP_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/prep' -++++ config.com[80]declare -rx 'COM_WAVE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/wave/history' -++++ config.com[81]declare -rx 'COM_WAVE_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded' -++++ config.com[82]declare -rx 'COM_WAVE_GRID_RES_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gridded/${GRDRESNAME}' -++++ config.com[83]declare -rx 'COM_WAVE_STATION_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/station' -++++ config.com[84]declare -rx 'COM_WAVE_GEMPAK_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/gempak' -++++ config.com[85]declare -rx 'COM_WAVE_WMO_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/wave/wmo' -++++ config.com[87]declare -rx 'COM_OCEAN_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/history' -++++ config.com[88]declare -rx 'COM_OCEAN_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/restart' -++++ config.com[89]declare -rx 'COM_OCEAN_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ocean/input' -++++ config.com[90]declare -rx 'COM_OCEAN_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean' -++++ config.com[91]declare -rx 'COM_OCEAN_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/anlmon' -++++ config.com[92]declare -rx 'COM_OCEAN_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ocean/letkf' -++++ config.com[93]declare -rx 'COM_OCEAN_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ocean' -++++ config.com[94]declare -rx 'COM_OCEAN_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/netcdf' -++++ config.com[95]declare -rx 'COM_OCEAN_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2' -++++ config.com[96]declare -rx 'COM_OCEAN_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ocean/grib2/${GRID}' -++++ config.com[98]declare -rx 'COM_ICE_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice' -++++ config.com[99]declare -rx 'COM_ICE_LETKF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/ice/letkf' -++++ config.com[100]declare -rx 'COM_ICE_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/anlmon' -++++ config.com[101]declare -rx 'COM_ICE_BMATRIX_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/bmatrix/ice' -++++ config.com[102]declare -rx 'COM_ICE_INPUT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/input' -++++ config.com[103]declare -rx 'COM_ICE_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/history' -++++ config.com[104]declare -rx 'COM_ICE_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/ice/restart' -++++ config.com[105]declare -rx 'COM_ICE_NETCDF_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/netcdf' -++++ config.com[106]declare -rx 'COM_ICE_GRIB_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2' -++++ config.com[107]declare -rx 'COM_ICE_GRIB_GRID_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/ice/grib2/${GRID}' -++++ config.com[109]declare -rx 'COM_CHEM_HISTORY_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/chem/history' -++++ config.com[110]declare -rx 'COM_CHEM_ANALYSIS_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem' -++++ config.com[111]declare -rx 'COM_CHEM_BMAT_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/analysis/chem/bmatrix' -++++ config.com[112]declare -rx 'COM_CHEM_ANLMON_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/products/chem/anlmon' -++++ config.com[114]declare -rx 'COM_MED_RESTART_TMPL=${ROTDIR}/${RUN}.${YMD}/${HH}/${MEMDIR}/model/med/restart' -+++ config.base[143]export LOGSCRIPT= -+++ config.base[143]LOGSCRIPT= -+++ config.base[145]export 'REDOUT=1>' -+++ config.base[145]REDOUT='1>' -+++ config.base[146]export 'REDERR=2>' -+++ config.base[146]REDERR='2>' -+++ config.base[148]export SENDECF=NO -+++ config.base[148]SENDECF=NO -+++ config.base[149]export SENDSDM=NO -+++ config.base[149]SENDSDM=NO -+++ config.base[150]export SENDDBN_NTC=NO -+++ config.base[150]SENDDBN_NTC=NO -+++ config.base[151]export SENDDBN=NO -+++ config.base[151]SENDDBN=NO -+++ config.base[152]export DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[152]DBNROOT=/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/fakedbn -+++ config.base[153]export SENDAWIP=NO -+++ config.base[153]SENDAWIP=NO -+++ config.base[156]export APP=S2SW -+++ config.base[156]APP=S2SW -+++ config.base[158]shopt -s extglob -+++ config.base[161]case "${RUN}" in -+++ config.base[168]shopt -u extglob -+++ config.base[171]export DO_ATM=YES -+++ config.base[171]DO_ATM=YES -+++ config.base[172]export DO_COUPLED=NO -+++ config.base[172]DO_COUPLED=NO -+++ config.base[173]export DO_WAVE=NO -+++ config.base[173]DO_WAVE=NO -+++ config.base[174]export DO_OCN=NO -+++ config.base[174]DO_OCN=NO -+++ config.base[175]export DO_ICE=NO -+++ config.base[175]DO_ICE=NO -+++ config.base[176]DO_AERO=NO -+++ config.base[177]export DO_PREP_OBS_AERO=NO -+++ config.base[177]DO_PREP_OBS_AERO=NO -+++ config.base[178]aero_fcst_runs=gdas -+++ config.base[179]aero_anl_runs='gdas gfs' -+++ config.base[180]export DO_AERO_FCST=NO -+++ config.base[180]DO_AERO_FCST=NO -+++ config.base[181]export DO_AERO_ANL=NO -+++ config.base[181]DO_AERO_ANL=NO -+++ config.base[182]export DOBNDPNT_WAVE=YES -+++ config.base[182]DOBNDPNT_WAVE=YES -+++ config.base[183]export DOIBP_WAV=NO -+++ config.base[183]DOIBP_WAV=NO -+++ config.base[184]export FRAC_GRID=.true. -+++ config.base[184]FRAC_GRID=.true. -+++ config.base[185]export DO_NEST=NO -+++ config.base[185]DO_NEST=NO -+++ config.base[186][[ NO == \Y\E\S ]] -+++ config.base[192]export ntiles=6 -+++ config.base[192]ntiles=6 -+++ config.base[193]export FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[193]FIXugwd=/work2/noaa/global/mterry/global-workflow_forked/fix/ugwd -+++ config.base[194]export FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[194]FIXorog=/work2/noaa/global/mterry/global-workflow_forked/fix/orog -+++ config.base[198]export OPS_RES=C768 -+++ config.base[198]OPS_RES=C768 -+++ config.base[201]export LEVS=128 -+++ config.base[201]LEVS=128 -+++ config.base[202]export CASE=C48 -+++ config.base[202]CASE=C48 -+++ config.base[203]export 'CASE_ENS={{ CASE_ENS }}' -+++ config.base[203]CASE_ENS='{{ CASE_ENS }}' -+++ config.base[204]export OCNRES=500 -+++ config.base[204]OCNRES=500 -+++ config.base[205]export ICERES=500 -+++ config.base[205]ICERES=500 -+++ config.base[208]case "${CASE}" in -+++ config.base[210]export waveGRD=uglo_100km -+++ config.base[210]waveGRD=uglo_100km -+++ config.base[227]case "${APP}" in -+++ config.base[243]export DO_COUPLED=YES -+++ config.base[243]DO_COUPLED=YES -+++ config.base[244]export DO_OCN=YES -+++ config.base[244]DO_OCN=YES -+++ config.base[245]export DO_ICE=YES -+++ config.base[245]DO_ICE=YES -+++ config.base[247][[ S2SW =~ A$ ]] -+++ config.base[251][[ S2SW =~ ^S2SW ]] -+++ config.base[252]export DO_WAVE=YES -+++ config.base[252]DO_WAVE=YES -+++ config.base[262][[ NO == \Y\E\S ]] -+++ config.base[272][[ gfs =~ gdas ]] -+++ config.base[275][[ gfs =~ gfs ]] -+++ config.base[276]export FHCYC=24 -+++ config.base[276]FHCYC=24 -+++ config.base[280]export FHMIN=0 -+++ config.base[280]FHMIN=0 -+++ config.base[281]export FHMAX=9 -+++ config.base[281]FHMAX=9 -+++ config.base[282]export FHOUT=3 -+++ config.base[282]FHOUT=3 -+++ config.base[283]export FHOUT_OCN=3 -+++ config.base[283]FHOUT_OCN=3 -+++ config.base[284]export FHOUT_ICE=3 -+++ config.base[284]FHOUT_ICE=3 -+++ config.base[285]export FHOUT_AERO=3 -+++ config.base[285]FHOUT_AERO=3 -+++ config.base[288]export EUPD_CYC=gdas -+++ config.base[288]EUPD_CYC=gdas -+++ config.base[291]export INTERVAL_GFS=6 -+++ config.base[291]INTERVAL_GFS=6 -+++ config.base[292]export SDATE_GFS=2021032312 -+++ config.base[292]SDATE_GFS=2021032312 -+++ config.base[295]export FHMIN_GFS=0 -+++ config.base[295]FHMIN_GFS=0 -+++ config.base[296]export FHMAX_GFS=120 -+++ config.base[296]FHMAX_GFS=120 -+++ config.base[298]breakpnts= -+++ config.base[299]export FCST_SEGMENTS=0,120 -+++ config.base[299]FCST_SEGMENTS=0,120 -+++ config.base[300]export FHOUT_GFS=3 -+++ config.base[300]FHOUT_GFS=3 -+++ config.base[301]export FHMAX_HF_GFS=48 -+++ config.base[301]FHMAX_HF_GFS=48 -+++ config.base[302]export FHMAX_HF_GFS=48 -+++ config.base[302]FHMAX_HF_GFS=48 -+++ config.base[303]export FHOUT_HF_GFS=1 -+++ config.base[303]FHOUT_HF_GFS=1 -+++ config.base[306]export FHMIN_WAV=0 -+++ config.base[306]FHMIN_WAV=0 -+++ config.base[307]export FHOUT_WAV=1 -+++ config.base[307]FHOUT_WAV=1 -+++ config.base[308]export FHMAX_WAV=9 -+++ config.base[308]FHMAX_WAV=9 -+++ config.base[309]export FHMAX_WAV=9 -+++ config.base[309]FHMAX_WAV=9 -+++ config.base[310]export FHOUT_WAV_GFS=3 -+++ config.base[310]FHOUT_WAV_GFS=3 -+++ config.base[311]export FHMAX_WAV_GFS=120 -+++ config.base[311]FHMAX_WAV_GFS=120 -+++ config.base[312]export FHOUT_HF_WAV=1 -+++ config.base[312]FHOUT_HF_WAV=1 -+++ config.base[313]export FHMAX_HF_WAV=48 -+++ config.base[313]FHMAX_HF_WAV=48 -+++ config.base[314]export FHMAX_HF_WAV=48 -+++ config.base[314]FHMAX_HF_WAV=48 -+++ config.base[317]export FHOUT_OCN_GFS=6 -+++ config.base[317]FHOUT_OCN_GFS=6 -+++ config.base[318]export FHOUT_ICE_GFS=6 -+++ config.base[318]FHOUT_ICE_GFS=6 -+++ config.base[321]export ILPOST=1 -+++ config.base[321]ILPOST=1 -+++ config.base[322](( FHMAX_HF_GFS < 120 )) -+++ config.base[323]export ILPOST=3 -+++ config.base[323]ILPOST=3 -+++ config.base[327]export FHMAX_GOES=180 -+++ config.base[327]FHMAX_GOES=180 -+++ config.base[328]export FHOUT_GOES=3 -+++ config.base[328]FHOUT_GOES=3 -+++ config.base[329](( FHMAX_GOES > FHMAX_GFS )) -+++ config.base[330]export FHMAX_GOES=120 -+++ config.base[330]FHMAX_GOES=120 -+++ config.base[334]export restart_interval_gfs=12 -+++ config.base[334]restart_interval_gfs=12 -+++ config.base[339]export QUILTING=.true. -+++ config.base[339]QUILTING=.true. -+++ config.base[340]export OUTPUT_GRID=gaussian_grid -+++ config.base[340]OUTPUT_GRID=gaussian_grid -+++ config.base[341]export WRITE_DOPOST=.true. -+++ config.base[341]WRITE_DOPOST=.true. -+++ config.base[342]export WRITE_NSFLIP=.true. -+++ config.base[342]WRITE_NSFLIP=.true. -+++ config.base[345]export DOIAU=YES -+++ config.base[345]DOIAU=YES -+++ config.base[346]export IAUFHRS=3,6,9 -+++ config.base[346]IAUFHRS=3,6,9 -+++ config.base[347]export IAU_FHROT=3 -+++ config.base[347]IAU_FHROT=3 -+++ config.base[348]export IAU_DELTHRS=6 -+++ config.base[348]IAU_DELTHRS=6 -+++ config.base[349]export IAU_OFFSET=6 -+++ config.base[349]IAU_OFFSET=6 -+++ config.base[350]export DOIAU_ENKF=YES -+++ config.base[350]DOIAU_ENKF=YES -+++ config.base[351]export IAUFHRS_ENKF=3,6,9 -+++ config.base[351]IAUFHRS_ENKF=3,6,9 -+++ config.base[352]export IAU_DELTHRS_ENKF=6 -+++ config.base[352]IAU_DELTHRS_ENKF=6 -+++ config.base[355]export lobsdiag_forenkf=.true. -+++ config.base[355]lobsdiag_forenkf=.true. -+++ config.base[363]export imp_physics=8 -+++ config.base[363]imp_physics=8 -+++ config.base[367]export DO_JEDIATMVAR=NO -+++ config.base[367]DO_JEDIATMVAR=NO -+++ config.base[368]export DO_JEDIATMENS=NO -+++ config.base[368]DO_JEDIATMENS=NO -+++ config.base[369]export DO_JEDIOCNVAR=NO -+++ config.base[369]DO_JEDIOCNVAR=NO -+++ config.base[370]export DO_JEDISNOWDA=NO -+++ config.base[370]DO_JEDISNOWDA=NO -+++ config.base[371]export DO_MERGENSST=NO -+++ config.base[371]DO_MERGENSST=NO -+++ config.base[372]export DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[372]DO_STARTMEM_FROM_JEDIICE=NO -+++ config.base[375]export 'DOHYBVAR={{ DOHYBVAR }}' -+++ config.base[375]DOHYBVAR='{{ DOHYBVAR }}' -+++ config.base[376]export DOHYBVAR_OCN=NO -+++ config.base[376]DOHYBVAR_OCN=NO -+++ config.base[377]export DOLETKF_OCN=NO -+++ config.base[377]DOLETKF_OCN=NO -+++ config.base[378]export NMEM_ENS=0 -+++ config.base[378]NMEM_ENS=0 -+++ config.base[379]export SMOOTH_ENKF=NO -+++ config.base[379]SMOOTH_ENKF=NO -+++ config.base[380]export l4densvar=.true. -+++ config.base[380]l4densvar=.true. -+++ config.base[381]export lwrite4danl=.true. -+++ config.base[381]lwrite4danl=.true. -+++ config.base[382]export DO_CALC_INCREMENT=NO -+++ config.base[382]DO_CALC_INCREMENT=NO -+++ config.base[385]export NMEM_ENS_GFS=30 -+++ config.base[385]NMEM_ENS_GFS=30 -+++ config.base[386]export NMEM_ENS_GFS_OFFSET=20 -+++ config.base[386]NMEM_ENS_GFS_OFFSET=20 -+++ config.base[387]export DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[387]DO_CALC_INCREMENT_ENKF_GFS=NO -+++ config.base[390][[ {{ DOHYBVAR }} = \Y\E\S ]] -+++ config.base[404][[ {{ DOHYBVAR }} == \N\O ]] -+++ config.base[412]export ENKF_SPREAD=YES -+++ config.base[412]ENKF_SPREAD=YES -+++ config.base[415]export DO_GSISOILDA=NO -+++ config.base[415]DO_GSISOILDA=NO -+++ config.base[416]export DO_LAND_IAU=.false. -+++ config.base[416]DO_LAND_IAU=.false. -+++ config.base[417]export LSOIL_INCR=2 -+++ config.base[417]LSOIL_INCR=2 -+++ config.base[420][[ forecast-only = \c\y\c\l\e\d ]] -+++ config.base[420][[ YES = \N\O ]] -+++ config.base[420][[ forecast-only = \f\o\r\e\c\a\s\t\-\o\n\l\y ]] -+++ config.base[420][[ .false. = \.\f\a\l\s\e\. ]] -+++ config.base[421]export IAU_OFFSET=0 -+++ config.base[421]IAU_OFFSET=0 -+++ config.base[422]export IAU_FHROT=0 -+++ config.base[422]IAU_FHROT=0 -+++ config.base[423]export IAUFHRS=6, -+++ config.base[423]IAUFHRS=6, -+++ config.base[424]export DO_LAND_IAU=.false. -+++ config.base[424]DO_LAND_IAU=.false. -+++ config.base[427][[ YES = \N\O ]] -+++ config.base[431][[ YES == \Y\E\S ]] -+++ config.base[432]export restart_interval_enkfgdas=3 -+++ config.base[432]restart_interval_enkfgdas=3 -+++ config.base[437]export restart_interval_enkfgfs=3 -+++ config.base[437]restart_interval_enkfgfs=3 -+++ config.base[439][[ YES == \Y\E\S ]] -+++ config.base[440]export restart_interval_gdas=3 -+++ config.base[440]restart_interval_gdas=3 -+++ config.base[446]export DONST=YES -+++ config.base[446]DONST=YES -+++ config.base[447][[ YES = \Y\E\S ]] -+++ config.base[447]export 'FNTSFA= ' -+++ config.base[447]FNTSFA=' ' -+++ config.base[450]export nst_anl=.true. -+++ config.base[450]nst_anl=.true. -+++ config.base[453]export MAKE_NSSTBUFR=NO -+++ config.base[453]MAKE_NSSTBUFR=NO -+++ config.base[456]export MAKE_ACFTBUFR=NO -+++ config.base[456]MAKE_ACFTBUFR=NO -+++ config.base[459]export 'INCREMENTS_TO_ZERO='\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[459]INCREMENTS_TO_ZERO=''\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]export 'INCVARS_ZERO_STRAT='\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[462]INCVARS_ZERO_STRAT=''\''sphum_inc'\'','\''liq_wat_inc'\'','\''icmr_inc'\'','\''rwmr_inc'\'','\''snmr_inc'\'','\''grle_inc'\''' -+++ config.base[463]export INCVARS_EFOLD=5 -+++ config.base[463]INCVARS_EFOLD=5 -+++ config.base[468]export netcdf_diag=.true. -+++ config.base[468]netcdf_diag=.true. -+++ config.base[469]export binary_diag=.false. -+++ config.base[469]binary_diag=.false. -+++ config.base[472]export DO_CA=YES -+++ config.base[472]DO_CA=YES -+++ config.base[475]export DO_METP=NO -+++ config.base[475]DO_METP=NO -+++ config.base[476]export DO_FIT2OBS=YES -+++ config.base[476]DO_FIT2OBS=YES -+++ config.base[479]export FHMAX_FITS=132 -+++ config.base[479]FHMAX_FITS=132 -+++ config.base[480][[ 132 -gt 120 ]] -+++ config.base[481]export FHMAX_FITS=120 -+++ config.base[481]FHMAX_FITS=120 -+++ config.base[486]export DO_FETCH_HPSS=NO -+++ config.base[486]DO_FETCH_HPSS=NO -+++ config.base[487]export DO_FETCH_LOCAL=NO -+++ config.base[487]DO_FETCH_LOCAL=NO -+++ config.base[490]export DO_ARCHCOM=NO -+++ config.base[490]DO_ARCHCOM=NO -+++ config.base[491]export ARCHCOM_TO=globus_hpss -+++ config.base[491]ARCHCOM_TO=globus_hpss -+++ config.base[494]export CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[494]CLIENT_GLOBUS_UUID=869912fe-f6de-46c0-af10-b22efd84a022 -+++ config.base[497][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[504][[ NO = \Y\E\S ]] -+++ config.base[510]export REPLAY_ICS=NO -+++ config.base[510]REPLAY_ICS=NO -+++ config.base[511]export OFFSET_START_HOUR=0 -+++ config.base[511]OFFSET_START_HOUR=0 -+++ config.base[514]export NUM_SND_COLLECTIVES=9 -+++ config.base[514]NUM_SND_COLLECTIVES=9 -+++ config.base[516]echo 'END: config.base' -END: config.base -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wave -+++ config.wave[6]echo 'BEGIN: config.wave' -BEGIN: config.wave -+++ config.wave[13]export RUNRSTwave=gdas -+++ config.wave[13]RUNRSTwave=gdas -+++ config.wave[16]export MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[16]MESH_WAV=mesh.uglo_100km.nc -+++ config.wave[19]case "${waveGRD}" in -+++ config.wave[64]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[64]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+++ config.wave[65]export wavepostGRD= -+++ config.wave[65]wavepostGRD= -+++ config.wave[66]export waveuoutpGRD=uglo_100km -+++ config.wave[66]waveuoutpGRD=uglo_100km -+++ config.wave[75]export WAVEWND_DID= -+++ config.wave[75]WAVEWND_DID= -+++ config.wave[76]export WAVEWND_FID= -+++ config.wave[76]WAVEWND_FID= -+++ config.wave[79][[ gfs == \g\f\s ]] -+++ config.wave[80]export FHMAX_WAV=120 -+++ config.wave[80]FHMAX_WAV=120 -+++ config.wave[82]export WAVHINDH=0 -+++ config.wave[82]WAVHINDH=0 -+++ config.wave[83]export FHMAX_WAV_IBP=180 -+++ config.wave[83]FHMAX_WAV_IBP=180 -+++ config.wave[84](( FHMAX_WAV < FHMAX_WAV_IBP )) -+++ config.wave[84]export FHMAX_WAV_IBP=120 -+++ config.wave[84]FHMAX_WAV_IBP=120 -+++ config.wave[87]export DTFLD_WAV=3600 -+++ config.wave[87]DTFLD_WAV=3600 -+++ config.wave[88]export DTPNT_WAV=3600 -+++ config.wave[88]DTPNT_WAV=3600 -+++ config.wave[89]export FHINCP_WAV=1 -+++ config.wave[89]FHINCP_WAV=1 -+++ config.wave[92]export 'OUTPARS_WAV=WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[92]OUTPARS_WAV='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -+++ config.wave[95][[ gfs == \g\d\a\s ]] -+++ config.wave[99](( INTERVAL_GFS > 0 )) -+++ config.wave[100]export WAVHCYC=6 -+++ config.wave[100]WAVHCYC=6 -+++ config.wave[101]export FHMAX_WAV_CUR=192 -+++ config.wave[101]FHMAX_WAV_CUR=192 -+++ config.wave[109]export RSTTYPE_WAV=T -+++ config.wave[109]RSTTYPE_WAV=T -+++ config.wave[110][[ gfs != gfs ]] -+++ config.wave[115]rst_dt_gfs=43200 -+++ config.wave[116][[ 43200 -gt 0 ]] -+++ config.wave[117]export DT_1_RST_WAV=0 -+++ config.wave[117]DT_1_RST_WAV=0 -+++ config.wave[120]export DT_2_RST_WAV=43200 -+++ config.wave[120]DT_2_RST_WAV=43200 -+++ config.wave[126]export RSTIOFF_WAV=0 -+++ config.wave[126]RSTIOFF_WAV=0 -+++ config.wave[131]export RUNMEM=-1 -+++ config.wave[131]RUNMEM=-1 -+++ config.wave[134](( RUNMEM == -1 )) -+++ config.wave[136]export waveMEMB= -+++ config.wave[136]waveMEMB= -+++ config.wave[143]export WW3ATMINP=CPL -+++ config.wave[143]WW3ATMINP=CPL -+++ config.wave[144][[ YES == \Y\E\S ]] -+++ config.wave[145]export WW3ICEINP=CPL -+++ config.wave[145]WW3ICEINP=CPL -+++ config.wave[146]export WAVEICE_FID= -+++ config.wave[146]WAVEICE_FID= -+++ config.wave[152][[ YES == \Y\E\S ]] -+++ config.wave[153]export WW3CURINP=CPL -+++ config.wave[153]WW3CURINP=CPL -+++ config.wave[154]export WAVECUR_FID= -+++ config.wave[154]WAVECUR_FID= -+++ config.wave[161]export WW3ATMIENS=F -+++ config.wave[161]WW3ATMIENS=F -+++ config.wave[162]export WW3ICEIENS=F -+++ config.wave[162]WW3ICEIENS=F -+++ config.wave[163]export WW3CURIENS=F -+++ config.wave[163]WW3CURIENS=F -+++ config.wave[165]export GOFILETYPE=1 -+++ config.wave[165]GOFILETYPE=1 -+++ config.wave[166]export POFILETYPE=1 -+++ config.wave[166]POFILETYPE=1 -+++ config.wave[170]export FUNIPNT=T -+++ config.wave[170]FUNIPNT=T -+++ config.wave[172]export IOSRV=1 -+++ config.wave[172]IOSRV=1 -+++ config.wave[174]export FPNTPROC=T -+++ config.wave[174]FPNTPROC=T -+++ config.wave[176]export FGRDPROC=F -+++ config.wave[176]FGRDPROC=F -+++ config.wave[178]export FLAGMASKCOMP=F -+++ config.wave[178]FLAGMASKCOMP=F -+++ config.wave[180]export FLAGMASKOUT=F -+++ config.wave[180]FLAGMASKOUT=F -+++ config.wave[182]echo 'END: config.wave' -END: config.wave -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[105]for config in "${configs[@]:-''}" -++ jjob_header.sh[106]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.wavepostsbs -+++ config.wavepostsbs[6]echo 'BEGIN: config.wavepostsbs' -BEGIN: config.wavepostsbs -+++ config.wavepostsbs[9]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources wavepostsbs -++++ config.resources[10](( 1 != 1 )) -++++ config.resources[34]step=wavepostsbs -++++ config.resources[36]echo 'BEGIN: config.resources' -BEGIN: config.resources -++++ config.resources[38]case ${machine} in -++++ config.resources[61]max_tasks_per_node=80 -++++ config.resources[62]mem_node_max=500GB -++++ config.resources[96]export max_tasks_per_node -++++ config.resources[98]case ${step} in -++++ config.resources[156]ntasks=8 -++++ config.resources[157]threads_per_task=1 -++++ config.resources[158]tasks_per_node=80 -++++ config.resources[159]NTASKS=8 -++++ config.resources[160]memory=20GB -++++ config.resources[161]walltime=00:15:00 -++++ config.resources[1398][[ -f /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES ]] -++++ config.resources[1399]source /work2/noaa/global/mterry/RUNTESTS/EXPDIR/C48_S2SW/config.resources.HERCULES -+++++ config.resources.HERCULES[6]case ${step} in -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=threads_per_task_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export threads_per_task -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=ntasks_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export ntasks -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=tasks_per_node_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export tasks_per_node -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=NTASKS_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export NTASKS -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=memory_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export memory -++++ config.resources[1403]for resource_var in threads_per_task ntasks tasks_per_node NTASKS memory walltime -++++ config.resources[1404]run_resource_var=walltime_gfs -++++ config.resources[1405][[ -n '' ]] -++++ config.resources[1407][[ -n 0 ]] -++++ config.resources[1408]export walltime -++++ config.resources[1412]echo 'END: config.resources' -END: config.resources -+++ config.wavepostsbs[12]export MAX_TASKS=25 -+++ config.wavepostsbs[12]MAX_TASKS=25 -+++ config.wavepostsbs[15]export WAV_SUBGRBSRC= -+++ config.wavepostsbs[15]WAV_SUBGRBSRC= -+++ config.wavepostsbs[16]export WAV_SUBGRB= -+++ config.wavepostsbs[16]WAV_SUBGRB= -+++ config.wavepostsbs[19]export DOFLD_WAV=YES -+++ config.wavepostsbs[19]DOFLD_WAV=YES -+++ config.wavepostsbs[20]export DOPNT_WAV=YES -+++ config.wavepostsbs[20]DOPNT_WAV=YES -+++ config.wavepostsbs[21]export DOGRB_WAV=YES -+++ config.wavepostsbs[21]DOGRB_WAV=YES -+++ config.wavepostsbs[22][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+++ config.wavepostsbs[23]export DOGRI_WAV=YES -+++ config.wavepostsbs[23]DOGRI_WAV=YES -+++ config.wavepostsbs[27]export DOSPC_WAV=YES -+++ config.wavepostsbs[27]DOSPC_WAV=YES -+++ config.wavepostsbs[28]export DOBLL_WAV=YES -+++ config.wavepostsbs[28]DOBLL_WAV=YES -+++ config.wavepostsbs[30]echo 'END: config.wavepostsbs' -END: config.wavepostsbs -++ jjob_header.sh[106]true -++ jjob_header.sh[107]export err=0 -++ jjob_header.sh[107]err=0 -++ jjob_header.sh[108][[ 0 -ne 0 ]] -++ jjob_header.sh[117]source /work2/noaa/global/mterry/global-workflow_forked/env/HERCULES.env wavepostsbs -+++ HERCULES.env[3][[ 1 -ne 1 ]] -+++ HERCULES.env[10]step=wavepostsbs -+++ HERCULES.env[12]export 'launcher=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[12]launcher='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[13]export 'mpmd_opt=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[13]mpmd_opt='--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[16]export MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[16]MPI_BUFS_PER_PROC=2048 -+++ HERCULES.env[17]export MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[17]MPI_BUFS_PER_HOST=2048 -+++ HERCULES.env[18]export MPI_GROUP_MAX=256 -+++ HERCULES.env[18]MPI_GROUP_MAX=256 -+++ HERCULES.env[19]export MPI_MEMMAP_OFF=1 -+++ HERCULES.env[19]MPI_MEMMAP_OFF=1 -+++ HERCULES.env[20]export MP_STDOUTMODE=ORDERED -+++ HERCULES.env[20]MP_STDOUTMODE=ORDERED -+++ HERCULES.env[21]export KMP_AFFINITY=scatter -+++ HERCULES.env[21]KMP_AFFINITY=scatter -+++ HERCULES.env[22]export OMP_STACKSIZE=2048000 -+++ HERCULES.env[22]OMP_STACKSIZE=2048000 -+++ HERCULES.env[23]export NTHSTACK=1024000000 -+++ HERCULES.env[23]NTHSTACK=1024000000 -+++ HERCULES.env[25]export I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[25]I_MPI_EXTRA_FILESYSTEM=1 -+++ HERCULES.env[26]export I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[26]I_MPI_EXTRA_FILESYSTEM_LIST=lustre -+++ HERCULES.env[28]ulimit -s unlimited -+++ HERCULES.env[29]ulimit -a -real-time non-blocking time (microseconds, -R) unlimited -core file size (blocks, -c) 0 -data seg size (kbytes, -d) unlimited -scheduling priority (-e) 0 -file size (blocks, -f) unlimited -pending signals (-i) 2049614 -max locked memory (kbytes, -l) unlimited -max memory size (kbytes, -m) 20971520 -open files (-n) 131072 -pipe size (512 bytes, -p) 8 -POSIX message queues (bytes, -q) 819200 -real-time priority (-r) 0 -stack size (kbytes, -s) unlimited -cpu time (seconds, -t) unlimited -max user processes (-u) 1028698 -virtual memory (kbytes, -v) unlimited -file locks (-x) unlimited -+++ HERCULES.env[33][[ -n 8 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[33][[ -n 80 ]] -+++ HERCULES.env[34]max_threads_per_task=1 -+++ HERCULES.env[35]NTHREADSmax=1 -+++ HERCULES.env[36]NTHREADS1=1 -+++ HERCULES.env[37][[ 1 -gt 1 ]] -+++ HERCULES.env[40][[ 1 -gt 1 ]] -+++ HERCULES.env[43]APRUN_default='srun -l --export=ALL --hint=nomultithread -n 8' -+++ HERCULES.env[49]case ${step} in -+++ HERCULES.env[63]export USE_CFP=YES -+++ HERCULES.env[63]USE_CFP=YES -+++ HERCULES.env[64][[ wavepostsbs == \w\a\v\e\p\r\e\p ]] -+++ HERCULES.env[65]export 'wavempexec=srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[65]wavempexec='srun -l --export=ALL --hint=nomultithread' -+++ HERCULES.env[66]export 'wave_mpmd=--multi-prog --output=mpmd.%j.%t.out' -+++ HERCULES.env[66]wave_mpmd='--multi-prog --output=mpmd.%j.%t.out' -++ jjob_header.sh[117]true -++ jjob_header.sh[118]export err=0 -++ jjob_header.sh[118]err=0 -++ jjob_header.sh[119][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[4]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ JGLOBAL_WAVE_POST_SBS[7]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[7]HH=12 -+ JGLOBAL_WAVE_POST_SBS[7]declare_from_tmpl -rx COMIN_WAVE_PREP:COM_WAVE_PREP_TMPL COMIN_WAVE_HISTORY:COM_WAVE_HISTORY_TMPL COMOUT_WAVE_GRID:COM_WAVE_GRID_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMIN_WAVE_PREP=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep -declare_from_tmpl :: COMIN_WAVE_HISTORY=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history -declare_from_tmpl :: COMOUT_WAVE_GRID=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded -+ JGLOBAL_WAVE_POST_SBS[12][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n '' ]] -+ JGLOBAL_WAVE_POST_SBS[14][[ -n at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_atlocn_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_atlocn_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_atlocn_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_epacif_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_epacif_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_epacif_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_wcoast_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_wcoast_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_wcoast_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p50 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p50:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p50=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p50 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=arctic.9km -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_arctic_9km:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_arctic_9km=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_arctic_9km -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=global.0p16 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_global_0p16:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_global_0p16=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_global_0p16 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 ]] -+ JGLOBAL_WAVE_POST_SBS[15]for grdID in ${wavepostGRD} ${waveinterpGRD} -+ JGLOBAL_WAVE_POST_SBS[16]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ JGLOBAL_WAVE_POST_SBS[17]YMD=20210323 -+ JGLOBAL_WAVE_POST_SBS[17]HH=12 -+ JGLOBAL_WAVE_POST_SBS[17]GRDRESNAME=gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[17]declare_from_tmpl -rx COMOUT_WAVE_GRID_gsouth_0p25:COM_WAVE_GRID_RES_TMPL -+ bash_utils.sh[39][[ NO == \N\O ]] -+ bash_utils.sh[39]set +x -declare_from_tmpl :: COMOUT_WAVE_GRID_gsouth_0p25=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[19]out_dir_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ JGLOBAL_WAVE_POST_SBS[20]out_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ JGLOBAL_WAVE_POST_SBS[21][[ ! -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 ]] -+ JGLOBAL_WAVE_POST_SBS[28]/work2/noaa/global/mterry/global-workflow_forked/scripts/exgfs_wave_post_gridded_sbs.sh -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ exgfs_wave_post_gridded_sbs.sh[24]source /work2/noaa/global/mterry/global-workflow_forked/ush/wave_domain_grid.sh -+ exgfs_wave_post_gridded_sbs.sh[26]DOGRI_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[27]DOGRB_WAV=YES -+ exgfs_wave_post_gridded_sbs.sh[29]export waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[29]waveGRD=uglo_100km -+ exgfs_wave_post_gridded_sbs.sh[30]export 'waveinterpGRD=at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[30]waveinterpGRD='at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m' -+ exgfs_wave_post_gridded_sbs.sh[31]export wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[31]wavepostGRD= -+ exgfs_wave_post_gridded_sbs.sh[33]cat - INFO: Grid information: - INFO: Native wave grids: uglo_100km - INFO: Interpolated grids: at_10m ep_10m wc_10m glo_30m aoc_9km gnh_10m gsh_15m - INFO: Post-process grids: -++ exgfs_wave_post_gridded_sbs.sh[40]printf %03i 108 -+ exgfs_wave_post_gridded_sbs.sh[40]fhr3=108 -++ exgfs_wave_post_gridded_sbs.sh[41]date -u -d '20210323 12 + 108 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[41]valid_time=2021032800 -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.uglo_100km.bin mod_def.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.at_10m.bin mod_def.at_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.ep_10m.bin mod_def.ep_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.wc_10m.bin mod_def.wc_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.glo_30m.bin mod_def.glo_30m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.aoc_9km.bin mod_def.aoc_9km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gnh_10m.bin mod_def.gnh_10m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[44]for grdID in ${waveGRD} ${wavepostGRD} ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[45]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/prep/gfs.wave.t12z.mod_def.gsh_15m.bin mod_def.gsh_15m -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[49]cpreq /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f108.bin ./out_grd.uglo_100km -+ cpreq[3]cp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//model/wave/history/gfs.wave.t12z.uglo_100km.f108.bin ./out_grd.uglo_100km -+ cpreq[4]'[' 0 -ne 0 ']' -+ exgfs_wave_post_gridded_sbs.sh[52][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[53]for grbGRD in ${waveinterpGRD} ${wavepostGRD} -+ exgfs_wave_post_gridded_sbs.sh[54][[ ! -f /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ]] -+ exgfs_wave_post_gridded_sbs.sh[62]cat - INFO: Summary: - INFO: Grid interp: DOGRI_WAV="YES" - INFO: Grib files: DOGRB_WAV="YES" - INFO: Fields to be included in grib files: - INFO: OUTPARS_WAV="WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA" -+ exgfs_wave_post_gridded_sbs.sh[70][[ YES == \N\O ]] -+ exgfs_wave_post_gridded_sbs.sh[76]rm -f 'cmdfile.*' cmdfile -+ exgfs_wave_post_gridded_sbs.sh[77]count=0 -+ exgfs_wave_post_gridded_sbs.sh[80][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[81]dt_int=3600. -+ exgfs_wave_post_gridded_sbs.sh[82]n_int=9999 -++ exgfs_wave_post_gridded_sbs.sh[83]date -u -d '20210328 00 - 0 hours' +%Y%m%d%H -+ exgfs_wave_post_gridded_sbs.sh[83]ymdh_int=2021032800 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=1 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032800 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID at_10m -+ wave_domain_grid.sh[13]grdID=at_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[23]GRDREGION=atlocn -+ wave_domain_grid.sh[23]GRDRES=0p16 -+ wave_domain_grid.sh[23]GRIDNR=255 -+ wave_domain_grid.sh[23]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=atlocn.0p16 -grdNAME=atlocn.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032800 108 atlocn 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib2_at_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib2_at_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/cmdfile.1 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=2 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032800 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID ep_10m -+ wave_domain_grid.sh[13]grdID=ep_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[24]GRDREGION=epacif -+ wave_domain_grid.sh[24]GRDRES=0p16 -+ wave_domain_grid.sh[24]GRIDNR=255 -+ wave_domain_grid.sh[24]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=epacif.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=epacif.0p16 -grdNAME=epacif.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032800 108 epacif 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib2_ep_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib2_ep_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/cmdfile.2 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=3 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032800 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID wc_10m -+ wave_domain_grid.sh[13]grdID=wc_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[25]GRDREGION=wcoast -+ wave_domain_grid.sh[25]GRDRES=0p16 -+ wave_domain_grid.sh[25]GRIDNR=255 -+ wave_domain_grid.sh[25]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=wcoast.0p16 -grdNAME=wcoast.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032800 108 wcoast 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib2_wc_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib2_wc_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/cmdfile.3 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=4 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032800 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID glo_30m -+ wave_domain_grid.sh[13]grdID=glo_30m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[18]GRDREGION=global -+ wave_domain_grid.sh[18]GRDRES=0p50 -+ wave_domain_grid.sh[18]GRIDNR=255 -+ wave_domain_grid.sh[18]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p50 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p50 -grdNAME=global.0p50 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032800 108 global 0p50 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib2_glo_30m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib2_glo_30m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/cmdfile.4 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=5 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032800 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID aoc_9km -+ wave_domain_grid.sh[13]grdID=aoc_9km -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[27]GRDREGION=arctic -+ wave_domain_grid.sh[27]GRDRES=9km -+ wave_domain_grid.sh[27]GRIDNR=255 -+ wave_domain_grid.sh[27]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=arctic.9km -+ wave_domain_grid.sh[40]echo grdNAME=arctic.9km -grdNAME=arctic.9km -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032800 108 arctic 9km '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib2_aoc_9km.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib2_aoc_9km.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/cmdfile.5 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=6 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032800 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gnh_10m -+ wave_domain_grid.sh[13]grdID=gnh_10m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[29]GRDREGION=global -+ wave_domain_grid.sh[29]GRDRES=0p16 -+ wave_domain_grid.sh[29]GRIDNR=255 -+ wave_domain_grid.sh[29]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=global.0p16 -+ wave_domain_grid.sh[40]echo grdNAME=global.0p16 -grdNAME=global.0p16 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032800 108 global 0p16 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib2_gnh_10m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib2_gnh_10m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/cmdfile.6 -+ exgfs_wave_post_gridded_sbs.sh[84]for grdID in ${waveinterpGRD} -+ exgfs_wave_post_gridded_sbs.sh[85]count=7 -+ exgfs_wave_post_gridded_sbs.sh[86]echo '#!/bin/bash' -+ exgfs_wave_post_gridded_sbs.sh[87]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032800 3600. 9999 > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[88]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[89][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[90]process_grdID gsh_15m -+ wave_domain_grid.sh[13]grdID=gsh_15m -+ wave_domain_grid.sh[14]case ${grdID} in -+ wave_domain_grid.sh[30]GRDREGION=gsouth -+ wave_domain_grid.sh[30]GRDRES=0p25 -+ wave_domain_grid.sh[30]GRIDNR=255 -+ wave_domain_grid.sh[30]MODNR=11 -+ wave_domain_grid.sh[39]grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[40]echo grdNAME=gsouth.0p25 -grdNAME=gsouth.0p25 -+ wave_domain_grid.sh[41]echo GRIDNR=255 -GRIDNR=255 -+ wave_domain_grid.sh[42]echo MODNR=11 -MODNR=11 -+ exgfs_wave_post_gridded_sbs.sh[91]echo '/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032800 108 gsouth 0p25 '\''WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA'\'' > /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib2_gsh_15m.out 2>&1' -+ exgfs_wave_post_gridded_sbs.sh[92]echo 'cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib2_gsh_15m.out' -+ exgfs_wave_post_gridded_sbs.sh[94]chmod 755 cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[95]echo /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/cmdfile.7 -+ exgfs_wave_post_gridded_sbs.sh[100][[ YES == \Y\E\S ]] -+ exgfs_wave_post_gridded_sbs.sh[113][[ 8 -lt 7 ]] -+ exgfs_wave_post_gridded_sbs.sh[121]echo 'INFO: Running MPMD job with 7 commands' -INFO: Running MPMD job with 7 commands -+ exgfs_wave_post_gridded_sbs.sh[122]/work2/noaa/global/mterry/global-workflow_forked/ush/run_mpmd.sh /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/cmdfile -+ bash[8]'[' -z '' ']' -+ bash[9]case "$-" in -+ bash[12]__lmod_vx=x -+ bash[16]'[' -n x ']' -+ bash[16]set +x -Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -Shell debugging restarted -+ bash[224]unset __lmod_vx -+ run_mpmd.sh[31]source /work2/noaa/global/mterry/global-workflow_forked/ush/preamble.sh -++ preamble.sh[20]set +x -Begin run_mpmd.sh at Tue Jul 29 03:05:25 UTC 2025 -+ run_mpmd.sh[33]cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/cmdfile -+ run_mpmd.sh[36][[ YES != \Y\E\S ]] -+ run_mpmd.sh[46]export OMP_NUM_THREADS=1 -+ run_mpmd.sh[46]OMP_NUM_THREADS=1 -++ run_mpmd.sh[49]wc -l -+ run_mpmd.sh[49]nprocs=7 -+ run_mpmd.sh[52]mpmd_cmdfile=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mpmd_cmdfile -+ run_mpmd.sh[53][[ -s /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mpmd_cmdfile ]] -+ run_mpmd.sh[55]cat - INFO: Executing MPMD job, STDOUT redirected for each process separately - INFO: On failure, logs for each job will be available in /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mpmd.proc_num.out - INFO: The proc_num corresponds to the line in '/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mpmd_cmdfile' -+ run_mpmd.sh[61][[ srun -l --export=ALL --hint=nomultithread =~ ^srun.* ]] -+ run_mpmd.sh[65]nm=0 -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '0 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/cmdfile.1' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '1 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/cmdfile.2' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '2 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/cmdfile.3' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '3 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/cmdfile.4' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '4 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/cmdfile.5' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '5 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/cmdfile.6' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[68]echo '6 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/cmdfile.7' -+ run_mpmd.sh[69](( nm=nm+1 )) -+ run_mpmd.sh[67]IFS= -+ run_mpmd.sh[67]read -r line -+ run_mpmd.sh[72]set +e -+ run_mpmd.sh[74]srun -l --export=ALL --hint=nomultithread --multi-prog --output=mpmd.%j.%t.out -n 7 /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mpmd_cmdfile -+ run_mpmd.sh[75]err=0 -+ run_mpmd.sh[76]set_strict -+ preamble.sh[35][[ YES == \Y\E\S ]] -+ preamble.sh[37]set -eu -+ run_mpmd.sh[103][[ 0 -eq 0 ]] -+ run_mpmd.sh[104]rm -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mpmd_cmdfile -++ run_mpmd.sh[105]find . -name 'mpmd.*.out' -+ run_mpmd.sh[105]out_files='./mpmd.5951799.1.out -./mpmd.5951799.6.out -./mpmd.5951799.0.out -./mpmd.5951799.5.out -./mpmd.5951799.3.out -./mpmd.5951799.4.out -./mpmd.5951799.2.out' -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951799.1.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951799.1.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951799.6.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951799.6.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951799.0.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951799.0.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951799.5.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951799.5.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951799.3.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951799.3.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951799.4.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951799.4.out -+ run_mpmd.sh[106]for file in ${out_files} -+ run_mpmd.sh[107]cat ./mpmd.5951799.2.out -+ run_mpmd.sh[108]rm -f ./mpmd.5951799.2.out -+ run_mpmd.sh[110]cat mpmd.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + cmdfile.2[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh ep_10m 2021032800 3600. 9999 -1: + cmdfile.2[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grid_interp_sbs.sh[25]grdID=ep_10m -1: + wave_grid_interp_sbs.sh[26]valid_time=2021032800 -1: + wave_grid_interp_sbs.sh[27]dt=3600. -1: + wave_grid_interp_sbs.sh[28]nst=9999 -1: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010 -1: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_ep_10m -1: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.uglo_100km ./out_grd.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mod_def.uglo_100km ./mod_def.uglo_100km -1: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -1: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mod_def.ep_10m ./mod_def.ep_10m -1: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ]] -1: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m'\''' -1: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m' -1: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.ep_10m ./WHTGRIDINT.bin -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grid_interp_sbs.sh[51]weights_found=1 -1: + wave_grid_interp_sbs.sh[59]ymdhms='20210328 000000' -1: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210328 000000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/ep_10m/g ww3_gint.inp.tmpl -1: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -1: $ Input file for interpolation of uglo_100km to ep_10m -1: $------------------------------------------------ -1: $ Start Time 3600. NSteps -1: 20210328 000000 3600. 9999 -1: $ Total number of grids -1: 2 -1: $ Grid extensions -1: 'uglo_100km' -1: 'ep_10m' -1: $ -1: 0 -1: $ -1: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[70]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.312689 ']' -1: ++ prep_step[4]echo gfs_ww3_gint.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -1: INFO: Executing 'gfs_ww3_gint.x' -1: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -1: + wave_grid_interp_sbs.sh[73]cat grid_interp.ep_10m.out -1: -1: *** WAVEWATCH III Grid interpolation *** -1: =============================================== -1: -1: Comment character is '$' -1: -1: Time Information : -1: --------------------------------------------- -1: Starting Time : 2021/03/28 00:00:00 UTC -1: Interval (in sec) : 3600.00 -1: Number of requests : 9999 -1: --------------------------------------------- -1: Number of grids (including output grid) = 2 -1: -1: -1: Extension for grid 1 is --> uglo_100km -1: -1: Grid Particulars are : -1: Dimensions = 45166 1 -1: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Extension for grid 2 is --> ep_10m -1: -1: Grid Particulars are : -1: Dimensions = 511 301 -1: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -1: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -1: -1: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -1: -1: -1: Preparing interpolation weights for output grid -1: Total number of wet points for interpolation 7439 -1: -1: -1: Variable: Grid Interpolation Map Units: 0.100E+01 -1: -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 301 | | -1: | | -1: | 0 0 | -1: 262 | 0 0 0 0 | -1: | 0 0 0 0 | -1: | 0 0 0 0 | -1: 223 | 0 0 0 | -1: | 0 | -1: | 0 | -1: 184 | | -1: | 0 0 0 | -1: | 0 0 0 0 | -1: 145 | 0 | -1: | | -1: | | -1: 106 | | -1: | | -1: | | -1: 67 | | -1: | 0 | -1: | | -1: 28 | 0 | -1: | | -1: | | -1: +-------------------------------------------------------------------------------------------------------------------------+ -1: 1 23 45 67 89 111 133 155 177 199 221 243 265 287 309 331 353 375 397 419 441 463 485 507 -1: -1: -1: Interpolating fields .... -1: -1: Output group 1 -1: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -1: Output group 2 -1: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -1: Output group 3 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 4 -1: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -1: Output group 5 -1: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -1: Output group 6 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 7 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 8 -1: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -1: Output group 9 -1: Output variables skipped -1: Output group 10 -1: Output variables skipped -1: ------------------------------------------------ -1: 1Current vel. -1: 1Wind speed -1: 1Ice concentration -1: 2Wave height -1: 2Mean wave period(+2) -1: 2Mean wave period(+1) -1: 2Peak frequency -1: 2Mean wave dir. a1b1 -1: 2Peak direction -1: 4Part. wave height -1: 4Part. peak period -1: 4Part. mean direction -1: 5Charnock parameter -1: ------------------------------------------------ -1: OUTPUT TIME : 2021/03/28 00:00:00 UTC -1: -1: End of file reached -1: -1: -1: *** End of Grid interpolation Routine *** -1: =============================================== -1: -1: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -1: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -1: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.ep_10m ]] -1: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_ep_10m/out_grd.ep_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.ep_10m -1: + cmdfile.2[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh ep_10m 255 11 2021032800 108 epacif 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + cmdfile.2[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib2_ep_10m.out -1: + bash[8]'[' -z '' ']' -1: + bash[9]case "$-" in -1: + bash[12]__lmod_vx=x -1: + bash[16]'[' -n x ']' -1: + bash[16]set +x -1: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -1: Shell debugging restarted -1: + bash[224]unset __lmod_vx -1: + wave_grib2_sbs.sh[30]grdID=ep_10m -1: + wave_grib2_sbs.sh[31]GRIDNR=255 -1: + wave_grib2_sbs.sh[32]MODNR=11 -1: + wave_grib2_sbs.sh[33]valid_time=2021032800 -1: + wave_grib2_sbs.sh[34]fhr=108 -1: + wave_grib2_sbs.sh[35]grid_region=epacif -1: + wave_grib2_sbs.sh[36]grid_res=0p16 -1: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -1: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010 -1: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_ep_10m -1: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_ep_10m -1: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_ep_10m -1: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_ep_10m -1: ++ wave_grib2_sbs.sh[47]printf %03i 108 -1: + wave_grib2_sbs.sh[47]FH3=108 -1: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_epacif_0p16 -1: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16 -1: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.epacif.0p16.f108.grib2 -1: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f108.grib2 ]] -1: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.ep_10m.inp.tmpl ./ww3_grib2.ep_10m.inp.tmpl -1: + cpreq[4]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mod_def.ep_10m ./mod_def.ww3 -1: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.ep_10m ./out_grd.ww3 -1: + wave_grib2_sbs.sh[73]ngrib=1 -1: + wave_grib2_sbs.sh[74]dtgrib=3600 -1: + wave_grib2_sbs.sh[75]tstart='20210328 000000' -1: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210328 000000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.ep_10m.inp.tmpl -1: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -1: $ WAVEWATCH-III gridded output input file -1: $ ---------------------------------------- -1: 20210328 000000 3600 1 -1: N -1: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -1: $ -1: 20210328 000000 7 11 255 0 0 -1: $ -1: $ end of input file -1: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[88]source prep_step -1: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -1: ++ prep_step[3]'[' -n OUTPUT.312689 ']' -1: ++ prep_step[4]echo gfs_ww3_grib.x -1: ++ prep_step[7]'[' -f errfile ']' -1: ++ prep_step[11]export FORT01=0 -1: ++ prep_step[11]FORT01=0 -1: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -1: +++ prep_step[12]awk -F= '{print $1}' -1: +++ prep_step[12]env -1: ++ prep_step[12]unset FORT01 -1: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -1: + wave_grib2_sbs.sh[90]export err=0 -1: + wave_grib2_sbs.sh[90]err=0 -1: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[95]cat grib2_epacif_108.out -1: -1: *** WAVEWATCH III GRIB output postp. *** -1: ============================================== -1: -1: Comment character is '$' -1: -1: Grid name : East Pacific 10 min wave grid -1: -1: LINEIN: -1: 20210328 000000 3600 1 -1: -1: 20210328000000 3600 1 -1: GEN_PRO -99999 -1: -1: Output time data : -1: ----------------------------------------------------- -1: First time : 2021/03/28 00:00:00 UTC -1: Interval : 01:00:00 -1: Number of requests : 1 -1: Fields : Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: Requested output fields not yet available: -1: ----------------------------------------------------- -1: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -1: -1: Successfully requested output fields : -1: ----------------------------------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: -1: Additional GRIB parameters : -1: ----------------------------------------------------- -1: Run time : 2021/03/28 00:00:00 UTC -1: GRIB center ID : 7 -1: GRIB gen. proc. ID : 11 -1: GRIB grid ID : 255 -1: GRIB GDS parameter : 0 -1: Fields in file : -1: -------------------------- -1: Current vel. -1: Wind speed -1: Ice concentration -1: Wave height -1: Mean wave period(+2) -1: Mean wave period(+1) -1: Peak frequency -1: Mean wave dir. a1b1 -1: Peak direction -1: Part. wave height -1: Part. peak period -1: Part. mean direction -1: Charnock parameter -1: -1: CHOSEN GRID TYPE: : LLRECTILINEAR -1: -1: -1: -1: Generating file -1: ----------------------------------------------------- -1: Data for 2021/03/28 00:00:00 UTC 0H forecast. -1: Warning: bitmask off everywhere. -1: Pretend one point in jpcpack to avoid crash. -1: Warning: bitmask off everywhere. -1: Pretend one point in jpcpack to avoid crash. -1: Warning: bitmask off everywhere. -1: Pretend one point in jpcpack to avoid crash. -1: -1: End of program -1: ========================================= -1: WAVEWATCH III GRIB output -1: -1: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -1: + wave_grib2_sbs.sh[102][[ 108 -gt 0 ]] -1: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '108 hour fcst' -grib gfs.wave.t12z.epacif.0p16.f108.grib2 -1: 1:0:d=2021032312:SPC:surface:108 hour fcst: -1: 2:21998:d=2021032312:DIRC:surface:108 hour fcst: -1: 3:50715:d=2021032312:UOGRD:surface:108 hour fcst: -1: 4:72501:d=2021032312:VOGRD:surface:108 hour fcst: -1: 5:94625:d=2021032312:WIND:surface:108 hour fcst: -1: 6:119050:d=2021032312:WDIR:surface:108 hour fcst: -1: 7:146070:d=2021032312:UGRD:surface:108 hour fcst: -1: 8:170600:d=2021032312:VGRD:surface:108 hour fcst: -1: 9:195021:d=2021032312:ICEC:surface:108 hour fcst: -1: 10:214429:d=2021032312:HTSGW:surface:108 hour fcst: -1: 11:236681:d=2021032312:IMWF:surface:108 hour fcst: -1: 12:259687:d=2021032312:MWSPER:surface:108 hour fcst: -1: 13:282699:d=2021032312:PERPW:surface:108 hour fcst: -1: 14:305701:d=2021032312:WWSDIR:surface:108 hour fcst: -1: 15:331760:d=2021032312:DIRPW:surface:108 hour fcst: -1: 16:358151:d=2021032312:WVHGT:surface:108 hour fcst: -1: 17:380261:d=2021032312:SWELL:1 in sequence:108 hour fcst: -1: 18:403341:d=2021032312:SWELL:2 in sequence:108 hour fcst: -1: 19:424028:d=2021032312:SWELL:3 in sequence:108 hour fcst: -1: 20:443436:d=2021032312:WVPER:surface:108 hour fcst: -1: 21:465534:d=2021032312:SWPER:1 in sequence:108 hour fcst: -1: 22:488960:d=2021032312:SWPER:2 in sequence:108 hour fcst: -1: 23:509545:d=2021032312:SWPER:3 in sequence:108 hour fcst: -1: 24:528953:d=2021032312:WVDIR:surface:108 hour fcst: -1: 25:553700:d=2021032312:SWDIR:1 in sequence:108 hour fcst: -1: 26:581285:d=2021032312:SWDIR:2 in sequence:108 hour fcst: -1: 27:603090:d=2021032312:SWDIR:3 in sequence:108 hour fcst: -1: + wave_grib2_sbs.sh[104]err=0 -1: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -1: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.epacif.0p16.f108.grib2 -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f108.grib2 ]] -1: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.epacif.0p16.f108.grib2.idx ]] -1: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.epacif.0p16.f108.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f108.grib2 -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f108.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f108.grib2 = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f108.grib2 ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f108.grib2 -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f108.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f108.grib2.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f108.grib2.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f108.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f108.grib2 -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.epacif.0p16.f108.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f108.grib2.idx -1: + cpfs[3]'[' 2 -ne 2 ']' -1: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f108.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f108.grib2.idx = ./ ']' -1: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f108.grib2.idx ']' -1: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f108.grib2.idx -1: + cpfs[16]cp gfs.wave.t12z.epacif.0p16.f108.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f108.grib2.idx.cptmp -1: + cpfs[18]'[' 0 -ne 0 ']' -1: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f108.grib2.idx.cptmp -1: + cpfs[23]'[' 0 -ne 0 ']' -1: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f108.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/epacif.0p16/gfs.wave.t12z.epacif.0p16.f108.grib2.idx -1: + cpfs[28]'[' 0 -ne 0 ']' -1: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.epacif.0p16.f108.grib2 and gfs.wave.t12z.epacif.0p16.f108.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_ep_10m to COM' -1: INFO: Copied gfs.wave.t12z.epacif.0p16.f108.grib2 and gfs.wave.t12z.epacif.0p16.f108.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_ep_10m to COM -1: + wave_grib2_sbs.sh[130][[ ep_10m == '' ]] -1: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -1: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.epacif.0p16.f108.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -1: INFO: gfs.wave.t12z.epacif.0p16.f108.grib2 is global.0p50 or SENDDBN is NO, no alert sent -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + cmdfile.7[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gsh_15m 2021032800 3600. 9999 -6: + cmdfile.7[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grid_interp_sbs.sh[25]grdID=gsh_15m -6: + wave_grid_interp_sbs.sh[26]valid_time=2021032800 -6: + wave_grid_interp_sbs.sh[27]dt=3600. -6: + wave_grid_interp_sbs.sh[28]nst=9999 -6: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010 -6: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_gsh_15m -6: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.uglo_100km ./out_grd.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mod_def.uglo_100km ./mod_def.uglo_100km -6: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -6: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mod_def.gsh_15m ./mod_def.gsh_15m -6: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m'\''' -6: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m' -6: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gsh_15m ./WHTGRIDINT.bin -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grid_interp_sbs.sh[51]weights_found=1 -6: + wave_grid_interp_sbs.sh[59]ymdhms='20210328 000000' -6: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210328 000000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gsh_15m/g ww3_gint.inp.tmpl -6: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -6: $ Input file for interpolation of uglo_100km to gsh_15m -6: $------------------------------------------------ -6: $ Start Time 3600. NSteps -6: 20210328 000000 3600. 9999 -6: $ Total number of grids -6: 2 -6: $ Grid extensions -6: 'uglo_100km' -6: 'gsh_15m' -6: $ -6: 0 -6: $ -6: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[70]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.312689 ']' -6: ++ prep_step[4]echo gfs_ww3_gint.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -6: INFO: Executing 'gfs_ww3_gint.x' -6: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -6: + wave_grid_interp_sbs.sh[73]cat grid_interp.gsh_15m.out -6: -6: *** WAVEWATCH III Grid interpolation *** -6: =============================================== -6: -6: Comment character is '$' -6: -6: Time Information : -6: --------------------------------------------- -6: Starting Time : 2021/03/28 00:00:00 UTC -6: Interval (in sec) : 3600.00 -6: Number of requests : 9999 -6: --------------------------------------------- -6: Number of grids (including output grid) = 2 -6: -6: -6: Extension for grid 1 is --> uglo_100km -6: -6: Grid Particulars are : -6: Dimensions = 45166 1 -6: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Extension for grid 2 is --> gsh_15m -6: -6: Grid Particulars are : -6: Dimensions = 1440 277 -6: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -6: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -6: -6: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -6: -6: -6: Preparing interpolation weights for output grid -6: Total number of wet points for interpolation 317192 -6: -6: -6: Variable: Grid Interpolation Map Units: 0.100E+01 -6: -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 277 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 241 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 205 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 169 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 133 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 97 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: 61 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 0 0 0 0 0 0 0 | -6: 25 | 0 0 0 0 0 0 0 0 0 | -6: | 0 0 0 0 | -6: | | -6: +-------------------------------------------------------------------------------------------------------------------------+ -6: 1 62 123 184 245 306 367 428 489 550 611 672 733 794 855 916 977 1038 1099 1160 1221 1282 1343 1404 -6: -6: -6: Interpolating fields .... -6: -6: Output group 1 -6: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -6: Output group 2 -6: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -6: Output group 3 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 4 -6: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -6: Output group 5 -6: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -6: Output group 6 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 7 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 8 -6: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -6: Output group 9 -6: Output variables skipped -6: Output group 10 -6: Output variables skipped -6: ------------------------------------------------ -6: 1Current vel. -6: 1Wind speed -6: 1Ice concentration -6: 2Wave height -6: 2Mean wave period(+2) -6: 2Mean wave period(+1) -6: 2Peak frequency -6: 2Mean wave dir. a1b1 -6: 2Peak direction -6: 4Part. wave height -6: 4Part. peak period -6: 4Part. mean direction -6: 5Charnock parameter -6: ------------------------------------------------ -6: OUTPUT TIME : 2021/03/28 00:00:00 UTC -6: -6: End of file reached -6: -6: -6: *** End of Grid interpolation Routine *** -6: =============================================== -6: -6: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -6: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -6: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.gsh_15m ]] -6: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_gsh_15m/out_grd.gsh_15m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.gsh_15m -6: + cmdfile.7[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gsh_15m 255 11 2021032800 108 gsouth 0p25 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + cmdfile.7[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib2_gsh_15m.out -6: + bash[8]'[' -z '' ']' -6: + bash[9]case "$-" in -6: + bash[12]__lmod_vx=x -6: + bash[16]'[' -n x ']' -6: + bash[16]set +x -6: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -6: Shell debugging restarted -6: + bash[224]unset __lmod_vx -6: + wave_grib2_sbs.sh[30]grdID=gsh_15m -6: + wave_grib2_sbs.sh[31]GRIDNR=255 -6: + wave_grib2_sbs.sh[32]MODNR=11 -6: + wave_grib2_sbs.sh[33]valid_time=2021032800 -6: + wave_grib2_sbs.sh[34]fhr=108 -6: + wave_grib2_sbs.sh[35]grid_region=gsouth -6: + wave_grib2_sbs.sh[36]grid_res=0p25 -6: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -6: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010 -6: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_gsh_15m -6: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_gsh_15m -6: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_gsh_15m -6: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_gsh_15m -6: ++ wave_grib2_sbs.sh[47]printf %03i 108 -6: + wave_grib2_sbs.sh[47]FH3=108 -6: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -6: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -6: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.gsouth.0p25.f108.grib2 -6: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f108.grib2 ]] -6: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gsh_15m.inp.tmpl ./ww3_grib2.gsh_15m.inp.tmpl -6: + cpreq[4]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mod_def.gsh_15m ./mod_def.ww3 -6: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.gsh_15m ./out_grd.ww3 -6: + wave_grib2_sbs.sh[73]ngrib=1 -6: + wave_grib2_sbs.sh[74]dtgrib=3600 -6: + wave_grib2_sbs.sh[75]tstart='20210328 000000' -6: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210328 000000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gsh_15m.inp.tmpl -6: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -6: $ WAVEWATCH-III gridded output input file -6: $ ---------------------------------------- -6: 20210328 000000 3600 1 -6: N -6: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -6: $ -6: 20210328 000000 7 11 255 0 0 -6: $ -6: $ end of input file -6: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[88]source prep_step -6: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -6: ++ prep_step[3]'[' -n OUTPUT.312689 ']' -6: ++ prep_step[4]echo gfs_ww3_grib.x -6: ++ prep_step[7]'[' -f errfile ']' -6: ++ prep_step[11]export FORT01=0 -6: ++ prep_step[11]FORT01=0 -6: +++ prep_step[12]awk -F= '{print $1}' -6: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -6: +++ prep_step[12]env -6: ++ prep_step[12]unset FORT01 -6: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -6: + wave_grib2_sbs.sh[90]export err=0 -6: + wave_grib2_sbs.sh[90]err=0 -6: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[95]cat grib2_gsouth_108.out -6: -6: *** WAVEWATCH III GRIB output postp. *** -6: ============================================== -6: -6: Comment character is '$' -6: -6: Grid name : GFSv16-wave S Hemisphere 1/4 d -6: -6: LINEIN: -6: 20210328 000000 3600 1 -6: -6: 20210328000000 3600 1 -6: GEN_PRO -99999 -6: -6: Output time data : -6: ----------------------------------------------------- -6: First time : 2021/03/28 00:00:00 UTC -6: Interval : 01:00:00 -6: Number of requests : 1 -6: Fields : Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: Requested output fields not yet available: -6: ----------------------------------------------------- -6: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -6: -6: Successfully requested output fields : -6: ----------------------------------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: -6: Additional GRIB parameters : -6: ----------------------------------------------------- -6: Run time : 2021/03/28 00:00:00 UTC -6: GRIB center ID : 7 -6: GRIB gen. proc. ID : 11 -6: GRIB grid ID : 255 -6: GRIB GDS parameter : 0 -6: Fields in file : -6: -------------------------- -6: Current vel. -6: Wind speed -6: Ice concentration -6: Wave height -6: Mean wave period(+2) -6: Mean wave period(+1) -6: Peak frequency -6: Mean wave dir. a1b1 -6: Peak direction -6: Part. wave height -6: Part. peak period -6: Part. mean direction -6: Charnock parameter -6: -6: CHOSEN GRID TYPE: : LLRECTILINEAR -6: -6: -6: -6: Generating file -6: ----------------------------------------------------- -6: Data for 2021/03/28 00:00:00 UTC 0H forecast. -6: -6: End of program -6: ========================================= -6: WAVEWATCH III GRIB output -6: -6: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -6: + wave_grib2_sbs.sh[102][[ 108 -gt 0 ]] -6: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '108 hour fcst' -grib gfs.wave.t12z.gsouth.0p25.f108.grib2 -6: 1:0:d=2021032312:SPC:surface:108 hour fcst: -6: 2:119064:d=2021032312:DIRC:surface:108 hour fcst: -6: 3:505627:d=2021032312:UOGRD:surface:108 hour fcst: -6: 4:620160:d=2021032312:VOGRD:surface:108 hour fcst: -6: 5:742262:d=2021032312:WIND:surface:108 hour fcst: -6: 6:985760:d=2021032312:WDIR:surface:108 hour fcst: -6: 7:1365674:d=2021032312:UGRD:surface:108 hour fcst: -6: 8:1603110:d=2021032312:VGRD:surface:108 hour fcst: -6: 9:1841765:d=2021032312:ICEC:surface:108 hour fcst: -6: 10:1902834:d=2021032312:HTSGW:surface:108 hour fcst: -6: 11:2057669:d=2021032312:IMWF:surface:108 hour fcst: -6: 12:2229724:d=2021032312:MWSPER:surface:108 hour fcst: -6: 13:2402222:d=2021032312:PERPW:surface:108 hour fcst: -6: 14:2585180:d=2021032312:WWSDIR:surface:108 hour fcst: -6: 15:2916090:d=2021032312:DIRPW:surface:108 hour fcst: -6: 16:3249466:d=2021032312:WVHGT:surface:108 hour fcst: -6: 17:3404530:d=2021032312:SWELL:1 in sequence:108 hour fcst: -6: 18:3579060:d=2021032312:SWELL:2 in sequence:108 hour fcst: -6: 19:3713501:d=2021032312:SWELL:3 in sequence:108 hour fcst: -6: 20:3797301:d=2021032312:WVPER:surface:108 hour fcst: -6: 21:3972212:d=2021032312:SWPER:1 in sequence:108 hour fcst: -6: 22:4170078:d=2021032312:SWPER:2 in sequence:108 hour fcst: -6: 23:4343036:d=2021032312:SWPER:3 in sequence:108 hour fcst: -6: 24:4447366:d=2021032312:WVDIR:surface:108 hour fcst: -6: 25:4718234:d=2021032312:SWDIR:1 in sequence:108 hour fcst: -6: 26:5072391:d=2021032312:SWDIR:2 in sequence:108 hour fcst: -6: 27:5365392:d=2021032312:SWDIR:3 in sequence:108 hour fcst: -6: + wave_grib2_sbs.sh[104]err=0 -6: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -6: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.gsouth.0p25.f108.grib2 -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f108.grib2 ]] -6: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.gsouth.0p25.f108.grib2.idx ]] -6: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.gsouth.0p25.f108.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f108.grib2 -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f108.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f108.grib2 = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f108.grib2 ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f108.grib2 -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f108.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f108.grib2.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f108.grib2.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f108.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f108.grib2 -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.gsouth.0p25.f108.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f108.grib2.idx -6: + cpfs[3]'[' 2 -ne 2 ']' -6: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f108.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f108.grib2.idx = ./ ']' -6: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f108.grib2.idx ']' -6: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f108.grib2.idx -6: + cpfs[16]cp gfs.wave.t12z.gsouth.0p25.f108.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f108.grib2.idx.cptmp -6: + cpfs[18]'[' 0 -ne 0 ']' -6: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f108.grib2.idx.cptmp -6: + cpfs[23]'[' 0 -ne 0 ']' -6: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f108.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f108.grib2.idx -6: + cpfs[28]'[' 0 -ne 0 ']' -6: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.gsouth.0p25.f108.grib2 and gfs.wave.t12z.gsouth.0p25.f108.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_gsh_15m to COM' -6: INFO: Copied gfs.wave.t12z.gsouth.0p25.f108.grib2 and gfs.wave.t12z.gsouth.0p25.f108.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_gsh_15m to COM -6: + wave_grib2_sbs.sh[130][[ gsh_15m == '' ]] -6: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -6: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.gsouth.0p25.f108.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -6: INFO: gfs.wave.t12z.gsouth.0p25.f108.grib2 is global.0p50 or SENDDBN is NO, no alert sent -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + cmdfile.1[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh at_10m 2021032800 3600. 9999 -0: + cmdfile.1[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grid_interp_sbs.sh[25]grdID=at_10m -0: + wave_grid_interp_sbs.sh[26]valid_time=2021032800 -0: + wave_grid_interp_sbs.sh[27]dt=3600. -0: + wave_grid_interp_sbs.sh[28]nst=9999 -0: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010 -0: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_at_10m -0: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.uglo_100km ./out_grd.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mod_def.uglo_100km ./mod_def.uglo_100km -0: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -0: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mod_def.at_10m ./mod_def.at_10m -0: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ]] -0: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m'\''' -0: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m' -0: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.at_10m ./WHTGRIDINT.bin -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grid_interp_sbs.sh[51]weights_found=1 -0: + wave_grid_interp_sbs.sh[59]ymdhms='20210328 000000' -0: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210328 000000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/at_10m/g ww3_gint.inp.tmpl -0: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -0: $ Input file for interpolation of uglo_100km to at_10m -0: $------------------------------------------------ -0: $ Start Time 3600. NSteps -0: 20210328 000000 3600. 9999 -0: $ Total number of grids -0: 2 -0: $ Grid extensions -0: 'uglo_100km' -0: 'at_10m' -0: $ -0: 0 -0: $ -0: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[70]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.312689 ']' -0: ++ prep_step[4]echo gfs_ww3_gint.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -0: INFO: Executing 'gfs_ww3_gint.x' -0: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -0: + wave_grid_interp_sbs.sh[73]cat grid_interp.at_10m.out -0: -0: *** WAVEWATCH III Grid interpolation *** -0: =============================================== -0: -0: Comment character is '$' -0: -0: Time Information : -0: --------------------------------------------- -0: Starting Time : 2021/03/28 00:00:00 UTC -0: Interval (in sec) : 3600.00 -0: Number of requests : 9999 -0: --------------------------------------------- -0: Number of grids (including output grid) = 2 -0: -0: -0: Extension for grid 1 is --> uglo_100km -0: -0: Grid Particulars are : -0: Dimensions = 45166 1 -0: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Extension for grid 2 is --> at_10m -0: -0: Grid Particulars are : -0: Dimensions = 301 331 -0: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -0: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -0: -0: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -0: -0: -0: Preparing interpolation weights for output grid -0: Total number of wet points for interpolation 29591 -0: -0: -0: Variable: Grid Interpolation Map Units: 0.100E+01 -0: -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 331 | | -0: | | -0: | | -0: 289 | | -0: | | -0: | 0 0 0 | -0: 247 | 0 0 0 0 | -0: | 0 0 0 0 0 | -0: | 0 0 0 0 | -0: 205 | 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 163 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 121 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: 79 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -0: | 0 0 0 0 0 0 0 0 0 | -0: | 0 0 | -0: 37 | 0 | -0: | | -0: | | -0: +-------------------------------------------------------------------------------------------------------------------------+ -0: 1 14 27 40 53 66 79 92 105 118 131 144 157 170 183 196 209 222 235 248 261 274 287 300 -0: -0: -0: Interpolating fields .... -0: -0: Output group 1 -0: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -0: Output group 2 -0: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -0: Output group 3 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 4 -0: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -0: Output group 5 -0: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -0: Output group 6 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 7 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 8 -0: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -0: Output group 9 -0: Output variables skipped -0: Output group 10 -0: Output variables skipped -0: ------------------------------------------------ -0: 1Current vel. -0: 1Wind speed -0: 1Ice concentration -0: 2Wave height -0: 2Mean wave period(+2) -0: 2Mean wave period(+1) -0: 2Peak frequency -0: 2Mean wave dir. a1b1 -0: 2Peak direction -0: 4Part. wave height -0: 4Part. peak period -0: 4Part. mean direction -0: 5Charnock parameter -0: ------------------------------------------------ -0: OUTPUT TIME : 2021/03/28 00:00:00 UTC -0: -0: End of file reached -0: -0: -0: *** End of Grid interpolation Routine *** -0: =============================================== -0: -0: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -0: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -0: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.at_10m ]] -0: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_at_10m/out_grd.at_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.at_10m -0: + cmdfile.1[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh at_10m 255 11 2021032800 108 atlocn 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + cmdfile.1[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib2_at_10m.out -0: + bash[8]'[' -z '' ']' -0: + bash[9]case "$-" in -0: + bash[12]__lmod_vx=x -0: + bash[16]'[' -n x ']' -0: + bash[16]set +x -0: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -0: Shell debugging restarted -0: + bash[224]unset __lmod_vx -0: + wave_grib2_sbs.sh[30]grdID=at_10m -0: + wave_grib2_sbs.sh[31]GRIDNR=255 -0: + wave_grib2_sbs.sh[32]MODNR=11 -0: + wave_grib2_sbs.sh[33]valid_time=2021032800 -0: + wave_grib2_sbs.sh[34]fhr=108 -0: + wave_grib2_sbs.sh[35]grid_region=atlocn -0: + wave_grib2_sbs.sh[36]grid_res=0p16 -0: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -0: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010 -0: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_at_10m -0: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_at_10m -0: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_at_10m -0: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_at_10m -0: ++ wave_grib2_sbs.sh[47]printf %03i 108 -0: + wave_grib2_sbs.sh[47]FH3=108 -0: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_atlocn_0p16 -0: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16 -0: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.atlocn.0p16.f108.grib2 -0: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f108.grib2 ]] -0: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.at_10m.inp.tmpl ./ww3_grib2.at_10m.inp.tmpl -0: + cpreq[4]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mod_def.at_10m ./mod_def.ww3 -0: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.at_10m ./out_grd.ww3 -0: + wave_grib2_sbs.sh[73]ngrib=1 -0: + wave_grib2_sbs.sh[74]dtgrib=3600 -0: + wave_grib2_sbs.sh[75]tstart='20210328 000000' -0: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210328 000000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.at_10m.inp.tmpl -0: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -0: $ WAVEWATCH-III gridded output input file -0: $ ---------------------------------------- -0: 20210328 000000 3600 1 -0: N -0: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -0: $ -0: 20210328 000000 7 11 255 0 0 -0: $ -0: $ end of input file -0: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[88]source prep_step -0: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -0: ++ prep_step[3]'[' -n OUTPUT.312689 ']' -0: ++ prep_step[4]echo gfs_ww3_grib.x -0: ++ prep_step[7]'[' -f errfile ']' -0: ++ prep_step[11]export FORT01=0 -0: ++ prep_step[11]FORT01=0 -0: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -0: +++ prep_step[12]awk -F= '{print $1}' -0: +++ prep_step[12]env -0: ++ prep_step[12]unset FORT01 -0: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -0: + wave_grib2_sbs.sh[90]export err=0 -0: + wave_grib2_sbs.sh[90]err=0 -0: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[95]cat grib2_atlocn_108.out -0: -0: *** WAVEWATCH III GRIB output postp. *** -0: ============================================== -0: -0: Comment character is '$' -0: -0: Grid name : NW Atlantic 10 min wave grid -0: -0: LINEIN: -0: 20210328 000000 3600 1 -0: -0: 20210328000000 3600 1 -0: GEN_PRO -99999 -0: -0: Output time data : -0: ----------------------------------------------------- -0: First time : 2021/03/28 00:00:00 UTC -0: Interval : 01:00:00 -0: Number of requests : 1 -0: Fields : Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: Requested output fields not yet available: -0: ----------------------------------------------------- -0: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -0: -0: Successfully requested output fields : -0: ----------------------------------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: -0: Additional GRIB parameters : -0: ----------------------------------------------------- -0: Run time : 2021/03/28 00:00:00 UTC -0: GRIB center ID : 7 -0: GRIB gen. proc. ID : 11 -0: GRIB grid ID : 255 -0: GRIB GDS parameter : 0 -0: Fields in file : -0: -------------------------- -0: Current vel. -0: Wind speed -0: Ice concentration -0: Wave height -0: Mean wave period(+2) -0: Mean wave period(+1) -0: Peak frequency -0: Mean wave dir. a1b1 -0: Peak direction -0: Part. wave height -0: Part. peak period -0: Part. mean direction -0: Charnock parameter -0: -0: CHOSEN GRID TYPE: : LLRECTILINEAR -0: -0: -0: -0: Generating file -0: ----------------------------------------------------- -0: Data for 2021/03/28 00:00:00 UTC 0H forecast. -0: -0: End of program -0: ========================================= -0: WAVEWATCH III GRIB output -0: -0: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -0: + wave_grib2_sbs.sh[102][[ 108 -gt 0 ]] -0: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '108 hour fcst' -grib gfs.wave.t12z.atlocn.0p16.f108.grib2 -0: 1:0:d=2021032312:SPC:surface:108 hour fcst: -0: 2:20192:d=2021032312:DIRC:surface:108 hour fcst: -0: 3:59828:d=2021032312:UOGRD:surface:108 hour fcst: -0: 4:80061:d=2021032312:VOGRD:surface:108 hour fcst: -0: 5:100410:d=2021032312:WIND:surface:108 hour fcst: -0: 6:130744:d=2021032312:WDIR:surface:108 hour fcst: -0: 7:171190:d=2021032312:UGRD:surface:108 hour fcst: -0: 8:201674:d=2021032312:VGRD:surface:108 hour fcst: -0: 9:231053:d=2021032312:ICEC:surface:108 hour fcst: -0: 10:243688:d=2021032312:HTSGW:surface:108 hour fcst: -0: 11:269124:d=2021032312:IMWF:surface:108 hour fcst: -0: 12:295655:d=2021032312:MWSPER:surface:108 hour fcst: -0: 13:322443:d=2021032312:PERPW:surface:108 hour fcst: -0: 14:350274:d=2021032312:WWSDIR:surface:108 hour fcst: -0: 15:389909:d=2021032312:DIRPW:surface:108 hour fcst: -0: 16:430062:d=2021032312:WVHGT:surface:108 hour fcst: -0: 17:454400:d=2021032312:SWELL:1 in sequence:108 hour fcst: -0: 18:474837:d=2021032312:SWELL:2 in sequence:108 hour fcst: -0: 19:491563:d=2021032312:SWELL:3 in sequence:108 hour fcst: -0: 20:505684:d=2021032312:WVPER:surface:108 hour fcst: -0: 21:531905:d=2021032312:SWPER:1 in sequence:108 hour fcst: -0: 22:553883:d=2021032312:SWPER:2 in sequence:108 hour fcst: -0: 23:572965:d=2021032312:SWPER:3 in sequence:108 hour fcst: -0: 24:588241:d=2021032312:WVDIR:surface:108 hour fcst: -0: 25:625140:d=2021032312:SWDIR:1 in sequence:108 hour fcst: -0: 26:655800:d=2021032312:SWDIR:2 in sequence:108 hour fcst: -0: 27:680637:d=2021032312:SWDIR:3 in sequence:108 hour fcst: -0: + wave_grib2_sbs.sh[104]err=0 -0: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -0: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.atlocn.0p16.f108.grib2 -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f108.grib2 ]] -0: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.atlocn.0p16.f108.grib2.idx ]] -0: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.atlocn.0p16.f108.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f108.grib2 -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f108.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f108.grib2 = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f108.grib2 ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f108.grib2 -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f108.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f108.grib2.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f108.grib2.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f108.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f108.grib2 -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.atlocn.0p16.f108.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f108.grib2.idx -0: + cpfs[3]'[' 2 -ne 2 ']' -0: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f108.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f108.grib2.idx = ./ ']' -0: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f108.grib2.idx ']' -0: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f108.grib2.idx -0: + cpfs[16]cp gfs.wave.t12z.atlocn.0p16.f108.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f108.grib2.idx.cptmp -0: + cpfs[18]'[' 0 -ne 0 ']' -0: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f108.grib2.idx.cptmp -0: + cpfs[23]'[' 0 -ne 0 ']' -0: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f108.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/atlocn.0p16/gfs.wave.t12z.atlocn.0p16.f108.grib2.idx -0: + cpfs[28]'[' 0 -ne 0 ']' -0: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.atlocn.0p16.f108.grib2 and gfs.wave.t12z.atlocn.0p16.f108.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_at_10m to COM' -0: INFO: Copied gfs.wave.t12z.atlocn.0p16.f108.grib2 and gfs.wave.t12z.atlocn.0p16.f108.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_at_10m to COM -0: + wave_grib2_sbs.sh[130][[ at_10m == '' ]] -0: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -0: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.atlocn.0p16.f108.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -0: INFO: gfs.wave.t12z.atlocn.0p16.f108.grib2 is global.0p50 or SENDDBN is NO, no alert sent -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + cmdfile.6[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh gnh_10m 2021032800 3600. 9999 -5: + cmdfile.6[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grid_interp_sbs.sh[25]grdID=gnh_10m -5: + wave_grid_interp_sbs.sh[26]valid_time=2021032800 -5: + wave_grid_interp_sbs.sh[27]dt=3600. -5: + wave_grid_interp_sbs.sh[28]nst=9999 -5: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010 -5: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_gnh_10m -5: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.uglo_100km ./out_grd.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mod_def.uglo_100km ./mod_def.uglo_100km -5: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -5: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mod_def.gnh_10m ./mod_def.gnh_10m -5: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m'\''' -5: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m' -5: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.gnh_10m ./WHTGRIDINT.bin -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grid_interp_sbs.sh[51]weights_found=1 -5: + wave_grid_interp_sbs.sh[59]ymdhms='20210328 000000' -5: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210328 000000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/gnh_10m/g ww3_gint.inp.tmpl -5: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -5: $ Input file for interpolation of uglo_100km to gnh_10m -5: $------------------------------------------------ -5: $ Start Time 3600. NSteps -5: 20210328 000000 3600. 9999 -5: $ Total number of grids -5: 2 -5: $ Grid extensions -5: 'uglo_100km' -5: 'gnh_10m' -5: $ -5: 0 -5: $ -5: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[70]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.312689 ']' -5: ++ prep_step[4]echo gfs_ww3_gint.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -5: INFO: Executing 'gfs_ww3_gint.x' -5: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -5: + wave_grid_interp_sbs.sh[73]cat grid_interp.gnh_10m.out -5: -5: *** WAVEWATCH III Grid interpolation *** -5: =============================================== -5: -5: Comment character is '$' -5: -5: Time Information : -5: --------------------------------------------- -5: Starting Time : 2021/03/28 00:00:00 UTC -5: Interval (in sec) : 3600.00 -5: Number of requests : 9999 -5: --------------------------------------------- -5: Number of grids (including output grid) = 2 -5: -5: -5: Extension for grid 1 is --> uglo_100km -5: -5: Grid Particulars are : -5: Dimensions = 45166 1 -5: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Extension for grid 2 is --> gnh_10m -5: -5: Grid Particulars are : -5: Dimensions = 2160 406 -5: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -5: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -5: -5: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -5: -5: -5: Preparing interpolation weights for output grid -5: Total number of wet points for interpolation 571209 -5: -5: -5: Variable: Grid Interpolation Map Units: 0.100E+01 -5: -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 406 | | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 | -5: 355 | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 304 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 253 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 202 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 151 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 100 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: 49 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -5: +-------------------------------------------------------------------------------------------------------------------------+ -5: 1 92 183 274 365 456 547 638 729 820 911 1002 1093 1184 1275 1366 1457 1548 1639 1730 1821 1912 2003 2094 -5: -5: -5: Interpolating fields .... -5: -5: Output group 1 -5: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -5: Output group 2 -5: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -5: Output group 3 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 4 -5: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -5: Output group 5 -5: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -5: Output group 6 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 7 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 8 -5: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -5: Output group 9 -5: Output variables skipped -5: Output group 10 -5: Output variables skipped -5: ------------------------------------------------ -5: 1Current vel. -5: 1Wind speed -5: 1Ice concentration -5: 2Wave height -5: 2Mean wave period(+2) -5: 2Mean wave period(+1) -5: 2Peak frequency -5: 2Mean wave dir. a1b1 -5: 2Peak direction -5: 4Part. wave height -5: 4Part. peak period -5: 4Part. mean direction -5: 5Charnock parameter -5: ------------------------------------------------ -5: OUTPUT TIME : 2021/03/28 00:00:00 UTC -5: -5: End of file reached -5: -5: -5: *** End of Grid interpolation Routine *** -5: =============================================== -5: -5: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -5: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -5: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.gnh_10m ]] -5: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_gnh_10m/out_grd.gnh_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.gnh_10m -5: + cmdfile.6[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh gnh_10m 255 11 2021032800 108 global 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + cmdfile.6[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib2_gnh_10m.out -5: + bash[8]'[' -z '' ']' -5: + bash[9]case "$-" in -5: + bash[12]__lmod_vx=x -5: + bash[16]'[' -n x ']' -5: + bash[16]set +x -5: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -5: Shell debugging restarted -5: + bash[224]unset __lmod_vx -5: + wave_grib2_sbs.sh[30]grdID=gnh_10m -5: + wave_grib2_sbs.sh[31]GRIDNR=255 -5: + wave_grib2_sbs.sh[32]MODNR=11 -5: + wave_grib2_sbs.sh[33]valid_time=2021032800 -5: + wave_grib2_sbs.sh[34]fhr=108 -5: + wave_grib2_sbs.sh[35]grid_region=global -5: + wave_grib2_sbs.sh[36]grid_res=0p16 -5: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -5: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010 -5: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_gnh_10m -5: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_gnh_10m -5: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_gnh_10m -5: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_gnh_10m -5: ++ wave_grib2_sbs.sh[47]printf %03i 108 -5: + wave_grib2_sbs.sh[47]FH3=108 -5: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p16 -5: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16 -5: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p16.f108.grib2 -5: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f108.grib2 ]] -5: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.gnh_10m.inp.tmpl ./ww3_grib2.gnh_10m.inp.tmpl -5: + cpreq[4]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mod_def.gnh_10m ./mod_def.ww3 -5: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.gnh_10m ./out_grd.ww3 -5: + wave_grib2_sbs.sh[73]ngrib=1 -5: + wave_grib2_sbs.sh[74]dtgrib=3600 -5: + wave_grib2_sbs.sh[75]tstart='20210328 000000' -5: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210328 000000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.gnh_10m.inp.tmpl -5: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -5: $ WAVEWATCH-III gridded output input file -5: $ ---------------------------------------- -5: 20210328 000000 3600 1 -5: N -5: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -5: $ -5: 20210328 000000 7 11 255 0 0 -5: $ -5: $ end of input file -5: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[88]source prep_step -5: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -5: ++ prep_step[3]'[' -n OUTPUT.312689 ']' -5: ++ prep_step[4]echo gfs_ww3_grib.x -5: ++ prep_step[7]'[' -f errfile ']' -5: ++ prep_step[11]export FORT01=0 -5: ++ prep_step[11]FORT01=0 -5: +++ prep_step[12]awk -F= '{print $1}' -5: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -5: +++ prep_step[12]env -5: ++ prep_step[12]unset FORT01 -5: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -5: + wave_grib2_sbs.sh[90]export err=0 -5: + wave_grib2_sbs.sh[90]err=0 -5: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[95]cat grib2_global_108.out -5: -5: *** WAVEWATCH III GRIB output postp. *** -5: ============================================== -5: -5: Comment character is '$' -5: -5: Grid name : GFSv16-wave N Hemisphere 1/6 d -5: -5: LINEIN: -5: 20210328 000000 3600 1 -5: -5: 20210328000000 3600 1 -5: GEN_PRO -99999 -5: -5: Output time data : -5: ----------------------------------------------------- -5: First time : 2021/03/28 00:00:00 UTC -5: Interval : 01:00:00 -5: Number of requests : 1 -5: Fields : Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: Requested output fields not yet available: -5: ----------------------------------------------------- -5: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -5: -5: Successfully requested output fields : -5: ----------------------------------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: -5: Additional GRIB parameters : -5: ----------------------------------------------------- -5: Run time : 2021/03/28 00:00:00 UTC -5: GRIB center ID : 7 -5: GRIB gen. proc. ID : 11 -5: GRIB grid ID : 255 -5: GRIB GDS parameter : 0 -5: Fields in file : -5: -------------------------- -5: Current vel. -5: Wind speed -5: Ice concentration -5: Wave height -5: Mean wave period(+2) -5: Mean wave period(+1) -5: Peak frequency -5: Mean wave dir. a1b1 -5: Peak direction -5: Part. wave height -5: Part. peak period -5: Part. mean direction -5: Charnock parameter -5: -5: CHOSEN GRID TYPE: : LLRECTILINEAR -5: -5: -5: -5: Generating file -5: ----------------------------------------------------- -5: Data for 2021/03/28 00:00:00 UTC 0H forecast. -5: -5: End of program -5: ========================================= -5: WAVEWATCH III GRIB output -5: -5: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -5: + wave_grib2_sbs.sh[102][[ 108 -gt 0 ]] -5: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '108 hour fcst' -grib gfs.wave.t12z.global.0p16.f108.grib2 -5: 1:0:d=2021032312:SPC:surface:108 hour fcst: -5: 2:239839:d=2021032312:DIRC:surface:108 hour fcst: -5: 3:872843:d=2021032312:UOGRD:surface:108 hour fcst: -5: 4:1108114:d=2021032312:VOGRD:surface:108 hour fcst: -5: 5:1347356:d=2021032312:WIND:surface:108 hour fcst: -5: 6:1758017:d=2021032312:WDIR:surface:108 hour fcst: -5: 7:2397111:d=2021032312:UGRD:surface:108 hour fcst: -5: 8:2800658:d=2021032312:VGRD:surface:108 hour fcst: -5: 9:3205002:d=2021032312:ICEC:surface:108 hour fcst: -5: 10:3319447:d=2021032312:HTSGW:surface:108 hour fcst: -5: 11:3607099:d=2021032312:IMWF:surface:108 hour fcst: -5: 12:3920220:d=2021032312:MWSPER:surface:108 hour fcst: -5: 13:4235431:d=2021032312:PERPW:surface:108 hour fcst: -5: 14:4571681:d=2021032312:WWSDIR:surface:108 hour fcst: -5: 15:5150991:d=2021032312:DIRPW:surface:108 hour fcst: -5: 16:5754158:d=2021032312:WVHGT:surface:108 hour fcst: -5: 17:6030832:d=2021032312:SWELL:1 in sequence:108 hour fcst: -5: 18:6328604:d=2021032312:SWELL:2 in sequence:108 hour fcst: -5: 19:6535614:d=2021032312:SWELL:3 in sequence:108 hour fcst: -5: 20:6667737:d=2021032312:WVPER:surface:108 hour fcst: -5: 21:6977105:d=2021032312:SWPER:1 in sequence:108 hour fcst: -5: 22:7315339:d=2021032312:SWPER:2 in sequence:108 hour fcst: -5: 23:7570832:d=2021032312:SWPER:3 in sequence:108 hour fcst: -5: 24:7721104:d=2021032312:WVDIR:surface:108 hour fcst: -5: 25:8209232:d=2021032312:SWDIR:1 in sequence:108 hour fcst: -5: 26:8816707:d=2021032312:SWDIR:2 in sequence:108 hour fcst: -5: 27:9225359:d=2021032312:SWDIR:3 in sequence:108 hour fcst: -5: + wave_grib2_sbs.sh[104]err=0 -5: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -5: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p16.f108.grib2 -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f108.grib2 ]] -5: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p16.f108.grib2.idx ]] -5: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p16.f108.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f108.grib2 -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f108.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f108.grib2 = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f108.grib2 ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f108.grib2 -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f108.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f108.grib2.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f108.grib2.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f108.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f108.grib2 -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p16.f108.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f108.grib2.idx -5: + cpfs[3]'[' 2 -ne 2 ']' -5: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f108.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f108.grib2.idx = ./ ']' -5: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f108.grib2.idx ']' -5: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f108.grib2.idx -5: + cpfs[16]cp gfs.wave.t12z.global.0p16.f108.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f108.grib2.idx.cptmp -5: + cpfs[18]'[' 0 -ne 0 ']' -5: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f108.grib2.idx.cptmp -5: + cpfs[23]'[' 0 -ne 0 ']' -5: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f108.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p16/gfs.wave.t12z.global.0p16.f108.grib2.idx -5: + cpfs[28]'[' 0 -ne 0 ']' -5: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p16.f108.grib2 and gfs.wave.t12z.global.0p16.f108.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_gnh_10m to COM' -5: INFO: Copied gfs.wave.t12z.global.0p16.f108.grib2 and gfs.wave.t12z.global.0p16.f108.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_gnh_10m to COM -5: + wave_grib2_sbs.sh[130][[ gnh_10m == '' ]] -5: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -5: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p16.f108.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -5: INFO: gfs.wave.t12z.global.0p16.f108.grib2 is global.0p50 or SENDDBN is NO, no alert sent -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + cmdfile.4[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh glo_30m 2021032800 3600. 9999 -3: + cmdfile.4[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grid_interp_sbs.sh[25]grdID=glo_30m -3: + wave_grid_interp_sbs.sh[26]valid_time=2021032800 -3: + wave_grid_interp_sbs.sh[27]dt=3600. -3: + wave_grid_interp_sbs.sh[28]nst=9999 -3: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010 -3: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_glo_30m -3: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.uglo_100km ./out_grd.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mod_def.uglo_100km ./mod_def.uglo_100km -3: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -3: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mod_def.glo_30m ./mod_def.glo_30m -3: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ]] -3: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m'\''' -3: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m' -3: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.glo_30m ./WHTGRIDINT.bin -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grid_interp_sbs.sh[51]weights_found=1 -3: + wave_grid_interp_sbs.sh[59]ymdhms='20210328 000000' -3: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210328 000000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/glo_30m/g ww3_gint.inp.tmpl -3: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -3: $ Input file for interpolation of uglo_100km to glo_30m -3: $------------------------------------------------ -3: $ Start Time 3600. NSteps -3: 20210328 000000 3600. 9999 -3: $ Total number of grids -3: 2 -3: $ Grid extensions -3: 'uglo_100km' -3: 'glo_30m' -3: $ -3: 0 -3: $ -3: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[70]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.312689 ']' -3: ++ prep_step[4]echo gfs_ww3_gint.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -3: INFO: Executing 'gfs_ww3_gint.x' -3: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -3: + wave_grid_interp_sbs.sh[73]cat grid_interp.glo_30m.out -3: -3: *** WAVEWATCH III Grid interpolation *** -3: =============================================== -3: -3: Comment character is '$' -3: -3: Time Information : -3: --------------------------------------------- -3: Starting Time : 2021/03/28 00:00:00 UTC -3: Interval (in sec) : 3600.00 -3: Number of requests : 9999 -3: --------------------------------------------- -3: Number of grids (including output grid) = 2 -3: -3: -3: Extension for grid 1 is --> uglo_100km -3: -3: Grid Particulars are : -3: Dimensions = 45166 1 -3: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Extension for grid 2 is --> glo_30m -3: -3: Grid Particulars are : -3: Dimensions = 720 336 -3: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -3: Grid Closure = 2 ==> -1 None, 2 Simple, 8 Tripolar -3: -3: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -3: -3: -3: Preparing interpolation weights for output grid -3: Total number of wet points for interpolation 167619 -3: -3: -3: Variable: Grid Interpolation Map Units: 0.100E+01 -3: -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 336 | | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 291 | 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 | -3: 246 | 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 201 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 156 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 111 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 66 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: 21 | 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 | -3: | 0 0 0 0 0 0 | -3: +-------------------------------------------------------------------------------------------------------------------------+ -3: 1 32 63 94 125 156 187 218 249 280 311 342 373 404 435 466 497 528 559 590 621 652 683 714 -3: -3: -3: Interpolating fields .... -3: -3: Output group 1 -3: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -3: Output group 2 -3: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -3: Output group 3 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 4 -3: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -3: Output group 5 -3: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -3: Output group 6 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 7 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 8 -3: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -3: Output group 9 -3: Output variables skipped -3: Output group 10 -3: Output variables skipped -3: ------------------------------------------------ -3: 1Current vel. -3: 1Wind speed -3: 1Ice concentration -3: 2Wave height -3: 2Mean wave period(+2) -3: 2Mean wave period(+1) -3: 2Peak frequency -3: 2Mean wave dir. a1b1 -3: 2Peak direction -3: 4Part. wave height -3: 4Part. peak period -3: 4Part. mean direction -3: 5Charnock parameter -3: ------------------------------------------------ -3: OUTPUT TIME : 2021/03/28 00:00:00 UTC -3: -3: End of file reached -3: -3: -3: *** End of Grid interpolation Routine *** -3: =============================================== -3: -3: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -3: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -3: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.glo_30m ]] -3: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_glo_30m/out_grd.glo_30m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.glo_30m -3: + cmdfile.4[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh glo_30m 255 11 2021032800 108 global 0p50 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + cmdfile.4[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib2_glo_30m.out -3: + bash[8]'[' -z '' ']' -3: + bash[9]case "$-" in -3: + bash[12]__lmod_vx=x -3: + bash[16]'[' -n x ']' -3: + bash[16]set +x -3: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -3: Shell debugging restarted -3: + bash[224]unset __lmod_vx -3: + wave_grib2_sbs.sh[30]grdID=glo_30m -3: + wave_grib2_sbs.sh[31]GRIDNR=255 -3: + wave_grib2_sbs.sh[32]MODNR=11 -3: + wave_grib2_sbs.sh[33]valid_time=2021032800 -3: + wave_grib2_sbs.sh[34]fhr=108 -3: + wave_grib2_sbs.sh[35]grid_region=global -3: + wave_grib2_sbs.sh[36]grid_res=0p50 -3: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -3: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010 -3: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_glo_30m -3: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_glo_30m -3: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_glo_30m -3: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_glo_30m -3: ++ wave_grib2_sbs.sh[47]printf %03i 108 -3: + wave_grib2_sbs.sh[47]FH3=108 -3: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_global_0p50 -3: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50 -3: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.global.0p50.f108.grib2 -3: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f108.grib2 ]] -3: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.glo_30m.inp.tmpl ./ww3_grib2.glo_30m.inp.tmpl -3: + cpreq[4]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mod_def.glo_30m ./mod_def.ww3 -3: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.glo_30m ./out_grd.ww3 -3: + wave_grib2_sbs.sh[73]ngrib=1 -3: + wave_grib2_sbs.sh[74]dtgrib=3600 -3: + wave_grib2_sbs.sh[75]tstart='20210328 000000' -3: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210328 000000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.glo_30m.inp.tmpl -3: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -3: $ WAVEWATCH-III gridded output input file -3: $ ---------------------------------------- -3: 20210328 000000 3600 1 -3: N -3: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -3: $ -3: 20210328 000000 7 11 255 0 0 -3: $ -3: $ end of input file -3: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[88]source prep_step -3: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -3: ++ prep_step[3]'[' -n OUTPUT.312689 ']' -3: ++ prep_step[4]echo gfs_ww3_grib.x -3: ++ prep_step[7]'[' -f errfile ']' -3: ++ prep_step[11]export FORT01=0 -3: ++ prep_step[11]FORT01=0 -3: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -3: +++ prep_step[12]awk -F= '{print $1}' -3: +++ prep_step[12]env -3: ++ prep_step[12]unset FORT01 -3: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -3: + wave_grib2_sbs.sh[90]export err=0 -3: + wave_grib2_sbs.sh[90]err=0 -3: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[95]cat grib2_global_108.out -3: -3: *** WAVEWATCH III GRIB output postp. *** -3: ============================================== -3: -3: Comment character is '$' -3: -3: Grid name : Global 30 min wave grid -3: -3: LINEIN: -3: 20210328 000000 3600 1 -3: -3: 20210328000000 3600 1 -3: GEN_PRO -99999 -3: -3: Output time data : -3: ----------------------------------------------------- -3: First time : 2021/03/28 00:00:00 UTC -3: Interval : 01:00:00 -3: Number of requests : 1 -3: Fields : Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: Requested output fields not yet available: -3: ----------------------------------------------------- -3: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -3: -3: Successfully requested output fields : -3: ----------------------------------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: -3: Additional GRIB parameters : -3: ----------------------------------------------------- -3: Run time : 2021/03/28 00:00:00 UTC -3: GRIB center ID : 7 -3: GRIB gen. proc. ID : 11 -3: GRIB grid ID : 255 -3: GRIB GDS parameter : 0 -3: Fields in file : -3: -------------------------- -3: Current vel. -3: Wind speed -3: Ice concentration -3: Wave height -3: Mean wave period(+2) -3: Mean wave period(+1) -3: Peak frequency -3: Mean wave dir. a1b1 -3: Peak direction -3: Part. wave height -3: Part. peak period -3: Part. mean direction -3: Charnock parameter -3: -3: CHOSEN GRID TYPE: : LLRECTILINEAR -3: -3: -3: -3: Generating file -3: ----------------------------------------------------- -3: Data for 2021/03/28 00:00:00 UTC 0H forecast. -3: -3: End of program -3: ========================================= -3: WAVEWATCH III GRIB output -3: -3: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -3: + wave_grib2_sbs.sh[102][[ 108 -gt 0 ]] -3: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '108 hour fcst' -grib gfs.wave.t12z.global.0p50.f108.grib2 -3: 1:0:d=2021032312:SPC:surface:108 hour fcst: -3: 2:78415:d=2021032312:DIRC:surface:108 hour fcst: -3: 3:321307:d=2021032312:UOGRD:surface:108 hour fcst: -3: 4:397037:d=2021032312:VOGRD:surface:108 hour fcst: -3: 5:477069:d=2021032312:WIND:surface:108 hour fcst: -3: 6:635896:d=2021032312:WDIR:surface:108 hour fcst: -3: 7:872141:d=2021032312:UGRD:surface:108 hour fcst: -3: 8:1027723:d=2021032312:VGRD:surface:108 hour fcst: -3: 9:1184499:d=2021032312:ICEC:surface:108 hour fcst: -3: 10:1224563:d=2021032312:HTSGW:surface:108 hour fcst: -3: 11:1323214:d=2021032312:IMWF:surface:108 hour fcst: -3: 12:1433592:d=2021032312:MWSPER:surface:108 hour fcst: -3: 13:1544728:d=2021032312:PERPW:surface:108 hour fcst: -3: 14:1661633:d=2021032312:WWSDIR:surface:108 hour fcst: -3: 15:1862251:d=2021032312:DIRPW:surface:108 hour fcst: -3: 16:2067672:d=2021032312:WVHGT:surface:108 hour fcst: -3: 17:2163078:d=2021032312:SWELL:1 in sequence:108 hour fcst: -3: 18:2268548:d=2021032312:SWELL:2 in sequence:108 hour fcst: -3: 19:2343708:d=2021032312:SWELL:3 in sequence:108 hour fcst: -3: 20:2389063:d=2021032312:WVPER:surface:108 hour fcst: -3: 21:2496007:d=2021032312:SWPER:1 in sequence:108 hour fcst: -3: 22:2615577:d=2021032312:SWPER:2 in sequence:108 hour fcst: -3: 23:2710525:d=2021032312:SWPER:3 in sequence:108 hour fcst: -3: 24:2765124:d=2021032312:WVDIR:surface:108 hour fcst: -3: 25:2926278:d=2021032312:SWDIR:1 in sequence:108 hour fcst: -3: 26:3131454:d=2021032312:SWDIR:2 in sequence:108 hour fcst: -3: 27:3281850:d=2021032312:SWDIR:3 in sequence:108 hour fcst: -3: + wave_grib2_sbs.sh[104]err=0 -3: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -3: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.global.0p50.f108.grib2 -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f108.grib2 ]] -3: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.global.0p50.f108.grib2.idx ]] -3: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.global.0p50.f108.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f108.grib2 -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f108.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f108.grib2 = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f108.grib2 ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f108.grib2 -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f108.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f108.grib2.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f108.grib2.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f108.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f108.grib2 -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.global.0p50.f108.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f108.grib2.idx -3: + cpfs[3]'[' 2 -ne 2 ']' -3: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f108.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f108.grib2.idx = ./ ']' -3: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f108.grib2.idx ']' -3: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f108.grib2.idx -3: + cpfs[16]cp gfs.wave.t12z.global.0p50.f108.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f108.grib2.idx.cptmp -3: + cpfs[18]'[' 0 -ne 0 ']' -3: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f108.grib2.idx.cptmp -3: + cpfs[23]'[' 0 -ne 0 ']' -3: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f108.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/global.0p50/gfs.wave.t12z.global.0p50.f108.grib2.idx -3: + cpfs[28]'[' 0 -ne 0 ']' -3: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.global.0p50.f108.grib2 and gfs.wave.t12z.global.0p50.f108.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_glo_30m to COM' -3: INFO: Copied gfs.wave.t12z.global.0p50.f108.grib2 and gfs.wave.t12z.global.0p50.f108.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_glo_30m to COM -3: + wave_grib2_sbs.sh[130][[ glo_30m == '' ]] -3: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -3: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.global.0p50.f108.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -3: INFO: gfs.wave.t12z.global.0p50.f108.grib2 is global.0p50 or SENDDBN is NO, no alert sent -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + cmdfile.5[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh aoc_9km 2021032800 3600. 9999 -4: + cmdfile.5[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grid_interp_sbs.sh[25]grdID=aoc_9km -4: + wave_grid_interp_sbs.sh[26]valid_time=2021032800 -4: + wave_grid_interp_sbs.sh[27]dt=3600. -4: + wave_grid_interp_sbs.sh[28]nst=9999 -4: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010 -4: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_aoc_9km -4: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.uglo_100km ./out_grd.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mod_def.uglo_100km ./mod_def.uglo_100km -4: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -4: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mod_def.aoc_9km ./mod_def.aoc_9km -4: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km'\''' -4: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km' -4: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.aoc_9km ./WHTGRIDINT.bin -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grid_interp_sbs.sh[51]weights_found=1 -4: + wave_grid_interp_sbs.sh[59]ymdhms='20210328 000000' -4: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210328 000000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/aoc_9km/g ww3_gint.inp.tmpl -4: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -4: $ Input file for interpolation of uglo_100km to aoc_9km -4: $------------------------------------------------ -4: $ Start Time 3600. NSteps -4: 20210328 000000 3600. 9999 -4: $ Total number of grids -4: 2 -4: $ Grid extensions -4: 'uglo_100km' -4: 'aoc_9km' -4: $ -4: 0 -4: $ -4: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[70]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.312689 ']' -4: ++ prep_step[4]echo gfs_ww3_gint.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -4: INFO: Executing 'gfs_ww3_gint.x' -4: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -4: + wave_grid_interp_sbs.sh[73]cat grid_interp.aoc_9km.out -4: -4: *** WAVEWATCH III Grid interpolation *** -4: =============================================== -4: -4: Comment character is '$' -4: -4: Time Information : -4: --------------------------------------------- -4: Starting Time : 2021/03/28 00:00:00 UTC -4: Interval (in sec) : 3600.00 -4: Number of requests : 9999 -4: --------------------------------------------- -4: Number of grids (including output grid) = 2 -4: -4: -4: Extension for grid 1 is --> uglo_100km -4: -4: Grid Particulars are : -4: Dimensions = 45166 1 -4: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Extension for grid 2 is --> aoc_9km -4: -4: Grid Particulars are : -4: Dimensions = 1006 1006 -4: Grid Type = 2 ==> 1 Rect, 2 Curv, 3 Unstr -4: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -4: -4: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -4: -4: -4: Preparing interpolation weights for output grid -4: Total number of wet points for interpolation 360052 -4: -4: -4: Variable: Grid Interpolation Map Units: 0.100E+01 -4: -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: *** | | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: 880 | 0 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 | -4: 754 | 0 0 0 | -4: | 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 628 | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 | -4: 502 | 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 376 | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: 250 | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 0 0 | -4: 124 | 0 0 0 0 0 0 0 0 | -4: | 0 0 0 0 0 0 0 | -4: | 0 0 0 0 | -4: +-------------------------------------------------------------------------------------------------------------------------+ -4: 1 43 85 127 169 211 253 295 337 379 421 463 505 547 589 631 673 715 757 799 841 883 925 967 -4: -4: -4: Interpolating fields .... -4: -4: Output group 1 -4: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -4: Output group 2 -4: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -4: Output group 3 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 4 -4: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -4: Output group 5 -4: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -4: Output group 6 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 7 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 8 -4: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -4: Output group 9 -4: Output variables skipped -4: Output group 10 -4: Output variables skipped -4: ------------------------------------------------ -4: 1Current vel. -4: 1Wind speed -4: 1Ice concentration -4: 2Wave height -4: 2Mean wave period(+2) -4: 2Mean wave period(+1) -4: 2Peak frequency -4: 2Mean wave dir. a1b1 -4: 2Peak direction -4: 4Part. wave height -4: 4Part. peak period -4: 4Part. mean direction -4: 5Charnock parameter -4: ------------------------------------------------ -4: OUTPUT TIME : 2021/03/28 00:00:00 UTC -4: -4: End of file reached -4: -4: -4: *** End of Grid interpolation Routine *** -4: =============================================== -4: -4: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -4: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -4: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.aoc_9km ]] -4: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_aoc_9km/out_grd.aoc_9km /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.aoc_9km -4: + cmdfile.5[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh aoc_9km 255 11 2021032800 108 arctic 9km 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + cmdfile.5[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib2_aoc_9km.out -4: + bash[8]'[' -z '' ']' -4: + bash[9]case "$-" in -4: + bash[12]__lmod_vx=x -4: + bash[16]'[' -n x ']' -4: + bash[16]set +x -4: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -4: Shell debugging restarted -4: + bash[224]unset __lmod_vx -4: + wave_grib2_sbs.sh[30]grdID=aoc_9km -4: + wave_grib2_sbs.sh[31]GRIDNR=255 -4: + wave_grib2_sbs.sh[32]MODNR=11 -4: + wave_grib2_sbs.sh[33]valid_time=2021032800 -4: + wave_grib2_sbs.sh[34]fhr=108 -4: + wave_grib2_sbs.sh[35]grid_region=arctic -4: + wave_grib2_sbs.sh[36]grid_res=9km -4: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -4: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010 -4: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_aoc_9km -4: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_aoc_9km -4: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_aoc_9km -4: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_aoc_9km -4: ++ wave_grib2_sbs.sh[47]printf %03i 108 -4: + wave_grib2_sbs.sh[47]FH3=108 -4: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_arctic_9km -4: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km -4: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.arctic.9km.f108.grib2 -4: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f108.grib2 ]] -4: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.aoc_9km.inp.tmpl ./ww3_grib2.aoc_9km.inp.tmpl -4: + cpreq[4]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mod_def.aoc_9km ./mod_def.ww3 -4: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.aoc_9km ./out_grd.ww3 -4: + wave_grib2_sbs.sh[73]ngrib=1 -4: + wave_grib2_sbs.sh[74]dtgrib=3600 -4: + wave_grib2_sbs.sh[75]tstart='20210328 000000' -4: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210328 000000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.aoc_9km.inp.tmpl -4: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -4: $ WAVEWATCH-III gridded output input file -4: $ ---------------------------------------- -4: 20210328 000000 3600 1 -4: N -4: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -4: $ -4: 20210328 000000 7 11 255 0 20 -4: $ -4: 70 0 9.0 9.0 64 -4: $ 60 0 8.64919046313 8.64919046313 64 -4: $ end of input file -4: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[88]source prep_step -4: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -4: ++ prep_step[3]'[' -n OUTPUT.312689 ']' -4: ++ prep_step[4]echo gfs_ww3_grib.x -4: ++ prep_step[7]'[' -f errfile ']' -4: ++ prep_step[11]export FORT01=0 -4: ++ prep_step[11]FORT01=0 -4: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -4: +++ prep_step[12]awk -F= '{print $1}' -4: +++ prep_step[12]env -4: ++ prep_step[12]unset FORT01 -4: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -4: + wave_grib2_sbs.sh[90]export err=0 -4: + wave_grib2_sbs.sh[90]err=0 -4: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[95]cat grib2_arctic_108.out -4: -4: *** WAVEWATCH III GRIB output postp. *** -4: ============================================== -4: -4: Comment character is '$' -4: -4: Grid name : Arctic Ocean PolarStereo 9km -4: -4: LINEIN: -4: 20210328 000000 3600 1 -4: -4: 20210328000000 3600 1 -4: GEN_PRO -99999 -4: -4: Output time data : -4: ----------------------------------------------------- -4: First time : 2021/03/28 00:00:00 UTC -4: Interval : 01:00:00 -4: Number of requests : 1 -4: Fields : Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: Requested output fields not yet available: -4: ----------------------------------------------------- -4: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -4: -4: Successfully requested output fields : -4: ----------------------------------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: -4: Additional GRIB parameters : -4: ----------------------------------------------------- -4: Run time : 2021/03/28 00:00:00 UTC -4: GRIB center ID : 7 -4: GRIB gen. proc. ID : 11 -4: GRIB grid ID : 255 -4: GRIB GDS parameter : 0 -4: Fields in file : -4: -------------------------- -4: Current vel. -4: Wind speed -4: Ice concentration -4: Wave height -4: Mean wave period(+2) -4: Mean wave period(+1) -4: Peak frequency -4: Mean wave dir. a1b1 -4: Peak direction -4: Part. wave height -4: Part. peak period -4: Part. mean direction -4: Charnock parameter -4: -4: CHOSEN GRID TYPE: : POLARSTEREO -4: -4: -4: -4: Generating file -4: ----------------------------------------------------- -4: Data for 2021/03/28 00:00:00 UTC 0H forecast. -4: -4: End of program -4: ========================================= -4: WAVEWATCH III GRIB output -4: -4: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -4: + wave_grib2_sbs.sh[102][[ 108 -gt 0 ]] -4: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '108 hour fcst' -grib gfs.wave.t12z.arctic.9km.f108.grib2 -4: 1:0:d=2021032312:SPC:surface:108 hour fcst: -4: 2:191687:d=2021032312:DIRC:surface:108 hour fcst: -4: 3:650180:d=2021032312:UOGRD:surface:108 hour fcst: -4: 4:836609:d=2021032312:VOGRD:surface:108 hour fcst: -4: 5:1027795:d=2021032312:WIND:surface:108 hour fcst: -4: 6:1332085:d=2021032312:WDIR:surface:108 hour fcst: -4: 7:1761913:d=2021032312:UGRD:surface:108 hour fcst: -4: 8:2060522:d=2021032312:VGRD:surface:108 hour fcst: -4: 9:2364921:d=2021032312:ICEC:surface:108 hour fcst: -4: 10:2534072:d=2021032312:HTSGW:surface:108 hour fcst: -4: 11:2749988:d=2021032312:IMWF:surface:108 hour fcst: -4: 12:2969305:d=2021032312:MWSPER:surface:108 hour fcst: -4: 13:3190253:d=2021032312:PERPW:surface:108 hour fcst: -4: 14:3420519:d=2021032312:WWSDIR:surface:108 hour fcst: -4: 15:3732030:d=2021032312:DIRPW:surface:108 hour fcst: -4: 16:4050327:d=2021032312:WVHGT:surface:108 hour fcst: -4: 17:4264584:d=2021032312:SWELL:1 in sequence:108 hour fcst: -4: 18:4474043:d=2021032312:SWELL:2 in sequence:108 hour fcst: -4: 19:4637763:d=2021032312:SWELL:3 in sequence:108 hour fcst: -4: 20:4771592:d=2021032312:WVPER:surface:108 hour fcst: -4: 21:4995002:d=2021032312:SWPER:1 in sequence:108 hour fcst: -4: 22:5216338:d=2021032312:SWPER:2 in sequence:108 hour fcst: -4: 23:5392519:d=2021032312:SWPER:3 in sequence:108 hour fcst: -4: 24:5529528:d=2021032312:WVDIR:surface:108 hour fcst: -4: 25:5817985:d=2021032312:SWDIR:1 in sequence:108 hour fcst: -4: 26:6125849:d=2021032312:SWDIR:2 in sequence:108 hour fcst: -4: 27:6343169:d=2021032312:SWDIR:3 in sequence:108 hour fcst: -4: + wave_grib2_sbs.sh[104]err=0 -4: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -4: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.arctic.9km.f108.grib2 -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f108.grib2 ]] -4: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.arctic.9km.f108.grib2.idx ]] -4: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.arctic.9km.f108.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f108.grib2 -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f108.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f108.grib2 = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f108.grib2 ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f108.grib2 -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f108.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f108.grib2.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f108.grib2.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f108.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f108.grib2 -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.arctic.9km.f108.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f108.grib2.idx -4: + cpfs[3]'[' 2 -ne 2 ']' -4: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f108.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f108.grib2.idx = ./ ']' -4: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f108.grib2.idx ']' -4: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f108.grib2.idx -4: + cpfs[16]cp gfs.wave.t12z.arctic.9km.f108.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f108.grib2.idx.cptmp -4: + cpfs[18]'[' 0 -ne 0 ']' -4: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f108.grib2.idx.cptmp -4: + cpfs[23]'[' 0 -ne 0 ']' -4: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f108.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/arctic.9km/gfs.wave.t12z.arctic.9km.f108.grib2.idx -4: + cpfs[28]'[' 0 -ne 0 ']' -4: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.arctic.9km.f108.grib2 and gfs.wave.t12z.arctic.9km.f108.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_aoc_9km to COM' -4: INFO: Copied gfs.wave.t12z.arctic.9km.f108.grib2 and gfs.wave.t12z.arctic.9km.f108.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_aoc_9km to COM -4: + wave_grib2_sbs.sh[130][[ aoc_9km == '' ]] -4: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -4: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.arctic.9km.f108.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -4: INFO: gfs.wave.t12z.arctic.9km.f108.grib2 is global.0p50 or SENDDBN is NO, no alert sent -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + cmdfile.3[2]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grid_interp_sbs.sh wc_10m 2021032800 3600. 9999 -2: + cmdfile.3[3]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grid_interp_sbs.sh[25]grdID=wc_10m -2: + wave_grid_interp_sbs.sh[26]valid_time=2021032800 -2: + wave_grid_interp_sbs.sh[27]dt=3600. -2: + wave_grid_interp_sbs.sh[28]nst=9999 -2: + wave_grid_interp_sbs.sh[30]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010 -2: + wave_grid_interp_sbs.sh[31]interp_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[32]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[33]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[34]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_wc_10m -2: + wave_grid_interp_sbs.sh[37]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_gint.inp.tmpl ww3_gint.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[40]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.uglo_100km ./out_grd.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mod_def.uglo_100km ./mod_def.uglo_100km -2: + wave_grid_interp_sbs.sh[43]for ID in ${waveGRD} ${grdID} -2: + wave_grid_interp_sbs.sh[44]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mod_def.wc_10m ./mod_def.wc_10m -2: + wave_grid_interp_sbs.sh[48][[ -f /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ]] -2: + wave_grid_interp_sbs.sh[49]echo 'INFO: Interpolation weights found at: '\''/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m'\''' -2: INFO: Interpolation weights found at: '/work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m' -2: + wave_grid_interp_sbs.sh[50]cpreq /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/fix/wave/ww3_gint.WHTGRIDINT.bin.uglo_100km.wc_10m ./WHTGRIDINT.bin -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grid_interp_sbs.sh[51]weights_found=1 -2: + wave_grid_interp_sbs.sh[59]ymdhms='20210328 000000' -2: + wave_grid_interp_sbs.sh[60]sed -e 's/TIME/20210328 000000/g' -e s/DT/3600./g -e s/NSTEPS/9999/g -e s/GRIDIN/uglo_100km/g -e s/GRIDOUT/wc_10m/g ww3_gint.inp.tmpl -2: + wave_grid_interp_sbs.sh[66]cat ww3_gint.inp -2: $ Input file for interpolation of uglo_100km to wc_10m -2: $------------------------------------------------ -2: $ Start Time 3600. NSteps -2: 20210328 000000 3600. 9999 -2: $ Total number of grids -2: 2 -2: $ Grid extensions -2: 'uglo_100km' -2: 'wc_10m' -2: $ -2: 0 -2: $ -2: + wave_grid_interp_sbs.sh[69]export pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[69]pgm=gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[70]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_gint.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.312689 ']' -2: ++ prep_step[4]echo gfs_ww3_gint.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]env -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: ++ prep_step[12]unset FORT01 -2: + wave_grid_interp_sbs.sh[71]echo 'INFO: Executing '\''gfs_ww3_gint.x'\''' -2: INFO: Executing 'gfs_ww3_gint.x' -2: + wave_grid_interp_sbs.sh[72]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_gint.x -2: + wave_grid_interp_sbs.sh[73]cat grid_interp.wc_10m.out -2: -2: *** WAVEWATCH III Grid interpolation *** -2: =============================================== -2: -2: Comment character is '$' -2: -2: Time Information : -2: --------------------------------------------- -2: Starting Time : 2021/03/28 00:00:00 UTC -2: Interval (in sec) : 3600.00 -2: Number of requests : 9999 -2: --------------------------------------------- -2: Number of grids (including output grid) = 2 -2: -2: -2: Extension for grid 1 is --> uglo_100km -2: -2: Grid Particulars are : -2: Dimensions = 45166 1 -2: Grid Type = 3 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Extension for grid 2 is --> wc_10m -2: -2: Grid Particulars are : -2: Dimensions = 241 151 -2: Grid Type = 1 ==> 1 Rect, 2 Curv, 3 Unstr -2: Grid Closure = -1 ==> -1 None, 2 Simple, 8 Tripolar -2: -2: Interpolation scheme = 0 ==> 0 linear, 1 extrapolate unstructured, 2 nearest -2: -2: -2: Preparing interpolation weights for output grid -2: Total number of wet points for interpolation 11044 -2: -2: -2: Variable: Grid Interpolation Map Units: 0.100E+01 -2: -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: +---------------------------------------------------------------------------------------------------------------+ -2: 151 | 0 0 0 0 | -2: | 0 0 0 0 0 | -2: | 0 0 0 0 0 | -2: 130 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 109 | 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: 88 | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 | -2: 67 | 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 | -2: 46 | 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 | -2: 25 | 0 0 0 0 0 0 0 0 0 0 0 | -2: | 0 0 0 0 0 0 0 0 0 0 0 | -2: | | -2: 4 | | -2: +---------------------------------------------------------------------------------------------------------------+ -2: 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 -2: -2: -2: Interpolating fields .... -2: -2: Output group 1 -2: Output variable flags are -> F T T F F T F F F F F F F F F F F F F F -2: Output group 2 -2: Output variable flags are -> T F T F T T T F T F F F F F F F F F F F -2: Output group 3 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 4 -2: Output variable flags are -> T T F T F F F F F F F F F F F F F F F F -2: Output group 5 -2: Output variable flags are -> F T F F F F F F F F F F F F F F F F F F -2: Output group 6 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 7 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 8 -2: Output variable flags are -> F F F F F F F F F F F F F F F F F F F F -2: Output group 9 -2: Output variables skipped -2: Output group 10 -2: Output variables skipped -2: ------------------------------------------------ -2: 1Current vel. -2: 1Wind speed -2: 1Ice concentration -2: 2Wave height -2: 2Mean wave period(+2) -2: 2Mean wave period(+1) -2: 2Peak frequency -2: 2Mean wave dir. a1b1 -2: 2Peak direction -2: 4Part. wave height -2: 4Part. peak period -2: 4Part. mean direction -2: 5Charnock parameter -2: ------------------------------------------------ -2: OUTPUT TIME : 2021/03/28 00:00:00 UTC -2: -2: End of file reached -2: -2: -2: *** End of Grid interpolation Routine *** -2: =============================================== -2: -2: + wave_grid_interp_sbs.sh[74][[ 0 -ne 0 ]] -2: + wave_grid_interp_sbs.sh[79][[ 1 -eq 0 ]] -2: + wave_grid_interp_sbs.sh[84][[ -f ./out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[85][[ -f /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.wc_10m ]] -2: + wave_grid_interp_sbs.sh[89]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grid_interp_wc_10m/out_grd.wc_10m /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.wc_10m -2: + cmdfile.3[4]/work2/noaa/global/mterry/global-workflow_forked/ush/wave_grib2_sbs.sh wc_10m 255 11 2021032800 108 wcoast 0p16 'WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + cmdfile.3[5]cat /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib2_wc_10m.out -2: + bash[8]'[' -z '' ']' -2: + bash[9]case "$-" in -2: + bash[12]__lmod_vx=x -2: + bash[16]'[' -n x ']' -2: + bash[16]set +x -2: Shell debugging temporarily silenced: export LMOD_SH_DBG_ON=1 for this output (/apps/other/lmod/lmod/init/bash) -2: Shell debugging restarted -2: + bash[224]unset __lmod_vx -2: + wave_grib2_sbs.sh[30]grdID=wc_10m -2: + wave_grib2_sbs.sh[31]GRIDNR=255 -2: + wave_grib2_sbs.sh[32]MODNR=11 -2: + wave_grib2_sbs.sh[33]valid_time=2021032800 -2: + wave_grib2_sbs.sh[34]fhr=108 -2: + wave_grib2_sbs.sh[35]grid_region=wcoast -2: + wave_grib2_sbs.sh[36]grid_res=0p16 -2: + wave_grib2_sbs.sh[37]grib_flags='WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA' -2: + wave_grib2_sbs.sh[39]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010 -2: + wave_grib2_sbs.sh[41]grib_DATA=/work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_wc_10m -2: + wave_grib2_sbs.sh[42]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_wc_10m -2: + wave_grib2_sbs.sh[43]mkdir -p /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_wc_10m -2: + wave_grib2_sbs.sh[44]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_wc_10m -2: ++ wave_grib2_sbs.sh[47]printf %03i 108 -2: + wave_grib2_sbs.sh[47]FH3=108 -2: + wave_grib2_sbs.sh[50]com_varname=COMOUT_WAVE_GRID_wcoast_0p16 -2: + wave_grib2_sbs.sh[51]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[52]mkdir -p /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16 -2: + wave_grib2_sbs.sh[55]outfile=gfs.wave.t12z.wcoast.0p16.f108.grib2 -2: + wave_grib2_sbs.sh[58][[ -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f108.grib2 ]] -2: + wave_grib2_sbs.sh[64]cpreq /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[3]cp /work2/noaa/global/mterry/global-workflow_forked/parm/wave/ww3_grib2.wc_10m.inp.tmpl ./ww3_grib2.wc_10m.inp.tmpl -2: + cpreq[4]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[67]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/mod_def.wc_10m ./mod_def.ww3 -2: + wave_grib2_sbs.sh[70]/bin/ln -sf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/out_grd.wc_10m ./out_grd.ww3 -2: + wave_grib2_sbs.sh[73]ngrib=1 -2: + wave_grib2_sbs.sh[74]dtgrib=3600 -2: + wave_grib2_sbs.sh[75]tstart='20210328 000000' -2: + wave_grib2_sbs.sh[77]sed -e 's/TIME/20210328 000000/g' -e s/DT/3600/g -e s/NT/1/g -e s/GRIDNR/255/g -e s/MODNR/11/g -e 's/FLAGS/WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA/g' ww3_grib2.wc_10m.inp.tmpl -2: + wave_grib2_sbs.sh[84]cat ww3_grib.inp -2: $ WAVEWATCH-III gridded output input file -2: $ ---------------------------------------- -2: 20210328 000000 3600 1 -2: N -2: WND CUR ICE HS T01 T02 DIR FP DP PHS PTP PDIR CHA -2: $ -2: 20210328 000000 7 11 255 0 0 -2: $ -2: $ end of input file -2: + wave_grib2_sbs.sh[87]export pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[87]pgm=gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[88]source prep_step -2: ++ prep_step[3]'[' -n gfs_ww3_grib.x ']' -2: ++ prep_step[3]'[' -n OUTPUT.312689 ']' -2: ++ prep_step[4]echo gfs_ww3_grib.x -2: ++ prep_step[7]'[' -f errfile ']' -2: ++ prep_step[11]export FORT01=0 -2: ++ prep_step[11]FORT01=0 -2: +++ prep_step[12]grep '^FORT[0-9]\{1,\}=' -2: +++ prep_step[12]awk -F= '{print $1}' -2: +++ prep_step[12]env -2: ++ prep_step[12]unset FORT01 -2: + wave_grib2_sbs.sh[89]/work2/noaa/global/mterry/global-workflow_forked/exec/gfs_ww3_grib.x -2: + wave_grib2_sbs.sh[90]export err=0 -2: + wave_grib2_sbs.sh[90]err=0 -2: + wave_grib2_sbs.sh[91][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[95]cat grib2_wcoast_108.out -2: -2: *** WAVEWATCH III GRIB output postp. *** -2: ============================================== -2: -2: Comment character is '$' -2: -2: Grid name : West Coast 10 min wave grid -2: -2: LINEIN: -2: 20210328 000000 3600 1 -2: -2: 20210328000000 3600 1 -2: GEN_PRO -99999 -2: -2: Output time data : -2: ----------------------------------------------------- -2: First time : 2021/03/28 00:00:00 UTC -2: Interval : 01:00:00 -2: Number of requests : 1 -2: Fields : Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: Requested output fields not yet available: -2: ----------------------------------------------------- -2: Charnock parameter *** NOT YET CODED INTO WW3_GRIB *** -2: -2: Successfully requested output fields : -2: ----------------------------------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: -2: Additional GRIB parameters : -2: ----------------------------------------------------- -2: Run time : 2021/03/28 00:00:00 UTC -2: GRIB center ID : 7 -2: GRIB gen. proc. ID : 11 -2: GRIB grid ID : 255 -2: GRIB GDS parameter : 0 -2: Fields in file : -2: -------------------------- -2: Current vel. -2: Wind speed -2: Ice concentration -2: Wave height -2: Mean wave period(+2) -2: Mean wave period(+1) -2: Peak frequency -2: Mean wave dir. a1b1 -2: Peak direction -2: Part. wave height -2: Part. peak period -2: Part. mean direction -2: Charnock parameter -2: -2: CHOSEN GRID TYPE: : LLRECTILINEAR -2: -2: -2: -2: Generating file -2: ----------------------------------------------------- -2: Data for 2021/03/28 00:00:00 UTC 0H forecast. -2: -2: End of program -2: ========================================= -2: WAVEWATCH III GRIB output -2: -2: + wave_grib2_sbs.sh[97][[ ! -s gribfile ]] -2: + wave_grib2_sbs.sh[102][[ 108 -gt 0 ]] -2: + wave_grib2_sbs.sh[103]wgrib2 gribfile -set_date 2021032312 -set_ftime '108 hour fcst' -grib gfs.wave.t12z.wcoast.0p16.f108.grib2 -2: 1:0:d=2021032312:SPC:surface:108 hour fcst: -2: 2:7355:d=2021032312:DIRC:surface:108 hour fcst: -2: 3:24804:d=2021032312:UOGRD:surface:108 hour fcst: -2: 4:32253:d=2021032312:VOGRD:surface:108 hour fcst: -2: 5:39785:d=2021032312:WIND:surface:108 hour fcst: -2: 6:51894:d=2021032312:WDIR:surface:108 hour fcst: -2: 7:69348:d=2021032312:UGRD:surface:108 hour fcst: -2: 8:80721:d=2021032312:VGRD:surface:108 hour fcst: -2: 9:92728:d=2021032312:ICEC:surface:108 hour fcst: -2: 10:97458:d=2021032312:HTSGW:surface:108 hour fcst: -2: 11:106963:d=2021032312:IMWF:surface:108 hour fcst: -2: 12:117005:d=2021032312:MWSPER:surface:108 hour fcst: -2: 13:127038:d=2021032312:PERPW:surface:108 hour fcst: -2: 14:136977:d=2021032312:WWSDIR:surface:108 hour fcst: -2: 15:152601:d=2021032312:DIRPW:surface:108 hour fcst: -2: 16:168357:d=2021032312:WVHGT:surface:108 hour fcst: -2: 17:175744:d=2021032312:SWELL:1 in sequence:108 hour fcst: -2: 18:184436:d=2021032312:SWELL:2 in sequence:108 hour fcst: -2: 19:189627:d=2021032312:SWELL:3 in sequence:108 hour fcst: -2: 20:194531:d=2021032312:WVPER:surface:108 hour fcst: -2: 21:202360:d=2021032312:SWPER:1 in sequence:108 hour fcst: -2: 22:211029:d=2021032312:SWPER:2 in sequence:108 hour fcst: -2: 23:216498:d=2021032312:SWPER:3 in sequence:108 hour fcst: -2: 24:221418:d=2021032312:WVDIR:surface:108 hour fcst: -2: 25:231468:d=2021032312:SWDIR:1 in sequence:108 hour fcst: -2: 26:244871:d=2021032312:SWDIR:2 in sequence:108 hour fcst: -2: 27:250643:d=2021032312:SWDIR:3 in sequence:108 hour fcst: -2: + wave_grib2_sbs.sh[104]err=0 -2: + wave_grib2_sbs.sh[111][[ 0 -ne 0 ]] -2: + wave_grib2_sbs.sh[117]wgrib2 -s gfs.wave.t12z.wcoast.0p16.f108.grib2 -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f108.grib2 ]] -2: + wave_grib2_sbs.sh[120][[ -s gfs.wave.t12z.wcoast.0p16.f108.grib2.idx ]] -2: + wave_grib2_sbs.sh[121]cpfs gfs.wave.t12z.wcoast.0p16.f108.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f108.grib2 -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f108.grib2 = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f108.grib2 = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f108.grib2 ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f108.grib2 -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f108.grib2 /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f108.grib2.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f108.grib2.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f108.grib2.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f108.grib2 -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[122]cpfs gfs.wave.t12z.wcoast.0p16.f108.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f108.grib2.idx -2: + cpfs[3]'[' 2 -ne 2 ']' -2: + cpfs[8]'[' /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f108.grib2.idx = . -o /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f108.grib2.idx = ./ ']' -2: + cpfs[10]'[' -d /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f108.grib2.idx ']' -2: + cpfs[13]cpdstfile=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f108.grib2.idx -2: + cpfs[16]cp gfs.wave.t12z.wcoast.0p16.f108.grib2.idx /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f108.grib2.idx.cptmp -2: + cpfs[18]'[' 0 -ne 0 ']' -2: + cpfs[22]/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.6.0/envs/unified-env/install/intel/2021.9.0/prod-util-2.1.1-ddlo2cq/bin/fsync_file /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f108.grib2.idx.cptmp -2: + cpfs[23]'[' 0 -ne 0 ']' -2: + cpfs[27]mv /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f108.grib2.idx.cptmp /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/wcoast.0p16/gfs.wave.t12z.wcoast.0p16.f108.grib2.idx -2: + cpfs[28]'[' 0 -ne 0 ']' -2: + wave_grib2_sbs.sh[123]echo 'INFO: Copied gfs.wave.t12z.wcoast.0p16.f108.grib2 and gfs.wave.t12z.wcoast.0p16.f108.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_wc_10m to COM' -2: INFO: Copied gfs.wave.t12z.wcoast.0p16.f108.grib2 and gfs.wave.t12z.wcoast.0p16.f108.grib2.idx from /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010/grib_wc_10m to COM -2: + wave_grib2_sbs.sh[130][[ wc_10m == '' ]] -2: + wave_grib2_sbs.sh[151][[ NO == \Y\E\S ]] -2: + wave_grib2_sbs.sh[157]echo 'INFO: gfs.wave.t12z.wcoast.0p16.f108.grib2 is global.0p50 or SENDDBN is NO, no alert sent' -2: INFO: gfs.wave.t12z.wcoast.0p16.f108.grib2 is global.0p50 or SENDDBN is NO, no alert sent -+ run_mpmd.sh[113]exit 0 -+ run_mpmd.sh[1]postamble run_mpmd.sh 1753758325 0 -+ preamble.sh[62]set +x -End run_mpmd.sh at 03:05:36 with error code 0 (time elapsed: 00:00:11) -+ exgfs_wave_post_gridded_sbs.sh[122]true -+ exgfs_wave_post_gridded_sbs.sh[123]export err=0 -+ exgfs_wave_post_gridded_sbs.sh[123]err=0 -+ exgfs_wave_post_gridded_sbs.sh[124][[ 0 -ne 0 ]] -+ exgfs_wave_post_gridded_sbs.sh[130]com_varname=COMOUT_WAVE_GRID_gsouth_0p25 -+ exgfs_wave_post_gridded_sbs.sh[131]com_dir=/work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25 -+ exgfs_wave_post_gridded_sbs.sh[132]gribchk=gfs.wave.t12z.gsouth.0p25.f108.grib2 -+ exgfs_wave_post_gridded_sbs.sh[133][[ ! -s /work2/noaa/global/mterry/RUNTESTS/COMROOT/C48_S2SW/gfs.20210323/12//products/wave/gridded/gsouth.0p25/gfs.wave.t12z.gsouth.0p25.f108.grib2 ]] -+ exgfs_wave_post_gridded_sbs.sh[138]exit 0 -+ JGLOBAL_WAVE_POST_SBS[28]true -+ JGLOBAL_WAVE_POST_SBS[29]export err=0 -+ JGLOBAL_WAVE_POST_SBS[29]err=0 -+ JGLOBAL_WAVE_POST_SBS[30][[ 0 -ne 0 ]] -+ JGLOBAL_WAVE_POST_SBS[37]cd /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312 -+ JGLOBAL_WAVE_POST_SBS[38][[ NO != \Y\E\S ]] -+ JGLOBAL_WAVE_POST_SBS[39]rm -rf /work2/noaa/stmp/mterry/HERCULES/RUNDIRS/C48_S2SW/gfs.2021032312/wavepostsbs_f108.306010 -+ JGLOBAL_WAVE_POST_SBS[42]exit 0 -+ JGLOBAL_WAVE_POST_SBS[1]postamble /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS 1753758313 0 -+ preamble.sh[62]set +x -End /work2/noaa/global/mterry/global-workflow_forked/jobs/JGLOBAL_WAVE_POST_SBS at 03:05:36 with error code 0 (time elapsed: 00:00:23) diff --git a/ci/error_logs/HECULES_PR_LOGS/C48_S2SW_rocotostat.log b/ci/error_logs/HECULES_PR_LOGS/C48_S2SW_rocotostat.log deleted file mode 100644 index 28289f7e..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/C48_S2SW_rocotostat.log +++ /dev/null @@ -1,456 +0,0 @@ -2025-07-28 14:49:35,427 - INFO - root : [START] USER_THREAD_COUNT: mterry has 33/1028698 threads (0.0% utilization) -2025-07-28 14:49:36,200 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 14:49:36,200 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 14:49:36,941 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 34/1028698 threads (0.0% utilization) -2025-07-28 14:49:36,941 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 14:49:36,999 - INFO - root : [END] USER_THREAD_COUNT: mterry has 49/1028698 threads (0.0% utilization) -2025-07-28 14:55:38,243 - INFO - root : [START] USER_THREAD_COUNT: mterry has 102/1028698 threads (0.01% utilization) -2025-07-28 14:55:39,016 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 95/1028698 threads (0.01% utilization) -2025-07-28 14:55:39,017 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 14:55:39,810 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 79/1028698 threads (0.01% utilization) -2025-07-28 14:55:39,810 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 14:55:39,867 - INFO - root : [END] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 15:01:41,108 - INFO - root : [START] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 15:01:41,931 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 33/1028698 threads (0.0% utilization) -2025-07-28 15:01:41,931 - INFO - root : Rocoto call successful on attempt 1: call_time=0.76s, total_time=0.76s -2025-07-28 15:01:42,670 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 15:01:42,671 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 15:01:42,726 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 15:07:43,935 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 15:07:44,712 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 15:07:44,712 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 15:07:45,470 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 58/1028698 threads (0.01% utilization) -2025-07-28 15:07:45,470 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 15:07:45,529 - INFO - root : [END] USER_THREAD_COUNT: mterry has 58/1028698 threads (0.01% utilization) -2025-07-28 15:13:46,748 - INFO - root : [START] USER_THREAD_COUNT: mterry has 35/1028698 threads (0.0% utilization) -2025-07-28 15:13:47,516 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 15:13:47,516 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 15:13:48,255 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 15:13:48,255 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 15:13:48,311 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 15:19:49,495 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 15:19:50,260 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 15:19:50,261 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 15:19:50,996 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:19:50,997 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 15:19:51,052 - INFO - root : [END] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:25:52,265 - INFO - root : [START] USER_THREAD_COUNT: mterry has 49/1028698 threads (0.0% utilization) -2025-07-28 15:25:53,048 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 15:25:53,049 - INFO - root : Rocoto call successful on attempt 1: call_time=0.72s, total_time=0.72s -2025-07-28 15:25:53,803 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 15:25:53,803 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 15:25:53,861 - INFO - root : [END] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 15:31:55,047 - INFO - root : [START] USER_THREAD_COUNT: mterry has 95/1028698 threads (0.01% utilization) -2025-07-28 15:31:55,809 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 15:31:55,809 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 15:31:56,541 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:31:56,541 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 15:31:56,597 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:37:57,781 - INFO - root : [START] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 15:37:58,707 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 15:37:58,707 - INFO - root : Rocoto call successful on attempt 1: call_time=0.87s, total_time=0.87s -2025-07-28 15:37:59,435 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:37:59,435 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 15:37:59,489 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:44:00,670 - INFO - root : [START] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 15:44:01,424 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 15:44:01,424 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 15:44:02,179 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:44:02,179 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 15:44:02,237 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:50:03,439 - INFO - root : [START] USER_THREAD_COUNT: mterry has 64/1028698 threads (0.01% utilization) -2025-07-28 15:50:04,200 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 63/1028698 threads (0.01% utilization) -2025-07-28 15:50:04,200 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 15:50:04,931 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:50:04,931 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 15:50:04,986 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:56:06,196 - INFO - root : [START] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 15:56:06,956 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 15:56:06,956 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 15:56:07,710 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:56:07,710 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 15:56:07,766 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:02:08,954 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 16:02:09,714 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 16:02:09,714 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 16:02:10,451 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:02:10,451 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 16:02:10,507 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:08:11,699 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:08:12,455 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:08:12,455 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 16:08:13,188 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:08:13,188 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 16:08:13,244 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:14:14,999 - INFO - root : [START] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 16:14:15,765 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:14:15,766 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 16:14:16,511 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:14:16,512 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 16:14:16,567 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:20:17,762 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:20:18,516 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:20:18,516 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 16:20:19,247 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:20:19,247 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 16:20:19,300 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:20,491 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:21,241 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:21,241 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 16:26:21,969 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:21,969 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 16:26:22,021 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:32:23,308 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:32:24,064 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:32:24,064 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 16:32:24,791 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:32:24,791 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 16:32:24,844 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:38:26,052 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:38:26,815 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:38:26,816 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 16:38:27,540 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:38:27,540 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 16:38:27,592 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:28,786 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:29,539 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:29,539 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 16:44:30,268 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:30,268 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 16:44:30,320 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:31,510 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:32,266 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:32,266 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 16:50:32,999 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:32,999 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 16:50:33,051 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:34,255 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:35,006 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:35,006 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 16:56:35,736 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:35,736 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 16:56:35,788 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:36,981 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:37,735 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:37,735 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 17:02:38,466 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:38,467 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 17:02:38,519 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:39,721 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:40,473 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:40,473 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 17:08:41,200 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:41,200 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 17:08:41,253 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:42,501 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:43,254 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:43,254 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 17:14:43,985 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:43,985 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 17:14:44,037 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:45,231 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:45,983 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:45,983 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 17:20:46,712 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:46,712 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 17:20:46,765 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:47,954 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:48,710 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:48,710 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 17:26:49,439 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:49,440 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 17:26:49,491 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:32:50,670 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:32:51,446 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:32:51,447 - INFO - root : Rocoto call successful on attempt 1: call_time=0.72s, total_time=0.72s -2025-07-28 17:32:52,177 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:32:52,178 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 17:32:52,231 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:38:53,457 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:38:54,214 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:38:54,214 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 17:38:54,942 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:38:54,942 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 17:38:54,994 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:44:56,169 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:44:56,918 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:44:56,918 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 17:44:57,644 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:44:57,644 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 17:44:57,695 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:50:58,881 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:50:59,633 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:50:59,633 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 17:51:00,367 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:51:00,367 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 17:51:00,419 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:57:01,599 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 17:57:02,360 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 17:57:02,360 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 17:57:03,122 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 17:57:03,122 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 17:57:03,176 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:03:04,359 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:03:05,111 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:03:05,111 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 18:03:05,834 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:03:05,834 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 18:03:05,886 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:07,083 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:07,838 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:07,838 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 18:09:08,564 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:08,564 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 18:09:08,616 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:09,809 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:10,566 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:10,566 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 18:15:11,298 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:11,298 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 18:15:11,351 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:12,530 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:13,280 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:13,280 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 18:21:14,014 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:14,014 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 18:21:14,065 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:15,240 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:15,996 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:15,996 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 18:27:16,723 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:16,723 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 18:27:16,774 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:17,955 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:18,712 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:18,713 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 18:33:19,441 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:19,441 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 18:33:19,493 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:39:21,023 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:39:21,781 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:39:21,781 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 18:39:22,508 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:39:22,508 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 18:39:22,561 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:45:23,743 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:45:24,493 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:45:24,493 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 18:45:25,222 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:45:25,222 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 18:45:25,274 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:26,455 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:27,208 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:27,208 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 18:51:27,934 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:27,934 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 18:51:27,985 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:29,253 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:35,015 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:35,015 - INFO - root : Rocoto call successful on attempt 1: call_time=5.70s, total_time=5.70s -2025-07-28 18:57:35,745 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:35,745 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 18:57:35,796 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:03:37,353 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:03:38,127 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:03:38,127 - INFO - root : Rocoto call successful on attempt 1: call_time=0.72s, total_time=0.72s -2025-07-28 19:03:38,854 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:03:38,854 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 19:03:38,906 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:40,529 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:41,289 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:41,289 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 19:09:42,022 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:42,022 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 19:09:42,074 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:15:43,493 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:15:44,247 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:15:44,247 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 19:15:44,977 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:15:44,977 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 19:15:45,030 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:21:46,404 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:21:47,155 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:21:47,155 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 19:21:47,883 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:21:47,883 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 19:21:47,936 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:54,241 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:54,998 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 19:27:54,998 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 19:27:55,728 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 69/1028698 threads (0.01% utilization) -2025-07-28 19:27:55,728 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 19:27:55,781 - INFO - root : [END] USER_THREAD_COUNT: mterry has 69/1028698 threads (0.01% utilization) -2025-07-28 19:33:57,052 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:33:57,805 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:33:57,805 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 19:33:58,535 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 19:33:58,535 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 19:33:58,588 - INFO - root : [END] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 19:39:59,865 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:40:00,626 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:40:00,626 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 19:40:01,358 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 51/1028698 threads (0.0% utilization) -2025-07-28 19:40:01,358 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 19:40:01,413 - INFO - root : [END] USER_THREAD_COUNT: mterry has 51/1028698 threads (0.0% utilization) -2025-07-28 19:46:03,023 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:46:03,779 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:46:03,780 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 19:46:04,517 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 51/1028698 threads (0.0% utilization) -2025-07-28 19:46:04,517 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 19:46:04,571 - INFO - root : [END] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 19:52:06,297 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:52:07,064 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 19:52:07,064 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 19:52:07,838 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:52:07,838 - INFO - root : Rocoto call successful on attempt 1: call_time=0.72s, total_time=0.72s -2025-07-28 19:52:07,891 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:58:09,176 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:58:09,971 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:58:09,971 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 19:58:10,748 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:58:10,748 - INFO - root : Rocoto call successful on attempt 1: call_time=0.72s, total_time=0.72s -2025-07-28 19:58:10,800 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 20:04:12,098 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:04:12,854 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 25/1028698 threads (0.0% utilization) -2025-07-28 20:04:12,854 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 20:04:13,584 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 36/1028698 threads (0.0% utilization) -2025-07-28 20:04:13,584 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 20:04:13,637 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 20:10:14,932 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:10:15,690 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:10:15,690 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 20:10:16,428 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 20:10:16,428 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 20:10:16,482 - INFO - root : [END] USER_THREAD_COUNT: mterry has 36/1028698 threads (0.0% utilization) -2025-07-28 20:16:22,754 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:16:23,505 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 20:16:23,505 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 20:16:24,234 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 20:16:24,234 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 20:16:24,288 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 20:22:25,571 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:22:26,327 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 36/1028698 threads (0.0% utilization) -2025-07-28 20:22:26,327 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 20:22:27,059 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 36/1028698 threads (0.0% utilization) -2025-07-28 20:22:27,060 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 20:22:27,113 - INFO - root : [END] USER_THREAD_COUNT: mterry has 36/1028698 threads (0.0% utilization) -2025-07-28 20:28:28,376 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:28:29,131 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 66/1028698 threads (0.01% utilization) -2025-07-28 20:28:29,131 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 20:28:29,861 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 66/1028698 threads (0.01% utilization) -2025-07-28 20:28:29,861 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 20:28:29,914 - INFO - root : [END] USER_THREAD_COUNT: mterry has 66/1028698 threads (0.01% utilization) -2025-07-28 20:34:31,203 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:34:31,957 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 20:34:31,958 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 20:34:32,685 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 20:34:32,685 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 20:34:32,741 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 20:40:34,003 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:34,761 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:34,761 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 20:40:35,506 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 20:40:35,506 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 20:40:35,562 - INFO - root : [END] USER_THREAD_COUNT: mterry has 49/1028698 threads (0.0% utilization) -2025-07-28 20:46:36,911 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:37,680 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:37,680 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 20:46:38,411 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 20:46:38,412 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 20:46:38,465 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 20:52:39,757 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:40,509 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:40,509 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 20:52:41,236 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 20:52:41,236 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 20:52:41,289 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 20:58:42,570 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:43,322 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:43,322 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 20:58:44,049 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:44,049 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 20:58:44,101 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:45,470 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:46,230 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:46,230 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 21:04:46,964 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:46,964 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 21:04:47,018 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:48,452 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:49,204 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:49,204 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 21:10:49,930 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:49,930 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 21:10:49,982 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:16:51,262 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:52,015 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:52,015 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 21:16:52,741 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:52,741 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 21:16:52,794 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:22:54,274 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:55,032 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:55,032 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 21:22:55,760 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:55,760 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 21:22:55,813 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:28:57,279 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:28:58,033 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:28:58,033 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 21:28:58,759 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:28:58,759 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 21:28:58,811 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:35:00,552 - INFO - root : [START] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:35:01,306 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:35:01,306 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 21:35:02,052 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:35:02,052 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 21:35:02,117 - INFO - root : [END] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:41:03,834 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:41:04,601 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:41:04,601 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 21:41:05,365 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:41:05,365 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 21:41:05,417 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:47:06,715 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:47:07,465 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:47:07,465 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 21:47:08,189 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:47:08,189 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 21:47:08,241 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:53:09,600 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:53:10,359 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:53:10,359 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 21:53:11,087 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:53:11,087 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 21:53:11,139 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:59:12,345 - INFO - root : [START] USER_THREAD_COUNT: mterry has 53/1028698 threads (0.01% utilization) -2025-07-28 21:59:13,106 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:59:13,106 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 21:59:13,837 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:59:13,837 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 21:59:13,889 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:05:15,075 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 22:05:15,838 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:05:15,839 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 22:05:16,568 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:05:16,568 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 22:05:16,624 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:11:17,877 - INFO - root : [START] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 22:11:18,631 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:11:18,631 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 22:11:19,365 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:11:19,365 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 22:11:19,417 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:17:20,609 - INFO - root : [START] USER_THREAD_COUNT: mterry has 66/1028698 threads (0.01% utilization) -2025-07-28 22:17:21,363 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:17:21,363 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 22:17:22,088 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:17:22,088 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 22:17:22,140 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:23:23,494 - INFO - root : [START] USER_THREAD_COUNT: mterry has 53/1028698 threads (0.01% utilization) -2025-07-28 22:23:24,247 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:23:24,247 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 22:23:24,988 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:23:24,989 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 22:23:25,040 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) diff --git a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C48_ATM_logs_2021032312.log b/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C48_ATM_logs_2021032312.log deleted file mode 100644 index a532cfe8..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C48_ATM_logs_2021032312.log +++ /dev/null @@ -1,702 +0,0 @@ -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_stage_ic -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_stage_ic is pending at druby://130.18.14.151:36373 -2025-07-28 14:48:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_stage_ic is success, jobid=5950262 -2025-07-28 14:48:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_stage_ic, jobid=5950262, in state RUNNING (RUNNING) -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_stage_ic, jobid=5950262, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_fcst_seg0 -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_fcst_seg0 is pending at druby://130.18.14.151:43453 -2025-07-28 15:00:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_fcst_seg0 is success, jobid=5950281 -2025-07-28 15:00:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950281, in state QUEUED (PENDING) -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950281, in state RUNNING (RUNNING) -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950281, in state RUNNING (RUNNING) -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f000-f002 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f003-f005 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f006-f008 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f009-f011 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f012-f014 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f015-f017 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f018-f020 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f021-f023 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f024-f026 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f027-f029 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f030-f032 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f000-f002 succeeded, jobid=5950348 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f003-f005 succeeded, jobid=5950349 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f006-f008 succeeded, jobid=5950350 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f009-f011 succeeded, jobid=5950351 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f012-f014 succeeded, jobid=5950352 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f015-f017 succeeded, jobid=5950353 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f018-f020 succeeded, jobid=5950354 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f021-f023 succeeded, jobid=5950355 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f024-f026 succeeded, jobid=5950356 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f027-f029 succeeded, jobid=5950357 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f030-f032 succeeded, jobid=5950358 -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950281, in state RUNNING (RUNNING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950348, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950349, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950350, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950351, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950352, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5950353, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5950354, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5950355, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5950356, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5950357, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5950358, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f033-f035 -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f036-f038 -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f039-f041 -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f042-f044 -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f045-f047 -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f048-f054 -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f057-f063 -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f033-f035 succeeded, jobid=5950418 -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f036-f038 succeeded, jobid=5950419 -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f039-f041 succeeded, jobid=5950420 -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f042-f044 succeeded, jobid=5950421 -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f045-f047 succeeded, jobid=5950422 -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f048-f054 succeeded, jobid=5950423 -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f057-f063 succeeded, jobid=5950424 -2025-07-28 15:24:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950281, in state RUNNING (RUNNING) -2025-07-28 15:24:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950348, in state QUEUED (PENDING) -2025-07-28 15:24:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950349, in state QUEUED (PENDING) -2025-07-28 15:24:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950350, in state QUEUED (PENDING) -2025-07-28 15:24:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950351, in state QUEUED (PENDING) -2025-07-28 15:24:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950352, in state QUEUED (PENDING) -2025-07-28 15:24:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5950353, in state QUEUED (PENDING) -2025-07-28 15:24:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5950354, in state QUEUED (PENDING) -2025-07-28 15:24:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5950355, in state QUEUED (PENDING) -2025-07-28 15:24:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5950356, in state QUEUED (PENDING) -2025-07-28 15:24:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5950357, in state QUEUED (PENDING) -2025-07-28 15:24:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5950358, in state QUEUED (PENDING) -2025-07-28 15:24:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 15:24:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 15:24:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 15:24:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 15:24:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f066-f072 -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f075-f081 -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f084-f090 -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f093-f099 -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f066-f072 succeeded, jobid=5950441 -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f075-f081 succeeded, jobid=5950442 -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f084-f090 succeeded, jobid=5950443 -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f093-f099 succeeded, jobid=5950444 -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950281, in state SUCCEEDED (COMPLETED), ran for 1409.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950348, in state RUNNING (RUNNING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950349, in state RUNNING (RUNNING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950350, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950351, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950352, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5950353, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5950354, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5950355, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5950356, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5950357, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5950358, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f102-f108 -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f111-f114 -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f117-f120 -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_atmos_prod_f102-f108 is pending at druby://130.18.14.151:35547 -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_atmos_prod_f111-f114 is pending at druby://130.18.14.151:35547 -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_atmos_prod_f117-f120 is pending at druby://130.18.14.151:35547 -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_atmos_prod_f102-f108 is success, jobid=5950491 -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_atmos_prod_f111-f114 is success, jobid=5950492 -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_atmos_prod_f117-f120 is success, jobid=5950493 -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950348, in state SUCCEEDED (COMPLETED), ran for 271.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950349, in state SUCCEEDED (COMPLETED), ran for 272.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950350, in state RUNNING (RUNNING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950351, in state RUNNING (RUNNING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950352, in state RUNNING (RUNNING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5950353, in state RUNNING (RUNNING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5950354, in state RUNNING (RUNNING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5950355, in state RUNNING (RUNNING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5950356, in state RUNNING (RUNNING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5950357, in state RUNNING (RUNNING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5950358, in state RUNNING (RUNNING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950350, in state SUCCEEDED (COMPLETED), ran for 270.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950351, in state SUCCEEDED (COMPLETED), ran for 273.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950352, in state SUCCEEDED (COMPLETED), ran for 271.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5950353, in state SUCCEEDED (COMPLETED), ran for 269.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5950354, in state SUCCEEDED (COMPLETED), ran for 271.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5950355, in state SUCCEEDED (COMPLETED), ran for 272.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5950356, in state SUCCEEDED (COMPLETED), ran for 273.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5950357, in state SUCCEEDED (COMPLETED), ran for 272.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5950358, in state SUCCEEDED (COMPLETED), ran for 271.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 15:55:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 15:55:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 15:55:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 15:55:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 15:55:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 15:55:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 15:55:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 15:55:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 15:55:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 15:55:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 15:55:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 15:55:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 15:55:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 15:55:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 16:31:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 16:31:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 16:31:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 16:31:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 16:31:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 16:31:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 16:31:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 16:31:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 16:31:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 16:31:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 16:31:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 16:31:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 16:31:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 16:31:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 16:37:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 16:37:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 16:37:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 16:37:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 16:37:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 16:37:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 16:37:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 16:37:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 16:37:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 16:37:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 16:37:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 16:37:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 16:37:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 16:37:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 16:43:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 16:43:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 16:43:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 16:43:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 16:43:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 16:43:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 16:43:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 16:43:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 16:43:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 16:43:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 16:43:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 16:43:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 16:43:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 16:43:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 16:49:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 16:49:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 16:49:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 16:49:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 16:49:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 16:49:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 16:49:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 16:49:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 16:49:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 16:49:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 16:49:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 16:49:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 16:49:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 16:49:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 16:55:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 16:55:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 16:55:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 16:55:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 16:55:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 16:55:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 16:55:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 16:55:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 16:55:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 16:55:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 16:55:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 16:55:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 16:55:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 16:55:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 17:01:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 17:01:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 17:01:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 17:01:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 17:01:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 17:01:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 17:01:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 17:01:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 17:01:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 17:01:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 17:01:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 17:01:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 17:01:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 17:01:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 17:07:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 17:07:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 17:07:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 17:07:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 17:07:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 17:07:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 17:07:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 17:07:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 17:07:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 17:07:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 17:07:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 17:07:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 17:07:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 17:07:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 17:13:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 17:13:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 17:13:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 17:13:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 17:13:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 17:13:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 17:13:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 17:13:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 17:13:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 17:13:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 17:13:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 17:13:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 17:13:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 17:13:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 17:19:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 17:19:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 17:19:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 17:19:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 17:19:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 17:19:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 17:19:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 17:19:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 17:19:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 17:19:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 17:19:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 17:19:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 17:19:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 17:19:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 17:25:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 17:25:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 17:25:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 17:25:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 17:25:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 17:25:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 17:25:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 17:25:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 17:25:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 17:25:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 17:25:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 17:25:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 17:25:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 17:25:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 17:31:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 17:31:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 17:31:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 17:31:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 17:31:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 17:31:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 17:31:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 17:31:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 17:31:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 17:31:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 17:31:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 17:31:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 17:31:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 17:31:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 17:37:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 17:37:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 17:37:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 17:37:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 17:37:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 17:37:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 17:37:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 17:37:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 17:37:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 17:37:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 17:37:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 17:37:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 17:37:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 17:37:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 17:43:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 17:43:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 17:43:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 17:43:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 17:43:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 17:43:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 17:43:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 17:43:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 17:43:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 17:43:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 17:43:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 17:43:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 17:43:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 17:43:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 17:49:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 17:49:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 17:49:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 17:49:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 17:49:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 17:49:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 17:49:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 17:49:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 17:49:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 17:49:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 17:49:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 17:49:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 17:49:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 17:49:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 17:55:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 17:55:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 17:55:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 17:55:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 17:55:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 17:55:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 17:55:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 17:55:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 17:55:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 17:55:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 17:55:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 17:55:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 17:55:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 17:55:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 18:01:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 18:01:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 18:01:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 18:01:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 18:01:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 18:01:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 18:01:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 18:01:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 18:01:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 18:01:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 18:01:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 18:01:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 18:01:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 18:01:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 18:07:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 18:07:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 18:07:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 18:07:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 18:07:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 18:07:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 18:07:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 18:07:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 18:07:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 18:07:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 18:07:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 18:07:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 18:07:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 18:07:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 18:14:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 18:14:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 18:14:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 18:14:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 18:14:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 18:14:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 18:14:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 18:14:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 18:14:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 18:14:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 18:14:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 18:14:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 18:14:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 18:14:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 18:20:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 18:20:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 18:20:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 18:20:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 18:20:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 18:20:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 18:20:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 18:20:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 18:20:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 18:20:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 18:20:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 18:20:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 18:20:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 18:20:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 18:26:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 18:26:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 18:26:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 18:26:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 18:26:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 18:26:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 18:26:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 18:26:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 18:26:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 18:26:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 18:26:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 18:26:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 18:26:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 18:26:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 18:32:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 18:32:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 18:32:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 18:32:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 18:32:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 18:32:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 18:32:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 18:32:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 18:32:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 18:32:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 18:32:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 18:32:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 18:32:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 18:32:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 18:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 18:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 18:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 18:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 18:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 18:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 18:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 18:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 18:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 18:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 18:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 18:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 18:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 18:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 18:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 18:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 18:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 18:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 18:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 18:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 18:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 18:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 18:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 18:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 18:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 18:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 18:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 18:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 18:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 18:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 18:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 18:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 18:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 18:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 18:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 18:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 18:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 18:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 18:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 18:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 18:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 18:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 18:56:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 18:56:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 18:56:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 18:56:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 18:56:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 18:56:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 18:56:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 18:56:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 18:56:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 18:56:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 18:56:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 18:56:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 18:56:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 18:56:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 19:02:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 19:02:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 19:02:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 19:02:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 19:02:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 19:02:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 19:02:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 19:02:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 19:02:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 19:02:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 19:02:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 19:02:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 19:02:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 19:02:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 19:08:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 19:08:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 19:08:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 19:08:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 19:08:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 19:08:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 19:08:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 19:08:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 19:08:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 19:08:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 19:08:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 19:08:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 19:08:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 19:08:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 19:14:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state QUEUED (PENDING) -2025-07-28 19:14:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state QUEUED (PENDING) -2025-07-28 19:14:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state QUEUED (PENDING) -2025-07-28 19:14:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state QUEUED (PENDING) -2025-07-28 19:14:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state QUEUED (PENDING) -2025-07-28 19:14:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state QUEUED (PENDING) -2025-07-28 19:14:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state QUEUED (PENDING) -2025-07-28 19:14:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state QUEUED (PENDING) -2025-07-28 19:14:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state QUEUED (PENDING) -2025-07-28 19:14:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state QUEUED (PENDING) -2025-07-28 19:14:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state QUEUED (PENDING) -2025-07-28 19:14:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state QUEUED (PENDING) -2025-07-28 19:14:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state QUEUED (PENDING) -2025-07-28 19:14:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state QUEUED (PENDING) -2025-07-28 19:20:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state RUNNING (RUNNING) -2025-07-28 19:20:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state RUNNING (RUNNING) -2025-07-28 19:20:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state RUNNING (RUNNING) -2025-07-28 19:20:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state RUNNING (RUNNING) -2025-07-28 19:20:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state RUNNING (RUNNING) -2025-07-28 19:20:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state RUNNING (RUNNING) -2025-07-28 19:20:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state RUNNING (RUNNING) -2025-07-28 19:20:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state RUNNING (RUNNING) -2025-07-28 19:20:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state RUNNING (RUNNING) -2025-07-28 19:20:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state RUNNING (RUNNING) -2025-07-28 19:20:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state RUNNING (RUNNING) -2025-07-28 19:20:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state RUNNING (RUNNING) -2025-07-28 19:20:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state RUNNING (RUNNING) -2025-07-28 19:20:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state RUNNING (RUNNING) -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5950418, in state SUCCEEDED (COMPLETED), ran for 270.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5950419, in state SUCCEEDED (COMPLETED), ran for 276.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5950420, in state SUCCEEDED (COMPLETED), ran for 278.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5950421, in state SUCCEEDED (COMPLETED), ran for 271.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5950422, in state SUCCEEDED (COMPLETED), ran for 275.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5950423, in state SUCCEEDED (COMPLETED), ran for 322.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5950424, in state SUCCEEDED (COMPLETED), ran for 317.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5950441, in state SUCCEEDED (COMPLETED), ran for 318.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5950442, in state SUCCEEDED (COMPLETED), ran for 320.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5950443, in state SUCCEEDED (COMPLETED), ran for 318.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5950444, in state SUCCEEDED (COMPLETED), ran for 315.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5950491, in state SUCCEEDED (COMPLETED), ran for 316.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5950492, in state SUCCEEDED (COMPLETED), ran for 214.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5950493, in state SUCCEEDED (COMPLETED), ran for 214.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_tracker -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_genesis -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_tracker is pending at druby://130.18.14.151:41653 -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_genesis is pending at druby://130.18.14.151:41653 -2025-07-28 19:32:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_tracker is success, jobid=5951165 -2025-07-28 19:32:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_genesis is success, jobid=5951166 -2025-07-28 19:32:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_tracker, jobid=5951165, in state SUCCEEDED (COMPLETED), ran for 63.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:32:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_genesis, jobid=5951166, in state RUNNING (RUNNING) -2025-07-28 19:38:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_genesis, jobid=5951166, in state SUCCEEDED (COMPLETED), ran for 481.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:38:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_arch_vrfy -2025-07-28 19:38:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_arch_vrfy is pending at druby://130.18.14.151:35659 -2025-07-28 19:44:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_arch_vrfy is success, jobid=5951173 -2025-07-28 19:44:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_arch_vrfy, jobid=5951173, in state SUCCEEDED (COMPLETED), ran for 68.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:44:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_cleanup -2025-07-28 19:44:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_cleanup is pending at druby://130.18.14.151:44783 -2025-07-28 19:50:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_cleanup is success, jobid=5951188 -2025-07-28 19:50:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_cleanup, jobid=5951188, in state SUCCEEDED (COMPLETED), ran for 21.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:50:41 -0500 :: hercules-login-1.hpc.msstate.edu :: This cycle is complete: Success diff --git a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C48_S2SWA_gefs_logs_2021032312.log b/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C48_S2SWA_gefs_logs_2021032312.log deleted file mode 100644 index 89103e68..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C48_S2SWA_gefs_logs_2021032312.log +++ /dev/null @@ -1,4372 +0,0 @@ -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_stage_ic -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_init -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_prep_emissions -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_stage_ic succeeded, jobid=5950271 -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_init succeeded, jobid=5950272 -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_prep_emissions succeeded, jobid=5950273 -2025-07-28 14:48:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_stage_ic, jobid=5950271, in state RUNNING (RUNNING) -2025-07-28 14:48:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_init, jobid=5950272, in state QUEUED (PENDING) -2025-07-28 14:48:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_prep_emissions, jobid=5950273, in state QUEUED (PENDING) -2025-07-28 14:54:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_stage_ic, jobid=5950271, in state SUCCEEDED (COMPLETED), ran for 16.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 14:54:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_init, jobid=5950272, in state QUEUED (PENDING) -2025-07-28 14:54:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_prep_emissions, jobid=5950273, in state QUEUED (PENDING) -2025-07-28 15:00:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_init, jobid=5950272, in state SUCCEEDED (COMPLETED), ran for 30.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:00:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_prep_emissions, jobid=5950273, in state SUCCEEDED (COMPLETED), ran for 35.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:00:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_fcst_mem000_seg0 -2025-07-28 15:00:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_fcst_mem001_seg0 -2025-07-28 15:00:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_fcst_mem002_seg0 -2025-07-28 15:00:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_fcst_mem000_seg0 succeeded, jobid=5950302 -2025-07-28 15:00:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_fcst_mem001_seg0 succeeded, jobid=5950303 -2025-07-28 15:00:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_fcst_mem002_seg0 succeeded, jobid=5950304 -2025-07-28 15:06:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 15:06:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 15:06:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 15:12:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 15:12:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 15:12:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 15:18:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 15:18:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 15:18:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 15:24:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 15:24:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 15:24:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 15:30:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 15:30:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 15:30:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 15:36:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 15:36:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 15:36:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 15:43:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 15:43:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 15:43:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 15:49:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 15:49:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 15:49:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 15:55:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 15:55:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 15:55:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 16:19:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 16:19:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 16:19:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 16:25:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 16:25:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 16:25:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 16:31:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 16:31:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 16:31:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 16:37:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 16:37:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 16:37:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 16:43:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 16:43:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 16:43:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 16:49:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 16:49:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 16:49:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 16:55:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 16:55:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 16:55:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 17:01:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 17:01:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 17:01:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 17:07:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 17:07:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 17:07:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 17:13:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 17:13:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 17:13:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 17:19:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 17:19:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 17:19:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 17:25:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 17:25:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 17:25:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 17:32:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 17:32:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 17:32:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 17:38:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 17:38:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 17:38:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 17:44:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 17:44:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 17:44:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 17:50:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 17:50:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 17:50:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 17:56:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 17:56:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 17:56:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 18:02:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 18:02:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 18:02:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 18:08:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 18:08:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 18:08:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 18:14:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 18:14:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 18:14:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 18:20:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 18:20:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 18:20:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 18:26:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 18:26:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 18:26:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 18:32:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 18:32:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 18:32:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 18:38:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 18:38:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 18:38:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 18:44:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 18:44:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 18:44:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 18:50:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state QUEUED (PENDING) -2025-07-28 18:50:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state QUEUED (PENDING) -2025-07-28 18:50:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state QUEUED (PENDING) -2025-07-28 18:56:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state RUNNING (RUNNING) -2025-07-28 18:56:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state RUNNING (RUNNING) -2025-07-28 18:56:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state RUNNING (RUNNING) -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state RUNNING (RUNNING) -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state RUNNING (RUNNING) -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state RUNNING (RUNNING) -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem000_f000 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem000_f006 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem000_f012 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem001_f000 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem001_f006 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem001_f012 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem002_f000 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem002_f006 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem002_f012 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f000-f004 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f005-f009 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f010-f014 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f015-f019 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f020-f024 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f000-f004 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f005-f009 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f010-f014 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f015-f019 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f020-f024 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f000-f004 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f005-f009 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f010-f014 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem000_f000 succeeded, jobid=5951018 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem000_f006 succeeded, jobid=5951019 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem000_f012 succeeded, jobid=5951020 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem001_f000 succeeded, jobid=5951021 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem001_f006 succeeded, jobid=5951022 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem001_f012 succeeded, jobid=5951023 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem002_f000 succeeded, jobid=5951024 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem002_f006 succeeded, jobid=5951025 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem002_f012 succeeded, jobid=5951026 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f000-f004 succeeded, jobid=5951027 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f005-f009 succeeded, jobid=5951028 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f010-f014 succeeded, jobid=5951029 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f015-f019 succeeded, jobid=5951030 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f020-f024 succeeded, jobid=5951031 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f000-f004 succeeded, jobid=5951032 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f005-f009 succeeded, jobid=5951033 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f010-f014 succeeded, jobid=5951034 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f015-f019 succeeded, jobid=5951035 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f020-f024 succeeded, jobid=5951036 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f000-f004 succeeded, jobid=5951037 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f005-f009 succeeded, jobid=5951038 -2025-07-28 19:02:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f010-f014 succeeded, jobid=5951039 -2025-07-28 19:08:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg0, jobid=5950302, in state SUCCEEDED (COMPLETED), ran for 883.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:08:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg0, jobid=5950303, in state SUCCEEDED (COMPLETED), ran for 903.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:08:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg0, jobid=5950304, in state SUCCEEDED (COMPLETED), ran for 894.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:08:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f000, jobid=5951018, in state QUEUED (PENDING) -2025-07-28 19:08:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f006, jobid=5951019, in state QUEUED (PENDING) -2025-07-28 19:08:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f012, jobid=5951020, in state QUEUED (PENDING) -2025-07-28 19:08:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f000, jobid=5951021, in state QUEUED (PENDING) -2025-07-28 19:08:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f006, jobid=5951022, in state QUEUED (PENDING) -2025-07-28 19:08:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f012, jobid=5951023, in state QUEUED (PENDING) -2025-07-28 19:08:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f000, jobid=5951024, in state QUEUED (PENDING) -2025-07-28 19:08:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f006, jobid=5951025, in state QUEUED (PENDING) -2025-07-28 19:08:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f012, jobid=5951026, in state QUEUED (PENDING) -2025-07-28 19:08:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 19:08:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 19:08:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 19:08:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_fcst_mem000_seg1 -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_fcst_mem001_seg1 -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_fcst_mem002_seg1 -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_fcst_mem000_seg1 succeeded, jobid=5951067 -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_fcst_mem001_seg1 succeeded, jobid=5951068 -2025-07-28 19:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_fcst_mem002_seg1 succeeded, jobid=5951069 -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f000, jobid=5951018, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f006, jobid=5951019, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f012, jobid=5951020, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f000, jobid=5951021, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f006, jobid=5951022, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f012, jobid=5951023, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f000, jobid=5951024, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f006, jobid=5951025, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f012, jobid=5951026, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f000, jobid=5951018, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f006, jobid=5951019, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f012, jobid=5951020, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f000, jobid=5951021, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f006, jobid=5951022, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f012, jobid=5951023, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f000, jobid=5951024, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f006, jobid=5951025, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f012, jobid=5951026, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f000, jobid=5951018, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f006, jobid=5951019, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f012, jobid=5951020, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f000, jobid=5951021, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f006, jobid=5951022, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f012, jobid=5951023, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f000, jobid=5951024, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f006, jobid=5951025, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f012, jobid=5951026, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f000, jobid=5951018, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f006, jobid=5951019, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f012, jobid=5951020, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f000, jobid=5951021, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f006, jobid=5951022, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f012, jobid=5951023, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f000, jobid=5951024, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f006, jobid=5951025, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f012, jobid=5951026, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:33:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f000, jobid=5951018, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f006, jobid=5951019, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f012, jobid=5951020, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f000, jobid=5951021, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f006, jobid=5951022, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f012, jobid=5951023, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f000, jobid=5951024, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f006, jobid=5951025, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f012, jobid=5951026, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f000, jobid=5951018, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f006, jobid=5951019, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f012, jobid=5951020, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f000, jobid=5951021, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f006, jobid=5951022, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f012, jobid=5951023, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f000, jobid=5951024, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f006, jobid=5951025, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f012, jobid=5951026, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f000, jobid=5951018, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f006, jobid=5951019, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f012, jobid=5951020, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f000, jobid=5951021, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f006, jobid=5951022, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f012, jobid=5951023, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f000, jobid=5951024, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f006, jobid=5951025, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f012, jobid=5951026, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f000, jobid=5951018, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f006, jobid=5951019, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f012, jobid=5951020, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f000, jobid=5951021, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f006, jobid=5951022, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f012, jobid=5951023, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f000, jobid=5951024, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f006, jobid=5951025, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f012, jobid=5951026, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f000, jobid=5951018, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f006, jobid=5951019, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f012, jobid=5951020, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f000, jobid=5951021, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f006, jobid=5951022, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f012, jobid=5951023, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f000, jobid=5951024, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f006, jobid=5951025, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f012, jobid=5951026, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f000, jobid=5951018, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f006, jobid=5951019, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f012, jobid=5951020, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f000, jobid=5951021, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f006, jobid=5951022, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f012, jobid=5951023, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f000, jobid=5951024, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f006, jobid=5951025, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f012, jobid=5951026, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f000, jobid=5951018, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f006, jobid=5951019, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f012, jobid=5951020, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f000, jobid=5951021, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f006, jobid=5951022, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f012, jobid=5951023, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f000, jobid=5951024, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f006, jobid=5951025, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f012, jobid=5951026, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f000, jobid=5951018, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f006, jobid=5951019, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f012, jobid=5951020, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f000, jobid=5951021, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f006, jobid=5951022, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f012, jobid=5951023, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f000, jobid=5951024, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f006, jobid=5951025, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f012, jobid=5951026, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f000, jobid=5951018, in state SUCCEEDED (COMPLETED), ran for 95.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f006, jobid=5951019, in state SUCCEEDED (COMPLETED), ran for 95.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f012, jobid=5951020, in state SUCCEEDED (COMPLETED), ran for 95.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f000, jobid=5951021, in state SUCCEEDED (COMPLETED), ran for 93.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f006, jobid=5951022, in state SUCCEEDED (COMPLETED), ran for 95.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f012, jobid=5951023, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f000, jobid=5951024, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f006, jobid=5951025, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f012, jobid=5951026, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem000_f018 -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem000_f024 -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem000_f030 -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem000_f036 -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem000_f042 -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem000_f018 succeeded, jobid=5951351 -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem000_f024 succeeded, jobid=5951352 -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem000_f030 succeeded, jobid=5951353 -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem000_f036 succeeded, jobid=5951354 -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem000_f042 succeeded, jobid=5951355 -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f018, jobid=5951351, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f024, jobid=5951352, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f030, jobid=5951353, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f036, jobid=5951354, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f042, jobid=5951355, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f012, jobid=5951023, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f000, jobid=5951024, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f006, jobid=5951025, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f012, jobid=5951026, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f018, jobid=5951351, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f024, jobid=5951352, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f030, jobid=5951353, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f036, jobid=5951354, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f042, jobid=5951355, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f012, jobid=5951023, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f000, jobid=5951024, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f006, jobid=5951025, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f012, jobid=5951026, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f018, jobid=5951351, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f024, jobid=5951352, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f030, jobid=5951353, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f036, jobid=5951354, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f042, jobid=5951355, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f012, jobid=5951023, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f000, jobid=5951024, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f006, jobid=5951025, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f012, jobid=5951026, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f018, jobid=5951351, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f024, jobid=5951352, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f030, jobid=5951353, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f036, jobid=5951354, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f042, jobid=5951355, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f012, jobid=5951023, in state SUCCEEDED (COMPLETED), ran for 96.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f000, jobid=5951024, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f006, jobid=5951025, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f012, jobid=5951026, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem000_f048 -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem000_f048 succeeded, jobid=5951483 -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f018, jobid=5951351, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f024, jobid=5951352, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f030, jobid=5951353, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f036, jobid=5951354, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f042, jobid=5951355, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f048, jobid=5951483, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f000, jobid=5951024, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f006, jobid=5951025, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f012, jobid=5951026, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f018, jobid=5951351, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f024, jobid=5951352, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f030, jobid=5951353, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f036, jobid=5951354, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f042, jobid=5951355, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f048, jobid=5951483, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f000, jobid=5951024, in state SUCCEEDED (COMPLETED), ran for 97.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f006, jobid=5951025, in state SUCCEEDED (COMPLETED), ran for 96.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f012, jobid=5951026, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem001_f018 -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem001_f024 -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_ensstat_f000, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_ensstat_f006, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem001_f018 succeeded, jobid=5951516 -2025-07-28 21:03:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem001_f024 succeeded, jobid=5951517 -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f018, jobid=5951351, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f024, jobid=5951352, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f030, jobid=5951353, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f036, jobid=5951354, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f042, jobid=5951355, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f048, jobid=5951483, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f018, jobid=5951516, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f024, jobid=5951517, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f012, jobid=5951026, in state SUCCEEDED (COMPLETED), ran for 100.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem001_f030 -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_ensstat_f000, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_ensstat_f006, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_ensstat_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem001_f030 succeeded, jobid=5951546 -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f018, jobid=5951351, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f024, jobid=5951352, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f030, jobid=5951353, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f036, jobid=5951354, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f042, jobid=5951355, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f048, jobid=5951483, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f018, jobid=5951516, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f024, jobid=5951517, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f030, jobid=5951546, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_ensstat_f000, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_ensstat_f006, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_ensstat_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state QUEUED (PENDING) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state QUEUED (PENDING) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state QUEUED (PENDING) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f018, jobid=5951351, in state QUEUED (PENDING) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f024, jobid=5951352, in state QUEUED (PENDING) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f030, jobid=5951353, in state QUEUED (PENDING) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f036, jobid=5951354, in state QUEUED (PENDING) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f042, jobid=5951355, in state QUEUED (PENDING) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f048, jobid=5951483, in state QUEUED (PENDING) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f018, jobid=5951516, in state QUEUED (PENDING) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f024, jobid=5951517, in state QUEUED (PENDING) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f030, jobid=5951546, in state QUEUED (PENDING) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f000-f004, jobid=5951027, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f005-f009, jobid=5951028, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f010-f014, jobid=5951029, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f015-f019, jobid=5951030, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f020-f024, jobid=5951031, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f000-f004, jobid=5951032, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f005-f009, jobid=5951033, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f010-f014, jobid=5951034, in state SUCCEEDED (COMPLETED), ran for 51.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f015-f019, jobid=5951035, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f020-f024, jobid=5951036, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f000-f004, jobid=5951037, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f005-f009, jobid=5951038, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state QUEUED (PENDING) -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem001_f036 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem001_f042 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem001_f048 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem002_f018 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem002_f024 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem002_f030 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem002_f036 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem002_f042 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem002_f048 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_ensstat_f000 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_ensstat_f006 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_ensstat_f012 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem001_f036 succeeded, jobid=5951628 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem001_f042 succeeded, jobid=5951629 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem001_f048 succeeded, jobid=5951630 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem002_f018 succeeded, jobid=5951631 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem002_f024 succeeded, jobid=5951632 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem002_f030 succeeded, jobid=5951633 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem002_f036 succeeded, jobid=5951634 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem002_f042 succeeded, jobid=5951635 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem002_f048 succeeded, jobid=5951636 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_ensstat_f000 succeeded, jobid=5951637 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_ensstat_f006 succeeded, jobid=5951638 -2025-07-28 21:21:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_ensstat_f012 succeeded, jobid=5951639 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state RUNNING (RUNNING) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state RUNNING (RUNNING) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state RUNNING (RUNNING) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f018, jobid=5951351, in state SUCCEEDED (COMPLETED), ran for 98.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f024, jobid=5951352, in state SUCCEEDED (COMPLETED), ran for 102.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f030, jobid=5951353, in state SUCCEEDED (COMPLETED), ran for 97.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f036, jobid=5951354, in state SUCCEEDED (COMPLETED), ran for 98.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f042, jobid=5951355, in state SUCCEEDED (COMPLETED), ran for 98.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f048, jobid=5951483, in state SUCCEEDED (COMPLETED), ran for 98.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f018, jobid=5951516, in state SUCCEEDED (COMPLETED), ran for 98.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f024, jobid=5951517, in state SUCCEEDED (COMPLETED), ran for 101.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f030, jobid=5951546, in state SUCCEEDED (COMPLETED), ran for 96.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f036, jobid=5951628, in state SUCCEEDED (COMPLETED), ran for 97.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f042, jobid=5951629, in state SUCCEEDED (COMPLETED), ran for 99.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f048, jobid=5951630, in state SUCCEEDED (COMPLETED), ran for 97.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f018, jobid=5951631, in state SUCCEEDED (COMPLETED), ran for 99.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f024, jobid=5951632, in state SUCCEEDED (COMPLETED), ran for 99.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f030, jobid=5951633, in state SUCCEEDED (COMPLETED), ran for 99.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f036, jobid=5951634, in state SUCCEEDED (COMPLETED), ran for 100.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f042, jobid=5951635, in state SUCCEEDED (COMPLETED), ran for 99.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f048, jobid=5951636, in state SUCCEEDED (COMPLETED), ran for 101.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f000, jobid=5951637, in state SUCCEEDED (COMPLETED), ran for 32.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f006, jobid=5951638, in state SUCCEEDED (COMPLETED), ran for 43.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f012, jobid=5951639, in state SUCCEEDED (COMPLETED), ran for 43.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f010-f014, jobid=5951039, in state SUCCEEDED (COMPLETED), ran for 44.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_ensstat_f018 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_ensstat_f024 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_ensstat_f030 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_ensstat_f036 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_ensstat_f042 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_ensstat_f048 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ocean_prod_mem000_f024 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ocean_prod_mem000_f048 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ocean_prod_mem001_f024 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ocean_prod_mem001_f048 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ocean_prod_mem002_f024 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ocean_prod_mem002_f048 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ice_prod_mem000_f024 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ice_prod_mem000_f048 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ice_prod_mem001_f024 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ice_prod_mem001_f048 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ice_prod_mem002_f024 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ice_prod_mem002_f048 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f025-f029 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f030-f034 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f035-f039 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f040-f044 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_ensstat_f018 succeeded, jobid=5951678 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_ensstat_f024 succeeded, jobid=5951679 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_ensstat_f030 succeeded, jobid=5951680 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_ensstat_f036 succeeded, jobid=5951681 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_ensstat_f042 succeeded, jobid=5951682 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_ensstat_f048 succeeded, jobid=5951683 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ocean_prod_mem000_f024 succeeded, jobid=5951684 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ocean_prod_mem000_f048 succeeded, jobid=5951685 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ocean_prod_mem001_f024 succeeded, jobid=5951686 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ocean_prod_mem001_f048 succeeded, jobid=5951687 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ocean_prod_mem002_f024 succeeded, jobid=5951688 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ocean_prod_mem002_f048 succeeded, jobid=5951689 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ice_prod_mem000_f024 succeeded, jobid=5951690 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ice_prod_mem000_f048 succeeded, jobid=5951691 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ice_prod_mem001_f024 succeeded, jobid=5951692 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ice_prod_mem001_f048 succeeded, jobid=5951693 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ice_prod_mem002_f024 succeeded, jobid=5951694 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ice_prod_mem002_f048 succeeded, jobid=5951695 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f025-f029 succeeded, jobid=5951696 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f030-f034 succeeded, jobid=5951697 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f035-f039 succeeded, jobid=5951698 -2025-07-28 21:27:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f040-f044 succeeded, jobid=5951699 -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state RUNNING (RUNNING) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state RUNNING (RUNNING) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state RUNNING (RUNNING) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f018, jobid=5951678, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f024, jobid=5951679, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f030, jobid=5951680, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f036, jobid=5951681, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f042, jobid=5951682, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f048, jobid=5951683, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ocean_prod_mem000_f024, jobid=5951684, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ocean_prod_mem000_f048, jobid=5951685, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ocean_prod_mem001_f024, jobid=5951686, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ocean_prod_mem001_f048, jobid=5951687, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ocean_prod_mem002_f024, jobid=5951688, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ocean_prod_mem002_f048, jobid=5951689, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ice_prod_mem000_f024, jobid=5951690, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ice_prod_mem000_f048, jobid=5951691, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ice_prod_mem001_f024, jobid=5951692, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ice_prod_mem001_f048, jobid=5951693, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ice_prod_mem002_f024, jobid=5951694, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ice_prod_mem002_f048, jobid=5951695, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f025-f029, jobid=5951696, in state QUEUED (PENDING) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f030-f034, jobid=5951697, in state QUEUED (PENDING) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f035-f039, jobid=5951698, in state QUEUED (PENDING) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f040-f044, jobid=5951699, in state QUEUED (PENDING) -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem000_f054 -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem000_f060 -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem000_f066 -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem001_f054 -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem001_f060 -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem001_f066 -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem002_f054 -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem002_f060 -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f045-f048 -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f049-f053 -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f054-f058 -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f059-f063 -2025-07-28 21:34:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f064-f068 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f069-f073 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f025-f029 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f030-f034 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f035-f039 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f040-f044 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem000_f054 succeeded, jobid=5951742 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem000_f060 succeeded, jobid=5951743 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem000_f066 succeeded, jobid=5951744 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem001_f054 succeeded, jobid=5951745 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem001_f060 succeeded, jobid=5951746 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem001_f066 succeeded, jobid=5951747 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem002_f054 succeeded, jobid=5951748 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem002_f060 succeeded, jobid=5951749 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f045-f048 succeeded, jobid=5951750 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f049-f053 succeeded, jobid=5951751 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f054-f058 succeeded, jobid=5951752 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f059-f063 succeeded, jobid=5951753 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f064-f068 succeeded, jobid=5951754 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f069-f073 succeeded, jobid=5951755 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f025-f029 succeeded, jobid=5951756 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f030-f034 succeeded, jobid=5951757 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f035-f039 succeeded, jobid=5951758 -2025-07-28 21:34:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f040-f044 succeeded, jobid=5951759 -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state RUNNING (RUNNING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state RUNNING (RUNNING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state RUNNING (RUNNING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f054, jobid=5951742, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f060, jobid=5951743, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f066, jobid=5951744, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f054, jobid=5951745, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f060, jobid=5951746, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f066, jobid=5951747, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f054, jobid=5951748, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f060, jobid=5951749, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f025-f029, jobid=5951696, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f030-f034, jobid=5951697, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f035-f039, jobid=5951698, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f040-f044, jobid=5951699, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f045-f048, jobid=5951750, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f049-f053, jobid=5951751, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f054-f058, jobid=5951752, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f059-f063, jobid=5951753, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f064-f068, jobid=5951754, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f069-f073, jobid=5951755, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f025-f029, jobid=5951756, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f030-f034, jobid=5951757, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f035-f039, jobid=5951758, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f040-f044, jobid=5951759, in state QUEUED (PENDING) -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 21:40:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem000_seg1, jobid=5951067, in state SUCCEEDED (COMPLETED), ran for 1256.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem001_seg1, jobid=5951068, in state SUCCEEDED (COMPLETED), ran for 1276.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_fcst_mem002_seg1, jobid=5951069, in state SUCCEEDED (COMPLETED), ran for 1305.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f054, jobid=5951742, in state QUEUED (PENDING) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f060, jobid=5951743, in state QUEUED (PENDING) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f066, jobid=5951744, in state QUEUED (PENDING) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f054, jobid=5951745, in state QUEUED (PENDING) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f060, jobid=5951746, in state QUEUED (PENDING) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f066, jobid=5951747, in state QUEUED (PENDING) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f054, jobid=5951748, in state QUEUED (PENDING) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f060, jobid=5951749, in state QUEUED (PENDING) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f025-f029, jobid=5951696, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f030-f034, jobid=5951697, in state QUEUED (PENDING) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f035-f039, jobid=5951698, in state QUEUED (PENDING) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f040-f044, jobid=5951699, in state QUEUED (PENDING) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f045-f048, jobid=5951750, in state QUEUED (PENDING) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f049-f053, jobid=5951751, in state QUEUED (PENDING) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f054-f058, jobid=5951752, in state QUEUED (PENDING) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f059-f063, jobid=5951753, in state QUEUED (PENDING) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f064-f068, jobid=5951754, in state QUEUED (PENDING) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f069-f073, jobid=5951755, in state QUEUED (PENDING) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f025-f029, jobid=5951756, in state QUEUED (PENDING) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f030-f034, jobid=5951757, in state QUEUED (PENDING) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f035-f039, jobid=5951758, in state QUEUED (PENDING) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f040-f044, jobid=5951759, in state QUEUED (PENDING) -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem000_f072 -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem000_f078 -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem000_f084 -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem000_f090 -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem000_f072 succeeded, jobid=5951827 -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem000_f078 succeeded, jobid=5951828 -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem000_f084 succeeded, jobid=5951829 -2025-07-28 21:46:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem000_f090 succeeded, jobid=5951830 -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f054, jobid=5951742, in state SUCCEEDED (COMPLETED), ran for 93.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f060, jobid=5951743, in state QUEUED (PENDING) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f066, jobid=5951744, in state QUEUED (PENDING) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f072, jobid=5951827, in state QUEUED (PENDING) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f078, jobid=5951828, in state QUEUED (PENDING) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f084, jobid=5951829, in state QUEUED (PENDING) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f090, jobid=5951830, in state QUEUED (PENDING) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f054, jobid=5951745, in state QUEUED (PENDING) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f060, jobid=5951746, in state QUEUED (PENDING) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f066, jobid=5951747, in state QUEUED (PENDING) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f054, jobid=5951748, in state QUEUED (PENDING) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f060, jobid=5951749, in state QUEUED (PENDING) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f030-f034, jobid=5951697, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f035-f039, jobid=5951698, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f040-f044, jobid=5951699, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f045-f048, jobid=5951750, in state QUEUED (PENDING) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f049-f053, jobid=5951751, in state QUEUED (PENDING) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f054-f058, jobid=5951752, in state QUEUED (PENDING) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f059-f063, jobid=5951753, in state QUEUED (PENDING) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f064-f068, jobid=5951754, in state QUEUED (PENDING) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f069-f073, jobid=5951755, in state QUEUED (PENDING) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f025-f029, jobid=5951756, in state QUEUED (PENDING) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f030-f034, jobid=5951757, in state QUEUED (PENDING) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f035-f039, jobid=5951758, in state QUEUED (PENDING) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f040-f044, jobid=5951759, in state QUEUED (PENDING) -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem000_f096 -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem000_f102 -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem000_f108 -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem000_f114 -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem000_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem000_f096 succeeded, jobid=5951885 -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem000_f102 succeeded, jobid=5951886 -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem000_f108 succeeded, jobid=5951887 -2025-07-28 21:52:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem000_f114 succeeded, jobid=5951888 -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f060, jobid=5951743, in state SUCCEEDED (COMPLETED), ran for 96.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f066, jobid=5951744, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f072, jobid=5951827, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f078, jobid=5951828, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f084, jobid=5951829, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f090, jobid=5951830, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f096, jobid=5951885, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f102, jobid=5951886, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f108, jobid=5951887, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f114, jobid=5951888, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f054, jobid=5951745, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f060, jobid=5951746, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f066, jobid=5951747, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f054, jobid=5951748, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f060, jobid=5951749, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f045-f048, jobid=5951750, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f049-f053, jobid=5951751, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f054-f058, jobid=5951752, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f059-f063, jobid=5951753, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f064-f068, jobid=5951754, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f069-f073, jobid=5951755, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f025-f029, jobid=5951756, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f030-f034, jobid=5951757, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f035-f039, jobid=5951758, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f040-f044, jobid=5951759, in state QUEUED (PENDING) -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem000_f120 -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 21:58:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem000_f120 succeeded, jobid=5951906 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f066, jobid=5951744, in state SUCCEEDED (COMPLETED), ran for 97.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f072, jobid=5951827, in state QUEUED (PENDING) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f078, jobid=5951828, in state QUEUED (PENDING) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f084, jobid=5951829, in state QUEUED (PENDING) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f090, jobid=5951830, in state QUEUED (PENDING) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f096, jobid=5951885, in state QUEUED (PENDING) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f102, jobid=5951886, in state QUEUED (PENDING) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f108, jobid=5951887, in state QUEUED (PENDING) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f114, jobid=5951888, in state QUEUED (PENDING) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f120, jobid=5951906, in state QUEUED (PENDING) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f054, jobid=5951745, in state SUCCEEDED (COMPLETED), ran for 96.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f060, jobid=5951746, in state SUCCEEDED (COMPLETED), ran for 95.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f066, jobid=5951747, in state SUCCEEDED (COMPLETED), ran for 95.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f054, jobid=5951748, in state SUCCEEDED (COMPLETED), ran for 96.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f060, jobid=5951749, in state SUCCEEDED (COMPLETED), ran for 94.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f045-f048, jobid=5951750, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f049-f053, jobid=5951751, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f054-f058, jobid=5951752, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f059-f063, jobid=5951753, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f064-f068, jobid=5951754, in state SUCCEEDED (COMPLETED), ran for 48.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f069-f073, jobid=5951755, in state SUCCEEDED (COMPLETED), ran for 48.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f025-f029, jobid=5951756, in state SUCCEEDED (COMPLETED), ran for 48.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f030-f034, jobid=5951757, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f035-f039, jobid=5951758, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f040-f044, jobid=5951759, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem001_f072 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem001_f078 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem001_f084 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem001_f090 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem001_f096 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem001_f102 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem001_f108 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem001_f114 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem001_f120 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem002_f066 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem002_f072 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem002_f078 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem002_f084 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem002_f090 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem002_f096 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem002_f102 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_ensstat_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_ensstat_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem001_f072 succeeded, jobid=5951916 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem001_f078 succeeded, jobid=5951917 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem001_f084 succeeded, jobid=5951918 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem001_f090 succeeded, jobid=5951919 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem001_f096 succeeded, jobid=5951920 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem001_f102 succeeded, jobid=5951921 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem001_f108 succeeded, jobid=5951922 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem001_f114 succeeded, jobid=5951923 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem001_f120 succeeded, jobid=5951924 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem002_f066 succeeded, jobid=5951925 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem002_f072 succeeded, jobid=5951926 -2025-07-28 22:04:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem002_f078 succeeded, jobid=5951927 -2025-07-28 22:04:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem002_f084 succeeded, jobid=5951928 -2025-07-28 22:04:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem002_f090 succeeded, jobid=5951929 -2025-07-28 22:04:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem002_f096 succeeded, jobid=5951930 -2025-07-28 22:04:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem002_f102 succeeded, jobid=5951931 -2025-07-28 22:10:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f072, jobid=5951827, in state RUNNING (RUNNING) -2025-07-28 22:10:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f078, jobid=5951828, in state RUNNING (RUNNING) -2025-07-28 22:10:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f084, jobid=5951829, in state RUNNING (RUNNING) -2025-07-28 22:10:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f090, jobid=5951830, in state RUNNING (RUNNING) -2025-07-28 22:10:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f096, jobid=5951885, in state QUEUED (PENDING) -2025-07-28 22:10:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f102, jobid=5951886, in state QUEUED (PENDING) -2025-07-28 22:10:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f108, jobid=5951887, in state QUEUED (PENDING) -2025-07-28 22:10:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f114, jobid=5951888, in state QUEUED (PENDING) -2025-07-28 22:10:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f120, jobid=5951906, in state QUEUED (PENDING) -2025-07-28 22:10:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f072, jobid=5951916, in state QUEUED (PENDING) -2025-07-28 22:10:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f078, jobid=5951917, in state QUEUED (PENDING) -2025-07-28 22:10:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f084, jobid=5951918, in state QUEUED (PENDING) -2025-07-28 22:10:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f090, jobid=5951919, in state QUEUED (PENDING) -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f096, jobid=5951920, in state QUEUED (PENDING) -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f102, jobid=5951921, in state QUEUED (PENDING) -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f108, jobid=5951922, in state QUEUED (PENDING) -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f114, jobid=5951923, in state QUEUED (PENDING) -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f120, jobid=5951924, in state QUEUED (PENDING) -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f066, jobid=5951925, in state QUEUED (PENDING) -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f072, jobid=5951926, in state QUEUED (PENDING) -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f078, jobid=5951927, in state QUEUED (PENDING) -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f084, jobid=5951928, in state QUEUED (PENDING) -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f090, jobid=5951929, in state QUEUED (PENDING) -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f096, jobid=5951930, in state QUEUED (PENDING) -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f102, jobid=5951931, in state QUEUED (PENDING) -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_ensstat_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_ensstat_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 22:10:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f072, jobid=5951827, in state SUCCEEDED (COMPLETED), ran for 96.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f078, jobid=5951828, in state SUCCEEDED (COMPLETED), ran for 96.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f084, jobid=5951829, in state SUCCEEDED (COMPLETED), ran for 96.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f090, jobid=5951830, in state SUCCEEDED (COMPLETED), ran for 94.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f096, jobid=5951885, in state QUEUED (PENDING) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f102, jobid=5951886, in state QUEUED (PENDING) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f108, jobid=5951887, in state QUEUED (PENDING) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f114, jobid=5951888, in state QUEUED (PENDING) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f120, jobid=5951906, in state QUEUED (PENDING) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f072, jobid=5951916, in state QUEUED (PENDING) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f078, jobid=5951917, in state QUEUED (PENDING) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f084, jobid=5951918, in state QUEUED (PENDING) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f090, jobid=5951919, in state QUEUED (PENDING) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f096, jobid=5951920, in state QUEUED (PENDING) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f102, jobid=5951921, in state QUEUED (PENDING) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f108, jobid=5951922, in state QUEUED (PENDING) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f114, jobid=5951923, in state QUEUED (PENDING) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f120, jobid=5951924, in state QUEUED (PENDING) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f066, jobid=5951925, in state QUEUED (PENDING) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f072, jobid=5951926, in state QUEUED (PENDING) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f078, jobid=5951927, in state QUEUED (PENDING) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f084, jobid=5951928, in state QUEUED (PENDING) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f090, jobid=5951929, in state QUEUED (PENDING) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f096, jobid=5951930, in state QUEUED (PENDING) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f102, jobid=5951931, in state QUEUED (PENDING) -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem002_f108 -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem002_f114 -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_prod_mem002_f120 -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_ensstat_f054 -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_ensstat_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem002_f108 succeeded, jobid=5951988 -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem002_f114 succeeded, jobid=5951989 -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_prod_mem002_f120 succeeded, jobid=5951990 -2025-07-28 22:16:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_ensstat_f054 succeeded, jobid=5951991 -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f096, jobid=5951885, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f102, jobid=5951886, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f108, jobid=5951887, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f114, jobid=5951888, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f120, jobid=5951906, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f072, jobid=5951916, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f078, jobid=5951917, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f084, jobid=5951918, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f090, jobid=5951919, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f096, jobid=5951920, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f102, jobid=5951921, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f108, jobid=5951922, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f114, jobid=5951923, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f120, jobid=5951924, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f066, jobid=5951925, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f072, jobid=5951926, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f078, jobid=5951927, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f084, jobid=5951928, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f090, jobid=5951929, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f096, jobid=5951930, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f102, jobid=5951931, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f108, jobid=5951988, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f114, jobid=5951989, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f120, jobid=5951990, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f054, jobid=5951991, in state QUEUED (PENDING) -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_atmos_ensstat_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 22:22:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f096, jobid=5951885, in state SUCCEEDED (COMPLETED), ran for 93.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f102, jobid=5951886, in state SUCCEEDED (COMPLETED), ran for 96.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f108, jobid=5951887, in state SUCCEEDED (COMPLETED), ran for 92.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f114, jobid=5951888, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f120, jobid=5951906, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f072, jobid=5951916, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f078, jobid=5951917, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f084, jobid=5951918, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f090, jobid=5951919, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f096, jobid=5951920, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f102, jobid=5951921, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f108, jobid=5951922, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f114, jobid=5951923, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f120, jobid=5951924, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f066, jobid=5951925, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f072, jobid=5951926, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f078, jobid=5951927, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f084, jobid=5951928, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f090, jobid=5951929, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f096, jobid=5951930, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f102, jobid=5951931, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f108, jobid=5951988, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f114, jobid=5951989, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f120, jobid=5951990, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f054, jobid=5951991, in state QUEUED (PENDING) -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_ensstat_f060 -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ocean_prod_mem000_f072 -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ocean_prod_mem000_f096 -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem000_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_ensstat_f060 succeeded, jobid=5952011 -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ocean_prod_mem000_f072 succeeded, jobid=5952012 -2025-07-28 22:28:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ocean_prod_mem000_f096 succeeded, jobid=5952013 -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f114, jobid=5951888, in state SUCCEEDED (COMPLETED), ran for 90.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f120, jobid=5951906, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f072, jobid=5951916, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f078, jobid=5951917, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f084, jobid=5951918, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f090, jobid=5951919, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f096, jobid=5951920, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f102, jobid=5951921, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f108, jobid=5951922, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f114, jobid=5951923, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f120, jobid=5951924, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f066, jobid=5951925, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f072, jobid=5951926, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f078, jobid=5951927, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f084, jobid=5951928, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f090, jobid=5951929, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f096, jobid=5951930, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f102, jobid=5951931, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f108, jobid=5951988, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f114, jobid=5951989, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f120, jobid=5951990, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f054, jobid=5951991, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f060, jobid=5952011, in state QUEUED (PENDING) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ocean_prod_mem000_f072, jobid=5952012, in state SUCCEEDED (COMPLETED), ran for 14.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ocean_prod_mem000_f096, jobid=5952013, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ocean_prod_mem000_f120 -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ocean_prod_mem001_f072 -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ocean_prod_mem001_f096 -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ocean_prod_mem000_f120 succeeded, jobid=5952017 -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ocean_prod_mem001_f072 succeeded, jobid=5952018 -2025-07-28 22:34:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ocean_prod_mem001_f096 succeeded, jobid=5952019 -2025-07-28 22:40:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f120, jobid=5951906, in state QUEUED (PENDING) -2025-07-28 22:40:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f072, jobid=5951916, in state QUEUED (PENDING) -2025-07-28 22:40:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f078, jobid=5951917, in state QUEUED (PENDING) -2025-07-28 22:40:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f084, jobid=5951918, in state QUEUED (PENDING) -2025-07-28 22:40:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f090, jobid=5951919, in state QUEUED (PENDING) -2025-07-28 22:40:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f096, jobid=5951920, in state QUEUED (PENDING) -2025-07-28 22:40:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f102, jobid=5951921, in state QUEUED (PENDING) -2025-07-28 22:40:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f108, jobid=5951922, in state QUEUED (PENDING) -2025-07-28 22:40:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f114, jobid=5951923, in state QUEUED (PENDING) -2025-07-28 22:40:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f120, jobid=5951924, in state QUEUED (PENDING) -2025-07-28 22:40:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f066, jobid=5951925, in state QUEUED (PENDING) -2025-07-28 22:40:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f072, jobid=5951926, in state QUEUED (PENDING) -2025-07-28 22:40:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f078, jobid=5951927, in state QUEUED (PENDING) -2025-07-28 22:40:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f084, jobid=5951928, in state QUEUED (PENDING) -2025-07-28 22:40:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f090, jobid=5951929, in state QUEUED (PENDING) -2025-07-28 22:40:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f096, jobid=5951930, in state QUEUED (PENDING) -2025-07-28 22:40:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f102, jobid=5951931, in state QUEUED (PENDING) -2025-07-28 22:40:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f108, jobid=5951988, in state QUEUED (PENDING) -2025-07-28 22:40:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f114, jobid=5951989, in state QUEUED (PENDING) -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f120, jobid=5951990, in state QUEUED (PENDING) -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f054, jobid=5951991, in state QUEUED (PENDING) -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f060, jobid=5952011, in state QUEUED (PENDING) -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ocean_prod_mem000_f120, jobid=5952017, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ocean_prod_mem001_f072, jobid=5952018, in state SUCCEEDED (COMPLETED), ran for 15.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ocean_prod_mem001_f096, jobid=5952019, in state SUCCEEDED (COMPLETED), ran for 16.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ocean_prod_mem001_f120 -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ocean_prod_mem002_f072 -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ocean_prod_mem002_f096 -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ocean_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ocean_prod_mem001_f120 succeeded, jobid=5952029 -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ocean_prod_mem002_f072 succeeded, jobid=5952030 -2025-07-28 22:40:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ocean_prod_mem002_f096 succeeded, jobid=5952031 -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f120, jobid=5951906, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f072, jobid=5951916, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f078, jobid=5951917, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f084, jobid=5951918, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f090, jobid=5951919, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f096, jobid=5951920, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f102, jobid=5951921, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f108, jobid=5951922, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f114, jobid=5951923, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f120, jobid=5951924, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f066, jobid=5951925, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f072, jobid=5951926, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f078, jobid=5951927, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f084, jobid=5951928, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f090, jobid=5951929, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f096, jobid=5951930, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f102, jobid=5951931, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f108, jobid=5951988, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f114, jobid=5951989, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f120, jobid=5951990, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f054, jobid=5951991, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f060, jobid=5952011, in state QUEUED (PENDING) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ocean_prod_mem001_f120, jobid=5952029, in state SUCCEEDED (COMPLETED), ran for 15.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ocean_prod_mem002_f072, jobid=5952030, in state SUCCEEDED (COMPLETED), ran for 15.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ocean_prod_mem002_f096, jobid=5952031, in state SUCCEEDED (COMPLETED), ran for 15.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ocean_prod_mem002_f120 -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ice_prod_mem000_f072 -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ice_prod_mem000_f096 -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem000_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem001_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_ice_prod_mem002_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ocean_prod_mem002_f120 succeeded, jobid=5952042 -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ice_prod_mem000_f072 succeeded, jobid=5952043 -2025-07-28 22:46:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ice_prod_mem000_f096 succeeded, jobid=5952044 -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem000_f120, jobid=5951906, in state SUCCEEDED (COMPLETED), ran for 95.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f072, jobid=5951916, in state SUCCEEDED (COMPLETED), ran for 95.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f078, jobid=5951917, in state SUCCEEDED (COMPLETED), ran for 95.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f084, jobid=5951918, in state SUCCEEDED (COMPLETED), ran for 96.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f090, jobid=5951919, in state QUEUED (PENDING) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f096, jobid=5951920, in state QUEUED (PENDING) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f102, jobid=5951921, in state QUEUED (PENDING) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f108, jobid=5951922, in state QUEUED (PENDING) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f114, jobid=5951923, in state QUEUED (PENDING) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f120, jobid=5951924, in state QUEUED (PENDING) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f066, jobid=5951925, in state QUEUED (PENDING) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f072, jobid=5951926, in state QUEUED (PENDING) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f078, jobid=5951927, in state QUEUED (PENDING) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f084, jobid=5951928, in state QUEUED (PENDING) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f090, jobid=5951929, in state QUEUED (PENDING) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f096, jobid=5951930, in state QUEUED (PENDING) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f102, jobid=5951931, in state QUEUED (PENDING) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f108, jobid=5951988, in state QUEUED (PENDING) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f114, jobid=5951989, in state QUEUED (PENDING) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f120, jobid=5951990, in state QUEUED (PENDING) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f054, jobid=5951991, in state QUEUED (PENDING) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f060, jobid=5952011, in state QUEUED (PENDING) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ocean_prod_mem002_f120, jobid=5952042, in state SUCCEEDED (COMPLETED), ran for 15.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ice_prod_mem000_f072, jobid=5952043, in state SUCCEEDED (COMPLETED), ran for 15.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ice_prod_mem000_f096, jobid=5952044, in state SUCCEEDED (COMPLETED), ran for 15.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ice_prod_mem000_f120 -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ice_prod_mem001_f072 -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ice_prod_mem001_f096 -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ice_prod_mem001_f120 -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ice_prod_mem002_f072 -2025-07-28 22:52:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ice_prod_mem002_f096 -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_ice_prod_mem002_f120 -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem000_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ice_prod_mem000_f120 succeeded, jobid=5952052 -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ice_prod_mem001_f072 succeeded, jobid=5952053 -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ice_prod_mem001_f096 succeeded, jobid=5952054 -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ice_prod_mem001_f120 succeeded, jobid=5952055 -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ice_prod_mem002_f072 succeeded, jobid=5952056 -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ice_prod_mem002_f096 succeeded, jobid=5952057 -2025-07-28 22:52:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_ice_prod_mem002_f120 succeeded, jobid=5952058 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f090, jobid=5951919, in state SUCCEEDED (COMPLETED), ran for 90.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f096, jobid=5951920, in state SUCCEEDED (COMPLETED), ran for 94.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f102, jobid=5951921, in state SUCCEEDED (COMPLETED), ran for 93.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f108, jobid=5951922, in state SUCCEEDED (COMPLETED), ran for 93.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f114, jobid=5951923, in state SUCCEEDED (COMPLETED), ran for 94.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem001_f120, jobid=5951924, in state SUCCEEDED (COMPLETED), ran for 95.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f066, jobid=5951925, in state QUEUED (PENDING) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f072, jobid=5951926, in state QUEUED (PENDING) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f078, jobid=5951927, in state QUEUED (PENDING) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f084, jobid=5951928, in state QUEUED (PENDING) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f090, jobid=5951929, in state QUEUED (PENDING) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f096, jobid=5951930, in state QUEUED (PENDING) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f102, jobid=5951931, in state QUEUED (PENDING) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f108, jobid=5951988, in state QUEUED (PENDING) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f114, jobid=5951989, in state QUEUED (PENDING) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f120, jobid=5951990, in state QUEUED (PENDING) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f054, jobid=5951991, in state QUEUED (PENDING) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f060, jobid=5952011, in state QUEUED (PENDING) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ice_prod_mem000_f120, jobid=5952052, in state SUCCEEDED (COMPLETED), ran for 15.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ice_prod_mem001_f072, jobid=5952053, in state SUCCEEDED (COMPLETED), ran for 15.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ice_prod_mem001_f096, jobid=5952054, in state SUCCEEDED (COMPLETED), ran for 14.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ice_prod_mem001_f120, jobid=5952055, in state SUCCEEDED (COMPLETED), ran for 14.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ice_prod_mem002_f072, jobid=5952056, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ice_prod_mem002_f096, jobid=5952057, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_ice_prod_mem002_f120, jobid=5952058, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f074-f078 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f079-f083 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f084-f088 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f089-f093 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f094-f098 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f099-f103 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f104-f108 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f109-f112 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f113-f116 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem000_f117-f120 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f045-f048 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f049-f053 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f054-f058 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f074-f078 succeeded, jobid=5952064 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f079-f083 succeeded, jobid=5952065 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f084-f088 succeeded, jobid=5952066 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f089-f093 succeeded, jobid=5952067 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f094-f098 succeeded, jobid=5952068 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f099-f103 succeeded, jobid=5952069 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f104-f108 succeeded, jobid=5952070 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f109-f112 succeeded, jobid=5952071 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f113-f116 succeeded, jobid=5952072 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem000_f117-f120 succeeded, jobid=5952073 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f045-f048 succeeded, jobid=5952074 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f049-f053 succeeded, jobid=5952075 -2025-07-28 22:58:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f054-f058 succeeded, jobid=5952076 -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f066, jobid=5951925, in state SUCCEEDED (COMPLETED), ran for 90.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f072, jobid=5951926, in state SUCCEEDED (COMPLETED), ran for 93.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f078, jobid=5951927, in state SUCCEEDED (COMPLETED), ran for 93.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f084, jobid=5951928, in state SUCCEEDED (COMPLETED), ran for 94.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f090, jobid=5951929, in state SUCCEEDED (COMPLETED), ran for 93.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f096, jobid=5951930, in state SUCCEEDED (COMPLETED), ran for 93.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f102, jobid=5951931, in state SUCCEEDED (COMPLETED), ran for 94.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f108, jobid=5951988, in state QUEUED (PENDING) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f114, jobid=5951989, in state QUEUED (PENDING) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f120, jobid=5951990, in state QUEUED (PENDING) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f054, jobid=5951991, in state QUEUED (PENDING) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f060, jobid=5952011, in state QUEUED (PENDING) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f074-f078, jobid=5952064, in state QUEUED (PENDING) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f079-f083, jobid=5952065, in state QUEUED (PENDING) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f084-f088, jobid=5952066, in state QUEUED (PENDING) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f089-f093, jobid=5952067, in state QUEUED (PENDING) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f094-f098, jobid=5952068, in state QUEUED (PENDING) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f099-f103, jobid=5952069, in state QUEUED (PENDING) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f104-f108, jobid=5952070, in state QUEUED (PENDING) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f109-f112, jobid=5952071, in state QUEUED (PENDING) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f113-f116, jobid=5952072, in state QUEUED (PENDING) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f117-f120, jobid=5952073, in state QUEUED (PENDING) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f045-f048, jobid=5952074, in state QUEUED (PENDING) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f049-f053, jobid=5952075, in state QUEUED (PENDING) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f054-f058, jobid=5952076, in state QUEUED (PENDING) -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_ensstat_f066 -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_ensstat_f072 -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_ensstat_f078 -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_ensstat_f084 -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_ensstat_f090 -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_ensstat_f096 -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_ensstat_f102 -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_ensstat_f066 succeeded, jobid=5952089 -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_ensstat_f072 succeeded, jobid=5952090 -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_ensstat_f078 succeeded, jobid=5952091 -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_ensstat_f084 succeeded, jobid=5952092 -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_ensstat_f090 succeeded, jobid=5952093 -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_ensstat_f096 succeeded, jobid=5952094 -2025-07-28 23:04:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_ensstat_f102 succeeded, jobid=5952095 -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f108, jobid=5951988, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f114, jobid=5951989, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f120, jobid=5951990, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f054, jobid=5951991, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f060, jobid=5952011, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f066, jobid=5952089, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f072, jobid=5952090, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f078, jobid=5952091, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f084, jobid=5952092, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f090, jobid=5952093, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f096, jobid=5952094, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f102, jobid=5952095, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f074-f078, jobid=5952064, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f079-f083, jobid=5952065, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f084-f088, jobid=5952066, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f089-f093, jobid=5952067, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f094-f098, jobid=5952068, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f099-f103, jobid=5952069, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f104-f108, jobid=5952070, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f109-f112, jobid=5952071, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f113-f116, jobid=5952072, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f117-f120, jobid=5952073, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f045-f048, jobid=5952074, in state QUEUED (PENDING) -2025-07-28 23:10:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f049-f053, jobid=5952075, in state QUEUED (PENDING) -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f054-f058, jobid=5952076, in state QUEUED (PENDING) -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 23:10:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f108, jobid=5951988, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f114, jobid=5951989, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f120, jobid=5951990, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f054, jobid=5951991, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f060, jobid=5952011, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f066, jobid=5952089, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f072, jobid=5952090, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f078, jobid=5952091, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f084, jobid=5952092, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f090, jobid=5952093, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f096, jobid=5952094, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f102, jobid=5952095, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f074-f078, jobid=5952064, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f079-f083, jobid=5952065, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f084-f088, jobid=5952066, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f089-f093, jobid=5952067, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f094-f098, jobid=5952068, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f099-f103, jobid=5952069, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f104-f108, jobid=5952070, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f109-f112, jobid=5952071, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f113-f116, jobid=5952072, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f117-f120, jobid=5952073, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f045-f048, jobid=5952074, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f049-f053, jobid=5952075, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f054-f058, jobid=5952076, in state QUEUED (PENDING) -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f108, jobid=5951988, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f114, jobid=5951989, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f120, jobid=5951990, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f054, jobid=5951991, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f060, jobid=5952011, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f066, jobid=5952089, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f072, jobid=5952090, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f078, jobid=5952091, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f084, jobid=5952092, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f090, jobid=5952093, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f096, jobid=5952094, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f102, jobid=5952095, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f074-f078, jobid=5952064, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f079-f083, jobid=5952065, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f084-f088, jobid=5952066, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f089-f093, jobid=5952067, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f094-f098, jobid=5952068, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f099-f103, jobid=5952069, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f104-f108, jobid=5952070, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f109-f112, jobid=5952071, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f113-f116, jobid=5952072, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f117-f120, jobid=5952073, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f045-f048, jobid=5952074, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f049-f053, jobid=5952075, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f054-f058, jobid=5952076, in state QUEUED (PENDING) -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 23:23:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f108, jobid=5951988, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f114, jobid=5951989, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f120, jobid=5951990, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f054, jobid=5951991, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f060, jobid=5952011, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f066, jobid=5952089, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f072, jobid=5952090, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f078, jobid=5952091, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f084, jobid=5952092, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f090, jobid=5952093, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f096, jobid=5952094, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f102, jobid=5952095, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f074-f078, jobid=5952064, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f079-f083, jobid=5952065, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f084-f088, jobid=5952066, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f089-f093, jobid=5952067, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f094-f098, jobid=5952068, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f099-f103, jobid=5952069, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f104-f108, jobid=5952070, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f109-f112, jobid=5952071, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f113-f116, jobid=5952072, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f117-f120, jobid=5952073, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f045-f048, jobid=5952074, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f049-f053, jobid=5952075, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f054-f058, jobid=5952076, in state QUEUED (PENDING) -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 23:29:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f108, jobid=5951988, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f114, jobid=5951989, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f120, jobid=5951990, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f054, jobid=5951991, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f060, jobid=5952011, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f066, jobid=5952089, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f072, jobid=5952090, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f078, jobid=5952091, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f084, jobid=5952092, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f090, jobid=5952093, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f096, jobid=5952094, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f102, jobid=5952095, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f074-f078, jobid=5952064, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f079-f083, jobid=5952065, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f084-f088, jobid=5952066, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f089-f093, jobid=5952067, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f094-f098, jobid=5952068, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f099-f103, jobid=5952069, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f104-f108, jobid=5952070, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f109-f112, jobid=5952071, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f113-f116, jobid=5952072, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f117-f120, jobid=5952073, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f045-f048, jobid=5952074, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f049-f053, jobid=5952075, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f054-f058, jobid=5952076, in state QUEUED (PENDING) -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 23:35:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f108, jobid=5951988, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f114, jobid=5951989, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f120, jobid=5951990, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f054, jobid=5951991, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f060, jobid=5952011, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f066, jobid=5952089, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f072, jobid=5952090, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f078, jobid=5952091, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f084, jobid=5952092, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f090, jobid=5952093, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f096, jobid=5952094, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f102, jobid=5952095, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f074-f078, jobid=5952064, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f079-f083, jobid=5952065, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f084-f088, jobid=5952066, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f089-f093, jobid=5952067, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f094-f098, jobid=5952068, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f099-f103, jobid=5952069, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f104-f108, jobid=5952070, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f109-f112, jobid=5952071, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f113-f116, jobid=5952072, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f117-f120, jobid=5952073, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f045-f048, jobid=5952074, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f049-f053, jobid=5952075, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f054-f058, jobid=5952076, in state QUEUED (PENDING) -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f108, jobid=5951988, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f114, jobid=5951989, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f120, jobid=5951990, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f054, jobid=5951991, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f060, jobid=5952011, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f066, jobid=5952089, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f072, jobid=5952090, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f078, jobid=5952091, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f084, jobid=5952092, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f090, jobid=5952093, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f096, jobid=5952094, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f102, jobid=5952095, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f074-f078, jobid=5952064, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f079-f083, jobid=5952065, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f084-f088, jobid=5952066, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f089-f093, jobid=5952067, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f094-f098, jobid=5952068, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f099-f103, jobid=5952069, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f104-f108, jobid=5952070, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f109-f112, jobid=5952071, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f113-f116, jobid=5952072, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f117-f120, jobid=5952073, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f045-f048, jobid=5952074, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f049-f053, jobid=5952075, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f054-f058, jobid=5952076, in state QUEUED (PENDING) -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f108, jobid=5951988, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f114, jobid=5951989, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f120, jobid=5951990, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f054, jobid=5951991, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f060, jobid=5952011, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f066, jobid=5952089, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f072, jobid=5952090, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f078, jobid=5952091, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f084, jobid=5952092, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f090, jobid=5952093, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f096, jobid=5952094, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f102, jobid=5952095, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f074-f078, jobid=5952064, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f079-f083, jobid=5952065, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f084-f088, jobid=5952066, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f089-f093, jobid=5952067, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f094-f098, jobid=5952068, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f099-f103, jobid=5952069, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f104-f108, jobid=5952070, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f109-f112, jobid=5952071, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f113-f116, jobid=5952072, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f117-f120, jobid=5952073, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f045-f048, jobid=5952074, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f049-f053, jobid=5952075, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f054-f058, jobid=5952076, in state QUEUED (PENDING) -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f108, jobid=5951988, in state SUCCEEDED (COMPLETED), ran for 94.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f114, jobid=5951989, in state SUCCEEDED (COMPLETED), ran for 98.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_prod_mem002_f120, jobid=5951990, in state SUCCEEDED (COMPLETED), ran for 104.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f054, jobid=5951991, in state SUCCEEDED (COMPLETED), ran for 46.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f060, jobid=5952011, in state SUCCEEDED (COMPLETED), ran for 45.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f066, jobid=5952089, in state QUEUED (PENDING) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f072, jobid=5952090, in state QUEUED (PENDING) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f078, jobid=5952091, in state QUEUED (PENDING) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f084, jobid=5952092, in state QUEUED (PENDING) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f090, jobid=5952093, in state QUEUED (PENDING) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f096, jobid=5952094, in state QUEUED (PENDING) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f102, jobid=5952095, in state QUEUED (PENDING) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f074-f078, jobid=5952064, in state QUEUED (PENDING) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f079-f083, jobid=5952065, in state QUEUED (PENDING) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f084-f088, jobid=5952066, in state QUEUED (PENDING) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f089-f093, jobid=5952067, in state QUEUED (PENDING) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f094-f098, jobid=5952068, in state QUEUED (PENDING) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f099-f103, jobid=5952069, in state QUEUED (PENDING) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f104-f108, jobid=5952070, in state QUEUED (PENDING) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f109-f112, jobid=5952071, in state QUEUED (PENDING) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f113-f116, jobid=5952072, in state QUEUED (PENDING) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f117-f120, jobid=5952073, in state QUEUED (PENDING) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f045-f048, jobid=5952074, in state QUEUED (PENDING) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f049-f053, jobid=5952075, in state QUEUED (PENDING) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f054-f058, jobid=5952076, in state QUEUED (PENDING) -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_ensstat_f108 -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_ensstat_f114 -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_atmos_ensstat_f120 -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f059-f063 -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f064-f068 -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_ensstat_f108 succeeded, jobid=5952295 -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_ensstat_f114 succeeded, jobid=5952296 -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_atmos_ensstat_f120 succeeded, jobid=5952297 -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f059-f063 succeeded, jobid=5952298 -2025-07-28 23:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f064-f068 succeeded, jobid=5952299 -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f066, jobid=5952089, in state RUNNING (RUNNING) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f072, jobid=5952090, in state RUNNING (RUNNING) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f078, jobid=5952091, in state RUNNING (RUNNING) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f084, jobid=5952092, in state RUNNING (RUNNING) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f090, jobid=5952093, in state RUNNING (RUNNING) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f096, jobid=5952094, in state QUEUED (PENDING) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f102, jobid=5952095, in state QUEUED (PENDING) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f108, jobid=5952295, in state QUEUED (PENDING) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f114, jobid=5952296, in state QUEUED (PENDING) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f120, jobid=5952297, in state QUEUED (PENDING) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f074-f078, jobid=5952064, in state SUCCEEDED (COMPLETED), ran for 51.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f079-f083, jobid=5952065, in state SUCCEEDED (COMPLETED), ran for 52.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f084-f088, jobid=5952066, in state SUCCEEDED (COMPLETED), ran for 52.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f089-f093, jobid=5952067, in state SUCCEEDED (COMPLETED), ran for 53.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f094-f098, jobid=5952068, in state RUNNING (RUNNING) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f099-f103, jobid=5952069, in state RUNNING (RUNNING) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f104-f108, jobid=5952070, in state RUNNING (RUNNING) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f109-f112, jobid=5952071, in state RUNNING (RUNNING) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f113-f116, jobid=5952072, in state RUNNING (RUNNING) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f117-f120, jobid=5952073, in state RUNNING (RUNNING) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f045-f048, jobid=5952074, in state RUNNING (RUNNING) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f049-f053, jobid=5952075, in state RUNNING (RUNNING) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f054-f058, jobid=5952076, in state RUNNING (RUNNING) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f059-f063, jobid=5952298, in state QUEUED (PENDING) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f064-f068, jobid=5952299, in state QUEUED (PENDING) -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f069-f073 -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f074-f078 -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f079-f083 -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f084-f088 -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem001_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f015-f019, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f020-f024, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f025-f029, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f030-f034, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f035-f039, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f040-f044, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f045-f048, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f069-f073 succeeded, jobid=5952344 -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f074-f078 succeeded, jobid=5952345 -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f079-f083 succeeded, jobid=5952346 -2025-07-29 00:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f084-f088 succeeded, jobid=5952347 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f066, jobid=5952089, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f072, jobid=5952090, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f078, jobid=5952091, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f084, jobid=5952092, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f090, jobid=5952093, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f096, jobid=5952094, in state QUEUED (PENDING) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f102, jobid=5952095, in state QUEUED (PENDING) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f108, jobid=5952295, in state QUEUED (PENDING) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f114, jobid=5952296, in state QUEUED (PENDING) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f120, jobid=5952297, in state QUEUED (PENDING) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f094-f098, jobid=5952068, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f099-f103, jobid=5952069, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f104-f108, jobid=5952070, in state SUCCEEDED (COMPLETED), ran for 51.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f109-f112, jobid=5952071, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f113-f116, jobid=5952072, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem000_f117-f120, jobid=5952073, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f045-f048, jobid=5952074, in state SUCCEEDED (COMPLETED), ran for 42.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f049-f053, jobid=5952075, in state SUCCEEDED (COMPLETED), ran for 51.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f054-f058, jobid=5952076, in state SUCCEEDED (COMPLETED), ran for 54.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f059-f063, jobid=5952298, in state QUEUED (PENDING) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f064-f068, jobid=5952299, in state QUEUED (PENDING) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f069-f073, jobid=5952344, in state QUEUED (PENDING) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f074-f078, jobid=5952345, in state QUEUED (PENDING) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f079-f083, jobid=5952346, in state QUEUED (PENDING) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f084-f088, jobid=5952347, in state QUEUED (PENDING) -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f089-f093 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f094-f098 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f099-f103 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f104-f108 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f109-f112 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f113-f116 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem001_f117-f120 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f015-f019 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f020-f024 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f025-f029 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f030-f034 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f035-f039 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f040-f044 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f045-f048 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f089-f093 succeeded, jobid=5952393 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f094-f098 succeeded, jobid=5952394 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f099-f103 succeeded, jobid=5952395 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f104-f108 succeeded, jobid=5952396 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f109-f112 succeeded, jobid=5952397 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f113-f116 succeeded, jobid=5952398 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem001_f117-f120 succeeded, jobid=5952399 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f015-f019 succeeded, jobid=5952400 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f020-f024 succeeded, jobid=5952401 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f025-f029 succeeded, jobid=5952402 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f030-f034 succeeded, jobid=5952403 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f035-f039 succeeded, jobid=5952404 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f040-f044 succeeded, jobid=5952405 -2025-07-29 00:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f045-f048 succeeded, jobid=5952406 -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f096, jobid=5952094, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f102, jobid=5952095, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f108, jobid=5952295, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f114, jobid=5952296, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f120, jobid=5952297, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f059-f063, jobid=5952298, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f064-f068, jobid=5952299, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f069-f073, jobid=5952344, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f074-f078, jobid=5952345, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f079-f083, jobid=5952346, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f084-f088, jobid=5952347, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f089-f093, jobid=5952393, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f094-f098, jobid=5952394, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f099-f103, jobid=5952395, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f104-f108, jobid=5952396, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f109-f112, jobid=5952397, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f113-f116, jobid=5952398, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f117-f120, jobid=5952399, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f015-f019, jobid=5952400, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f020-f024, jobid=5952401, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f025-f029, jobid=5952402, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f030-f034, jobid=5952403, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f035-f039, jobid=5952404, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f040-f044, jobid=5952405, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f045-f048, jobid=5952406, in state QUEUED (PENDING) -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f049-f053, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f054-f058, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f096, jobid=5952094, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f102, jobid=5952095, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f108, jobid=5952295, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f114, jobid=5952296, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f120, jobid=5952297, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f059-f063, jobid=5952298, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f064-f068, jobid=5952299, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f069-f073, jobid=5952344, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f074-f078, jobid=5952345, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f079-f083, jobid=5952346, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f084-f088, jobid=5952347, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f089-f093, jobid=5952393, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f094-f098, jobid=5952394, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f099-f103, jobid=5952395, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f104-f108, jobid=5952396, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f109-f112, jobid=5952397, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f113-f116, jobid=5952398, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f117-f120, jobid=5952399, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f015-f019, jobid=5952400, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f020-f024, jobid=5952401, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f025-f029, jobid=5952402, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f030-f034, jobid=5952403, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f035-f039, jobid=5952404, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f040-f044, jobid=5952405, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f045-f048, jobid=5952406, in state QUEUED (PENDING) -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f049-f053 -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f054-f058 -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f059-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f064-f068, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f069-f073, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f074-f078, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f079-f083, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f084-f088, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f089-f093, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f094-f098, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f099-f103, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f104-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f109-f112, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f113-f116, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_grid_mem002_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem000, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem001, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gefs_wave_post_pnt_mem002, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f049-f053 succeeded, jobid=5952445 -2025-07-29 00:23:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f054-f058 succeeded, jobid=5952446 -2025-07-29 00:29:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f108, jobid=5952295, in state QUEUED (PENDING) -2025-07-29 00:29:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f114, jobid=5952296, in state QUEUED (PENDING) -2025-07-29 00:29:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f120, jobid=5952297, in state QUEUED (PENDING) -2025-07-29 00:29:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f059-f063, jobid=5952298, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f064-f068, jobid=5952299, in state SUCCEEDED (COMPLETED), ran for 52.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f069-f073, jobid=5952344, in state SUCCEEDED (COMPLETED), ran for 51.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f074-f078, jobid=5952345, in state SUCCEEDED (COMPLETED), ran for 52.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f079-f083, jobid=5952346, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f084-f088, jobid=5952347, in state SUCCEEDED (COMPLETED), ran for 51.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f089-f093, jobid=5952393, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f094-f098, jobid=5952394, in state SUCCEEDED (COMPLETED), ran for 54.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f099-f103, jobid=5952395, in state SUCCEEDED (COMPLETED), ran for 51.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f104-f108, jobid=5952396, in state SUCCEEDED (COMPLETED), ran for 51.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f109-f112, jobid=5952397, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f113-f116, jobid=5952398, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem001_f117-f120, jobid=5952399, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f015-f019, jobid=5952400, in state SUCCEEDED (COMPLETED), ran for 52.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f020-f024, jobid=5952401, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f025-f029, jobid=5952402, in state SUCCEEDED (COMPLETED), ran for 53.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f030-f034, jobid=5952403, in state SUCCEEDED (COMPLETED), ran for 51.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f035-f039, jobid=5952404, in state SUCCEEDED (COMPLETED), ran for 51.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f040-f044, jobid=5952405, in state SUCCEEDED (COMPLETED), ran for 51.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f045-f048, jobid=5952406, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f049-f053, jobid=5952445, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f054-f058, jobid=5952446, in state QUEUED (PENDING) -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f059-f063 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f064-f068 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f069-f073 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f074-f078 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f079-f083 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f084-f088 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f089-f093 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f094-f098 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f099-f103 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f104-f108 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f109-f112 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f113-f116 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_grid_mem002_f117-f120 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_pnt_mem000 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_pnt_mem001 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_wave_post_pnt_mem002 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f059-f063 succeeded, jobid=5952485 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f064-f068 succeeded, jobid=5952486 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f069-f073 succeeded, jobid=5952487 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f074-f078 succeeded, jobid=5952488 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f079-f083 succeeded, jobid=5952489 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f084-f088 succeeded, jobid=5952490 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f089-f093 succeeded, jobid=5952491 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f094-f098 succeeded, jobid=5952492 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f099-f103 succeeded, jobid=5952493 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f104-f108 succeeded, jobid=5952494 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f109-f112 succeeded, jobid=5952495 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f113-f116 succeeded, jobid=5952496 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_grid_mem002_f117-f120 succeeded, jobid=5952497 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_pnt_mem000 succeeded, jobid=5952498 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_pnt_mem001 succeeded, jobid=5952499 -2025-07-29 00:29:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gefs_wave_post_pnt_mem002 succeeded, jobid=5952500 -2025-07-29 00:35:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f108, jobid=5952295, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:35:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f114, jobid=5952296, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:35:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_atmos_ensstat_f120, jobid=5952297, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:35:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f054-f058, jobid=5952446, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:35:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f059-f063, jobid=5952485, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:35:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f064-f068, jobid=5952486, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:35:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f069-f073, jobid=5952487, in state SUCCEEDED (COMPLETED), ran for 48.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:35:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f074-f078, jobid=5952488, in state QUEUED (PENDING) -2025-07-29 00:35:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f079-f083, jobid=5952489, in state QUEUED (PENDING) -2025-07-29 00:35:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f084-f088, jobid=5952490, in state QUEUED (PENDING) -2025-07-29 00:35:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f089-f093, jobid=5952491, in state QUEUED (PENDING) -2025-07-29 00:35:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f094-f098, jobid=5952492, in state QUEUED (PENDING) -2025-07-29 00:35:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f099-f103, jobid=5952493, in state QUEUED (PENDING) -2025-07-29 00:35:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f104-f108, jobid=5952494, in state QUEUED (PENDING) -2025-07-29 00:35:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f109-f112, jobid=5952495, in state QUEUED (PENDING) -2025-07-29 00:35:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f113-f116, jobid=5952496, in state QUEUED (PENDING) -2025-07-29 00:35:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f117-f120, jobid=5952497, in state QUEUED (PENDING) -2025-07-29 00:35:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_pnt_mem000, jobid=5952498, in state SUCCEEDED (COMPLETED), ran for 58.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:35:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_pnt_mem001, jobid=5952499, in state SUCCEEDED (COMPLETED), ran for 58.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:35:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_pnt_mem002, jobid=5952500, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f074-f078, jobid=5952488, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f079-f083, jobid=5952489, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f084-f088, jobid=5952490, in state SUCCEEDED (COMPLETED), ran for 48.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f089-f093, jobid=5952491, in state SUCCEEDED (COMPLETED), ran for 51.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f094-f098, jobid=5952492, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f099-f103, jobid=5952493, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f104-f108, jobid=5952494, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f109-f112, jobid=5952495, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f113-f116, jobid=5952496, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_wave_post_grid_mem002_f117-f120, jobid=5952497, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_arch_vrfy -2025-07-29 00:41:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gefs_arch_vrfy is pending at druby://130.18.14.151:44883 -2025-07-29 00:47:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gefs_arch_vrfy is success, jobid=5952594 -2025-07-29 00:47:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_arch_vrfy, jobid=5952594, in state SUCCEEDED (COMPLETED), ran for 14.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:47:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gefs_cleanup -2025-07-29 00:47:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gefs_cleanup is pending at druby://130.18.14.151:33085 -2025-07-29 00:53:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gefs_cleanup is success, jobid=5952641 -2025-07-29 00:53:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gefs_cleanup, jobid=5952641, in state SUCCEEDED (COMPLETED), ran for 16.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:53:52 -0500 :: hercules-login-1.hpc.msstate.edu :: This cycle is complete: Success diff --git a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C48_S2SW_logs_2021032312.log b/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C48_S2SW_logs_2021032312.log deleted file mode 100644 index 2b7201ec..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C48_S2SW_logs_2021032312.log +++ /dev/null @@ -1,2417 +0,0 @@ -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_stage_ic -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_waveinit -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_stage_ic succeeded, jobid=5950267 -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_waveinit succeeded, jobid=5950268 -2025-07-28 14:48:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_stage_ic, jobid=5950267, in state RUNNING (RUNNING) -2025-07-28 14:48:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_waveinit, jobid=5950268, in state QUEUED (PENDING) -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_stage_ic, jobid=5950267, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_waveinit, jobid=5950268, in state SUCCEEDED (COMPLETED), ran for 35.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_fcst_seg0 -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_fcst_seg0 succeeded, jobid=5950292 -2025-07-28 15:00:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 15:06:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 15:18:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 15:30:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 15:36:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 15:43:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 15:49:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 15:55:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 16:01:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 16:07:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 16:13:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 16:19:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 16:25:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 16:31:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 16:37:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 16:43:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 16:49:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 16:55:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 17:01:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 17:07:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 17:13:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 17:19:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 17:25:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 17:31:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 17:37:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 17:43:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 17:49:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 17:56:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 18:02:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 18:08:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 18:14:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 18:20:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 18:26:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 18:32:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 18:38:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 18:44:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 18:50:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 18:56:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state RUNNING (RUNNING) -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state RUNNING (RUNNING) -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f000-f002 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f003-f005 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f006-f008 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f009-f011 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f012-f014 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f015-f017 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f018-f020 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f006 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f006 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f012 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f018 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f000-f002 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f003-f005 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f006-f008 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f009-f011 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f012-f014 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f015-f017 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f018-f020 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f021-f023 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f024-f026 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f027-f029 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f030-f032 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f000-f002 succeeded, jobid=5950995 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f003-f005 succeeded, jobid=5950996 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f006-f008 succeeded, jobid=5950997 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f009-f011 succeeded, jobid=5950998 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f012-f014 succeeded, jobid=5950999 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f015-f017 succeeded, jobid=5951000 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f018-f020 succeeded, jobid=5951001 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f006 succeeded, jobid=5951002 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f006 succeeded, jobid=5951003 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f012 succeeded, jobid=5951004 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f018 succeeded, jobid=5951005 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f000-f002 succeeded, jobid=5951006 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f003-f005 succeeded, jobid=5951007 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f006-f008 succeeded, jobid=5951008 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f009-f011 succeeded, jobid=5951009 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f012-f014 succeeded, jobid=5951010 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f015-f017 succeeded, jobid=5951011 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f018-f020 succeeded, jobid=5951012 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f021-f023 succeeded, jobid=5951013 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f024-f026 succeeded, jobid=5951014 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f027-f029 succeeded, jobid=5951015 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f030-f032 succeeded, jobid=5951016 -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state RUNNING (RUNNING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f006, jobid=5951002, in state SUCCEEDED (COMPLETED), ran for 46.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f006, jobid=5951003, in state SUCCEEDED (COMPLETED), ran for 46.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f012, jobid=5951004, in state SUCCEEDED (COMPLETED), ran for 46.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f018, jobid=5951005, in state SUCCEEDED (COMPLETED), ran for 46.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f021-f023 -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f024-f026 -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f027-f029 -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f030-f032 -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f033-f035 -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f036-f038 -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f021-f023 succeeded, jobid=5951060 -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f024-f026 succeeded, jobid=5951061 -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f027-f029 succeeded, jobid=5951062 -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f030-f032 succeeded, jobid=5951063 -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f033-f035 succeeded, jobid=5951064 -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f036-f038 succeeded, jobid=5951065 -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state RUNNING (RUNNING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state SUCCEEDED (COMPLETED), ran for 1488.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f039-f041 -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f039-f041 succeeded, jobid=5951115 -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state RUNNING (RUNNING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state SUCCEEDED (COMPLETED), ran for 267.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f042-f044 -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f042-f044 succeeded, jobid=5951482 -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5951482, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5951482, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state SUCCEEDED (COMPLETED), ran for 284.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state SUCCEEDED (COMPLETED), ran for 280.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5951482, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f045-f047 -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f048-f054 -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f045-f047 succeeded, jobid=5951544 -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f048-f054 succeeded, jobid=5951545 -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state RUNNING (RUNNING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5951482, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5951544, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5951545, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state SUCCEEDED (COMPLETED), ran for 274.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state RUNNING (RUNNING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state RUNNING (RUNNING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state RUNNING (RUNNING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5951482, in state QUEUED (PENDING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5951544, in state QUEUED (PENDING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5951545, in state QUEUED (PENDING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state SUCCEEDED (COMPLETED), ran for 70.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state SUCCEEDED (COMPLETED), ran for 76.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state SUCCEEDED (COMPLETED), ran for 75.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state SUCCEEDED (COMPLETED), ran for 74.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state SUCCEEDED (COMPLETED), ran for 74.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state SUCCEEDED (COMPLETED), ran for 75.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state SUCCEEDED (COMPLETED), ran for 75.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state SUCCEEDED (COMPLETED), ran for 74.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f057-f063 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f066-f072 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f075-f081 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f084-f090 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f093-f099 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f102-f108 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f111-f114 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f117-f120 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f012 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f018 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f024 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f030 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f057-f063 succeeded, jobid=5951616 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f066-f072 succeeded, jobid=5951617 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f075-f081 succeeded, jobid=5951618 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f084-f090 succeeded, jobid=5951619 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f093-f099 succeeded, jobid=5951620 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f102-f108 succeeded, jobid=5951621 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f111-f114 succeeded, jobid=5951622 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f117-f120 succeeded, jobid=5951623 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f012 succeeded, jobid=5951624 -2025-07-28 21:21:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f018 succeeded, jobid=5951625 -2025-07-28 21:21:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f024 succeeded, jobid=5951626 -2025-07-28 21:21:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f030 succeeded, jobid=5951627 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state SUCCEEDED (COMPLETED), ran for 273.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state SUCCEEDED (COMPLETED), ran for 272.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state SUCCEEDED (COMPLETED), ran for 268.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5951482, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5951544, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5951545, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5951616, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5951617, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5951618, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5951619, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5951620, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5951621, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5951622, in state SUCCEEDED (COMPLETED), ran for 216.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5951623, in state RUNNING (COMPLETING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f012, jobid=5951624, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f018, jobid=5951625, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f024, jobid=5951626, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f030, jobid=5951627, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f036 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f042 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f048 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f054 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f060 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f066 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f072 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f078 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f036 succeeded, jobid=5951670 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f042 succeeded, jobid=5951671 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f048 succeeded, jobid=5951672 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f054 succeeded, jobid=5951673 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f060 succeeded, jobid=5951674 -2025-07-28 21:27:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f066 succeeded, jobid=5951675 -2025-07-28 21:27:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f072 succeeded, jobid=5951676 -2025-07-28 21:27:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f078 succeeded, jobid=5951677 -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state SUCCEEDED (COMPLETED), ran for 276.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state SUCCEEDED (COMPLETED), ran for 276.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state SUCCEEDED (COMPLETED), ran for 273.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state SUCCEEDED (COMPLETED), ran for 280.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state SUCCEEDED (COMPLETED), ran for 281.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state SUCCEEDED (COMPLETED), ran for 274.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state SUCCEEDED (COMPLETED), ran for 279.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5951482, in state SUCCEEDED (COMPLETED), ran for 275.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5951544, in state SUCCEEDED (COMPLETED), ran for 275.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5951545, in state SUCCEEDED (COMPLETED), ran for 321.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5951616, in state SUCCEEDED (COMPLETED), ran for 286.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5951617, in state SUCCEEDED (COMPLETED), ran for 320.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5951618, in state SUCCEEDED (COMPLETED), ran for 318.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5951619, in state SUCCEEDED (COMPLETED), ran for 324.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5951620, in state SUCCEEDED (COMPLETED), ran for 325.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5951621, in state SUCCEEDED (COMPLETED), ran for 330.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5951623, in state SUCCEEDED (COMPLETED), ran for 220.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f036, jobid=5951670, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f042, jobid=5951671, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f048, jobid=5951672, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f054, jobid=5951673, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f060, jobid=5951674, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f066, jobid=5951675, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f072, jobid=5951676, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f078, jobid=5951677, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_tracker -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_genesis -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f084 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f090 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f096 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f102 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f108 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f114 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f120 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f024 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f030 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f036 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f042 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f048 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f054 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f060 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f066 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f072 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f078 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f084 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f090 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f096 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f102 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f108 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f114 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_tracker succeeded, jobid=5951717 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_genesis succeeded, jobid=5951718 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f084 succeeded, jobid=5951719 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f090 succeeded, jobid=5951720 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f096 succeeded, jobid=5951721 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f102 succeeded, jobid=5951722 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f108 succeeded, jobid=5951723 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f114 succeeded, jobid=5951724 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f120 succeeded, jobid=5951725 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f024 succeeded, jobid=5951726 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f030 succeeded, jobid=5951727 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f036 succeeded, jobid=5951728 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f042 succeeded, jobid=5951729 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f048 succeeded, jobid=5951730 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f054 succeeded, jobid=5951731 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f060 succeeded, jobid=5951732 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f066 succeeded, jobid=5951733 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f072 succeeded, jobid=5951734 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f078 succeeded, jobid=5951735 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f084 succeeded, jobid=5951736 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f090 succeeded, jobid=5951737 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f096 succeeded, jobid=5951738 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f102 succeeded, jobid=5951739 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f108 succeeded, jobid=5951740 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f114 succeeded, jobid=5951741 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_tracker, jobid=5951717, in state SUCCEEDED (COMPLETED), ran for 64.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_genesis, jobid=5951718, in state RUNNING (RUNNING) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f084, jobid=5951719, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f090, jobid=5951720, in state SUCCEEDED (COMPLETED), ran for 20.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f096, jobid=5951721, in state SUCCEEDED (COMPLETED), ran for 20.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f102, jobid=5951722, in state SUCCEEDED (COMPLETED), ran for 20.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f108, jobid=5951723, in state SUCCEEDED (COMPLETED), ran for 20.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f114, jobid=5951724, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f120, jobid=5951725, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f024, jobid=5951726, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f030, jobid=5951727, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f036, jobid=5951728, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f042, jobid=5951729, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f048, jobid=5951730, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f054, jobid=5951731, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f060, jobid=5951732, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f066, jobid=5951733, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f072, jobid=5951734, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f078, jobid=5951735, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f084, jobid=5951736, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f090, jobid=5951737, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f096, jobid=5951738, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f102, jobid=5951739, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f108, jobid=5951740, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f114, jobid=5951741, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f120 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f033-f035 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f036-f038 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f039-f041 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f042-f044 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f045-f047 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f048-f054 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f057-f063 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f066-f072 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f075-f081 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f084-f090 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f093-f099 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f102-f108 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f111-f114 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f117-f120 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostbndpnt -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostbndpntbll -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostpnt -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f120 succeeded, jobid=5951787 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f033-f035 succeeded, jobid=5951788 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f036-f038 succeeded, jobid=5951789 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f039-f041 succeeded, jobid=5951790 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f042-f044 succeeded, jobid=5951791 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f045-f047 succeeded, jobid=5951792 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f048-f054 succeeded, jobid=5951793 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f057-f063 succeeded, jobid=5951794 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f066-f072 succeeded, jobid=5951795 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f075-f081 succeeded, jobid=5951796 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f084-f090 succeeded, jobid=5951797 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f093-f099 succeeded, jobid=5951798 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f102-f108 succeeded, jobid=5951799 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f111-f114 succeeded, jobid=5951800 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f117-f120 succeeded, jobid=5951801 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostbndpnt succeeded, jobid=5951802 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostbndpntbll succeeded, jobid=5951803 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostpnt succeeded, jobid=5951804 -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_genesis, jobid=5951718, in state SUCCEEDED (COMPLETED), ran for 465.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f120, jobid=5951787, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f033-f035, jobid=5951788, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f036-f038, jobid=5951789, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f039-f041, jobid=5951790, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f042-f044, jobid=5951791, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f045-f047, jobid=5951792, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f048-f054, jobid=5951793, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f057-f063, jobid=5951794, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f066-f072, jobid=5951795, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f075-f081, jobid=5951796, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f084-f090, jobid=5951797, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f093-f099, jobid=5951798, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f102-f108, jobid=5951799, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f111-f114, jobid=5951800, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f117-f120, jobid=5951801, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostbndpnt, jobid=5951802, in state RUNNING (RUNNING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostbndpntbll, jobid=5951803, in state SUCCEEDED (COMPLETED), ran for 78.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostpnt, jobid=5951804, in state RUNNING (RUNNING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f033-f035, jobid=5951788, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f036-f038, jobid=5951789, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f039-f041, jobid=5951790, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f042-f044, jobid=5951791, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f045-f047, jobid=5951792, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f048-f054, jobid=5951793, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f057-f063, jobid=5951794, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f066-f072, jobid=5951795, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f075-f081, jobid=5951796, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f084-f090, jobid=5951797, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f093-f099, jobid=5951798, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f102-f108, jobid=5951799, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f111-f114, jobid=5951800, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f117-f120, jobid=5951801, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostbndpnt, jobid=5951802, in state SUCCEEDED (COMPLETED), ran for 175.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostpnt, jobid=5951804, in state SUCCEEDED (COMPLETED), ran for 111.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f033-f035, jobid=5951788, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f036-f038, jobid=5951789, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f039-f041, jobid=5951790, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f042-f044, jobid=5951791, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f045-f047, jobid=5951792, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f048-f054, jobid=5951793, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f057-f063, jobid=5951794, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f066-f072, jobid=5951795, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f075-f081, jobid=5951796, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f084-f090, jobid=5951797, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f093-f099, jobid=5951798, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f102-f108, jobid=5951799, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f111-f114, jobid=5951800, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f117-f120, jobid=5951801, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f033-f035, jobid=5951788, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f036-f038, jobid=5951789, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f039-f041, jobid=5951790, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f042-f044, jobid=5951791, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f045-f047, jobid=5951792, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f048-f054, jobid=5951793, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f057-f063, jobid=5951794, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f066-f072, jobid=5951795, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f075-f081, jobid=5951796, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f084-f090, jobid=5951797, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f093-f099, jobid=5951798, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f102-f108, jobid=5951799, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f111-f114, jobid=5951800, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f117-f120, jobid=5951801, in state QUEUED (PENDING) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f033-f035, jobid=5951788, in state SUCCEEDED (COMPLETED), ran for 70.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f036-f038, jobid=5951789, in state SUCCEEDED (COMPLETED), ran for 70.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f039-f041, jobid=5951790, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f042-f044, jobid=5951791, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f045-f047, jobid=5951792, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f048-f054, jobid=5951793, in state SUCCEEDED (COMPLETED), ran for 74.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f057-f063, jobid=5951794, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f066-f072, jobid=5951795, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f075-f081, jobid=5951796, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f084-f090, jobid=5951797, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f093-f099, jobid=5951798, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f102-f108, jobid=5951799, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f111-f114, jobid=5951800, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f117-f120, jobid=5951801, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_arch_vrfy -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_arch_vrfy is pending at druby://130.18.14.151:37481 -2025-07-28 22:16:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_arch_vrfy is success, jobid=5951962 -2025-07-28 22:16:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_arch_vrfy, jobid=5951962, in state SUCCEEDED (COMPLETED), ran for 74.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:16:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_cleanup -2025-07-28 22:16:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_cleanup is pending at druby://130.18.14.151:33107 -2025-07-28 22:22:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_cleanup is success, jobid=5951987 -2025-07-28 22:22:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_cleanup, jobid=5951987, in state SUCCEEDED (COMPLETED), ran for 22.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:22:23 -0500 :: hercules-login-1.hpc.msstate.edu :: This cycle is complete: Success diff --git a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C48mx500_3DVarAOWCDA_logs_2021032418.log b/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C48mx500_3DVarAOWCDA_logs_2021032418.log deleted file mode 100644 index 6c7781ad..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C48mx500_3DVarAOWCDA_logs_2021032418.log +++ /dev/null @@ -1,148 +0,0 @@ -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_waveinit -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_stage_ic -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_waveinit succeeded, jobid=5950274 -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_stage_ic succeeded, jobid=5950275 -2025-07-28 14:48:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_waveinit, jobid=5950274, in state QUEUED (PENDING) -2025-07-28 14:48:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_stage_ic, jobid=5950275, in state RUNNING (RUNNING) -2025-07-28 14:54:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_waveinit, jobid=5950274, in state QUEUED (PENDING) -2025-07-28 14:54:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_stage_ic, jobid=5950275, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 14:54:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_fcst_seg0 -2025-07-28 14:54:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_fcst_seg0 succeeded, jobid=5950293 -2025-07-28 15:00:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_waveinit, jobid=5950274, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:00:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5950293, in state QUEUED (PENDING) -2025-07-28 15:06:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5950293, in state RUNNING (RUNNING) -2025-07-28 15:12:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5950293, in state SUCCEEDED (COMPLETED), ran for 266.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:12:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f000 -2025-07-28 15:12:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f003 -2025-07-28 15:12:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f006 -2025-07-28 15:12:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f009 -2025-07-28 15:12:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f000 succeeded, jobid=5950383 -2025-07-28 15:12:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f003 succeeded, jobid=5950384 -2025-07-28 15:12:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f006 succeeded, jobid=5950385 -2025-07-28 15:12:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f009 succeeded, jobid=5950386 -2025-07-28 15:18:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 15:18:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 15:18:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 15:18:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 15:24:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 15:24:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 15:24:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 15:24:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 15:30:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 15:30:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 15:30:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 15:30:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 15:36:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 15:36:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 15:36:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 15:36:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 15:43:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 15:43:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 15:43:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 15:43:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 15:49:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 15:49:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 15:49:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 15:49:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 15:55:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 15:55:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 15:55:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 15:55:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 16:01:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 16:01:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 16:01:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 16:01:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 16:07:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 16:07:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 16:07:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 16:07:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 16:13:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 16:13:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 16:13:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 16:13:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 16:19:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 16:19:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 16:19:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 16:19:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 16:25:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 16:25:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 16:25:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 16:25:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 16:31:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 16:31:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 16:31:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 16:31:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 16:37:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 16:37:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 16:37:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 16:37:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 16:43:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 16:43:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 16:43:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 16:43:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 16:49:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 16:49:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 16:49:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 16:49:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 16:55:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 16:55:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 16:55:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 16:55:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 17:01:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 17:01:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 17:01:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 17:01:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 17:07:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 17:07:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 17:07:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 17:07:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 17:13:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 17:13:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 17:13:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 17:13:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 17:19:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 17:19:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 17:19:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 17:19:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 17:25:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 17:25:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 17:25:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 17:25:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 17:31:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 17:31:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 17:31:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 17:31:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 17:37:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 17:37:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 17:37:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 17:37:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 17:44:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 17:44:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 17:44:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 17:44:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 17:50:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state QUEUED (PENDING) -2025-07-28 17:50:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 17:50:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 17:50:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 17:56:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950383, in state SUCCEEDED (COMPLETED), ran for 65.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 17:56:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 17:56:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 17:56:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 18:02:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 18:02:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 18:02:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 18:08:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state QUEUED (PENDING) -2025-07-28 18:08:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 18:08:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 18:14:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950384, in state SUCCEEDED (COMPLETED), ran for 66.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 18:14:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 18:14:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 18:20:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 18:20:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 18:26:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 18:26:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 18:32:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state QUEUED (PENDING) -2025-07-28 18:32:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state QUEUED (PENDING) -2025-07-28 18:38:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950385, in state SUCCEEDED (COMPLETED), ran for 66.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 18:38:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950386, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 18:38:29 -0500 :: hercules-login-1.hpc.msstate.edu :: This cycle is complete: Success diff --git a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C48mx500_3DVarAOWCDA_logs_2021032500.log b/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C48mx500_3DVarAOWCDA_logs_2021032500.log deleted file mode 100644 index 9b60f5ed..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C48mx500_3DVarAOWCDA_logs_2021032500.log +++ /dev/null @@ -1,1447 +0,0 @@ -2025-07-28 15:06:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_prepoceanobs -2025-07-28 15:06:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_marinebmatinit -2025-07-28 15:06:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_prepoceanobs -2025-07-28 15:06:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marinebmatinit -2025-07-28 15:06:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_prepoceanobs succeeded, jobid=5950334 -2025-07-28 15:06:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_marinebmatinit succeeded, jobid=5950335 -2025-07-28 15:06:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_prepoceanobs succeeded, jobid=5950336 -2025-07-28 15:06:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_marinebmatinit succeeded, jobid=5950337 -2025-07-28 15:12:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prepoceanobs, jobid=5950334, in state QUEUED (PENDING) -2025-07-28 15:12:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmatinit, jobid=5950335, in state QUEUED (PENDING) -2025-07-28 15:12:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5950336, in state QUEUED (PENDING) -2025-07-28 15:12:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950337, in state QUEUED (PENDING) -2025-07-28 15:18:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prepoceanobs, jobid=5950334, in state QUEUED (PENDING) -2025-07-28 15:18:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmatinit, jobid=5950335, in state QUEUED (PENDING) -2025-07-28 15:18:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5950336, in state QUEUED (PENDING) -2025-07-28 15:18:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950337, in state QUEUED (PENDING) -2025-07-28 15:24:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prepoceanobs, jobid=5950334, in state RUNNING (RUNNING) -2025-07-28 15:24:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmatinit, jobid=5950335, in state SUCCEEDED (COMPLETED), ran for 32.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:24:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5950336, in state QUEUED (PENDING) -2025-07-28 15:24:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950337, in state QUEUED (PENDING) -2025-07-28 15:24:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_marinebmat -2025-07-28 15:24:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_marinebmat succeeded, jobid=5950446 -2025-07-28 15:30:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prepoceanobs, jobid=5950334, in state SUCCEEDED (COMPLETED), ran for 214.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:30:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 15:30:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5950336, in state RUNNING (RUNNING) -2025-07-28 15:30:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950337, in state SUCCEEDED (COMPLETED), ran for 32.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:30:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marinebmat -2025-07-28 15:30:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_marinebmat is pending at druby://130.18.14.151:35605 -2025-07-28 15:36:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_marinebmat is success, jobid=5950494 -2025-07-28 15:36:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 15:36:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5950336, in state SUCCEEDED (COMPLETED), ran for 215.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:36:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 15:43:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 15:43:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 15:49:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 15:49:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 15:55:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 15:55:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 16:01:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 16:01:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 16:07:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 16:07:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 16:13:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 16:13:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 16:19:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 16:19:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 16:25:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 16:25:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 16:31:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 16:31:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 16:37:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 16:37:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 16:43:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 16:43:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 16:49:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 16:49:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 16:55:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 16:55:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 17:01:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 17:01:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 17:07:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 17:07:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 17:13:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 17:13:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 17:19:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 17:19:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 17:25:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 17:25:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 17:31:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 17:31:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 17:37:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 17:37:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 17:44:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 17:44:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 17:50:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 17:50:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 17:56:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 17:56:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 18:02:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 18:02:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 18:08:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 18:08:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 18:14:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 18:14:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 18:20:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 18:20:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 18:26:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 18:26:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 18:32:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 18:32:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 18:38:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 18:38:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 18:38:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_prep -2025-07-28 18:38:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_prep -2025-07-28 18:38:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_prep succeeded, jobid=5950926 -2025-07-28 18:38:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_prep is pending at druby://130.18.14.151:32963 -2025-07-28 18:44:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_prep is success, jobid=5950927 -2025-07-28 18:44:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 18:44:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 18:44:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 18:44:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 18:50:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 18:50:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 18:50:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 18:50:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 18:56:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 18:56:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 18:56:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 18:56:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 19:02:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 19:02:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state QUEUED (PENDING) -2025-07-28 19:02:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 19:02:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 19:08:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 19:08:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950446, in state SUCCEEDED (COMPLETED), ran for 65.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:08:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 19:08:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 19:08:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_marineanlinit -2025-07-28 19:08:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_marineanlinit is pending at druby://130.18.14.151:46407 -2025-07-28 19:14:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_marineanlinit is success, jobid=5951066 -2025-07-28 19:14:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 19:14:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlinit, jobid=5951066, in state QUEUED (PENDING) -2025-07-28 19:14:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 19:14:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state QUEUED (PENDING) -2025-07-28 19:20:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 19:20:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlinit, jobid=5951066, in state QUEUED (PENDING) -2025-07-28 19:20:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 19:20:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950494, in state SUCCEEDED (COMPLETED), ran for 63.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:20:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marineanlinit -2025-07-28 19:20:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_marineanlinit is pending at druby://130.18.14.151:44625 -2025-07-28 19:26:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_marineanlinit is success, jobid=5951116 -2025-07-28 19:26:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 19:26:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlinit, jobid=5951066, in state QUEUED (PENDING) -2025-07-28 19:26:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 19:26:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlinit, jobid=5951116, in state QUEUED (PENDING) -2025-07-28 19:32:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 19:32:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlinit, jobid=5951066, in state QUEUED (PENDING) -2025-07-28 19:32:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 19:32:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlinit, jobid=5951116, in state QUEUED (PENDING) -2025-07-28 19:39:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 19:39:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlinit, jobid=5951066, in state QUEUED (PENDING) -2025-07-28 19:39:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 19:39:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlinit, jobid=5951116, in state QUEUED (PENDING) -2025-07-28 19:45:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 19:45:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlinit, jobid=5951066, in state QUEUED (PENDING) -2025-07-28 19:45:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 19:45:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlinit, jobid=5951116, in state QUEUED (PENDING) -2025-07-28 19:51:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 19:51:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlinit, jobid=5951066, in state QUEUED (PENDING) -2025-07-28 19:51:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 19:51:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlinit, jobid=5951116, in state QUEUED (PENDING) -2025-07-28 19:57:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 19:57:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlinit, jobid=5951066, in state QUEUED (PENDING) -2025-07-28 19:57:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 19:57:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlinit, jobid=5951116, in state QUEUED (PENDING) -2025-07-28 20:03:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 20:03:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlinit, jobid=5951066, in state QUEUED (PENDING) -2025-07-28 20:03:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 20:03:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlinit, jobid=5951116, in state QUEUED (PENDING) -2025-07-28 20:09:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 20:09:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlinit, jobid=5951066, in state QUEUED (PENDING) -2025-07-28 20:09:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 20:09:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlinit, jobid=5951116, in state QUEUED (PENDING) -2025-07-28 20:15:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 20:15:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlinit, jobid=5951066, in state QUEUED (PENDING) -2025-07-28 20:15:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 20:15:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlinit, jobid=5951116, in state QUEUED (PENDING) -2025-07-28 20:21:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 20:21:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlinit, jobid=5951066, in state QUEUED (PENDING) -2025-07-28 20:21:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 20:21:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlinit, jobid=5951116, in state QUEUED (PENDING) -2025-07-28 20:27:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 20:27:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlinit, jobid=5951066, in state SUCCEEDED (COMPLETED), ran for 34.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 20:27:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 20:27:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlinit, jobid=5951116, in state SUCCEEDED (COMPLETED), ran for 34.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 20:27:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_marineanlvar -2025-07-28 20:27:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marineanlvar -2025-07-28 20:27:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_marineanlvar succeeded, jobid=5951349 -2025-07-28 20:27:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_marineanlvar is pending at druby://130.18.14.151:40471 -2025-07-28 20:33:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_marineanlvar is success, jobid=5951350 -2025-07-28 20:33:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 20:33:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlvar, jobid=5951349, in state QUEUED (PENDING) -2025-07-28 20:33:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 20:33:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951350, in state QUEUED (PENDING) -2025-07-28 20:39:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 20:39:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlvar, jobid=5951349, in state QUEUED (PENDING) -2025-07-28 20:39:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 20:39:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951350, in state QUEUED (PENDING) -2025-07-28 20:45:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 20:45:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlvar, jobid=5951349, in state QUEUED (PENDING) -2025-07-28 20:45:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 20:45:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951350, in state QUEUED (PENDING) -2025-07-28 20:51:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 20:51:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlvar, jobid=5951349, in state QUEUED (PENDING) -2025-07-28 20:51:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 20:51:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951350, in state QUEUED (PENDING) -2025-07-28 20:57:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 20:57:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlvar, jobid=5951349, in state QUEUED (PENDING) -2025-07-28 20:57:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 20:57:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951350, in state QUEUED (PENDING) -2025-07-28 21:03:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 21:03:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlvar, jobid=5951349, in state QUEUED (PENDING) -2025-07-28 21:03:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 21:03:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951350, in state QUEUED (PENDING) -2025-07-28 21:09:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 21:09:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlvar, jobid=5951349, in state QUEUED (PENDING) -2025-07-28 21:09:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 21:09:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951350, in state QUEUED (PENDING) -2025-07-28 21:15:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state QUEUED (PENDING) -2025-07-28 21:15:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlvar, jobid=5951349, in state QUEUED (PENDING) -2025-07-28 21:15:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state QUEUED (PENDING) -2025-07-28 21:15:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951350, in state QUEUED (PENDING) -2025-07-28 21:21:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950926, in state SUCCEEDED (COMPLETED), ran for 99.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlvar, jobid=5951349, in state QUEUED (PENDING) -2025-07-28 21:21:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950927, in state SUCCEEDED (COMPLETED), ran for 102.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951350, in state QUEUED (PENDING) -2025-07-28 21:21:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_anal -2025-07-28 21:21:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_waveinit -2025-07-28 21:21:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_anal -2025-07-28 21:21:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_waveinit -2025-07-28 21:21:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_anal succeeded, jobid=5951612 -2025-07-28 21:21:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_waveinit succeeded, jobid=5951613 -2025-07-28 21:21:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_anal succeeded, jobid=5951614 -2025-07-28 21:21:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_waveinit succeeded, jobid=5951615 -2025-07-28 21:27:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951612, in state QUEUED (PENDING) -2025-07-28 21:27:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlvar, jobid=5951349, in state SUCCEEDED (COMPLETED), ran for 61.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_waveinit, jobid=5951613, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951614, in state QUEUED (PENDING) -2025-07-28 21:27:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951350, in state SUCCEEDED (COMPLETED), ran for 60.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_waveinit, jobid=5951615, in state SUCCEEDED (COMPLETED), ran for 35.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_marineanlchkpt -2025-07-28 21:27:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marineanlchkpt -2025-07-28 21:27:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_marineanlchkpt succeeded, jobid=5951668 -2025-07-28 21:27:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_marineanlchkpt is pending at druby://130.18.14.151:45339 -2025-07-28 21:33:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_marineanlchkpt is success, jobid=5951669 -2025-07-28 21:33:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951612, in state QUEUED (PENDING) -2025-07-28 21:33:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlchkpt, jobid=5951668, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951614, in state QUEUED (PENDING) -2025-07-28 21:33:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlchkpt, jobid=5951669, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_marineanlfinal -2025-07-28 21:33:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marineanlfinal -2025-07-28 21:33:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_marineanlfinal succeeded, jobid=5951715 -2025-07-28 21:33:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_marineanlfinal is pending at druby://130.18.14.151:35455 -2025-07-28 21:40:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_marineanlfinal is success, jobid=5951716 -2025-07-28 21:40:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951612, in state RUNNING (RUNNING) -2025-07-28 21:40:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlfinal, jobid=5951715, in state QUEUED (PENDING) -2025-07-28 21:40:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951614, in state RUNNING (RUNNING) -2025-07-28 21:40:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlfinal, jobid=5951716, in state QUEUED (PENDING) -2025-07-28 21:46:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951612, in state RUNNING (RUNNING) -2025-07-28 21:46:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlfinal, jobid=5951715, in state QUEUED (PENDING) -2025-07-28 21:46:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951614, in state RUNNING (RUNNING) -2025-07-28 21:46:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlfinal, jobid=5951716, in state QUEUED (PENDING) -2025-07-28 21:52:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951612, in state SUCCEEDED (COMPLETED), ran for 669.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:52:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlfinal, jobid=5951715, in state SUCCEEDED (COMPLETED), ran for 35.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:52:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951614, in state RUNNING (RUNNING) -2025-07-28 21:52:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlfinal, jobid=5951716, in state SUCCEEDED (COMPLETED), ran for 35.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:52:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_sfcanl -2025-07-28 21:52:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_vminmon -2025-07-28 21:52:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_anlstat -2025-07-28 21:52:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_anlstat -2025-07-28 21:52:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_sfcanl succeeded, jobid=5951881 -2025-07-28 21:52:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_vminmon succeeded, jobid=5951882 -2025-07-28 21:52:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_anlstat succeeded, jobid=5951883 -2025-07-28 21:52:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_anlstat succeeded, jobid=5951884 -2025-07-28 21:58:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_sfcanl, jobid=5951881, in state QUEUED (PENDING) -2025-07-28 21:58:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_vminmon, jobid=5951882, in state SUCCEEDED (COMPLETED), ran for 16.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:58:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anlstat, jobid=5951883, in state SUCCEEDED (COMPLETED), ran for 31.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:58:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951614, in state SUCCEEDED (COMPLETED), ran for 1003.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:58:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anlstat, jobid=5951884, in state SUCCEEDED (COMPLETED), ran for 31.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:58:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_sfcanl -2025-07-28 21:58:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_analdiag -2025-07-28 21:58:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_vminmon -2025-07-28 21:58:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_sfcanl succeeded, jobid=5951903 -2025-07-28 21:58:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_analdiag succeeded, jobid=5951904 -2025-07-28 21:58:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_vminmon succeeded, jobid=5951905 -2025-07-28 22:04:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_sfcanl, jobid=5951881, in state QUEUED (PENDING) -2025-07-28 22:04:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5951903, in state QUEUED (PENDING) -2025-07-28 22:04:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951904, in state QUEUED (PENDING) -2025-07-28 22:04:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_vminmon, jobid=5951905, in state SUCCEEDED (COMPLETED), ran for 16.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_sfcanl, jobid=5951881, in state QUEUED (PENDING) -2025-07-28 22:10:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5951903, in state QUEUED (PENDING) -2025-07-28 22:10:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951904, in state QUEUED (PENDING) -2025-07-28 22:16:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_sfcanl, jobid=5951881, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:16:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5951903, in state QUEUED (PENDING) -2025-07-28 22:16:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951904, in state QUEUED (PENDING) -2025-07-28 22:16:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_analcalc -2025-07-28 22:16:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_fcst_seg0 -2025-07-28 22:16:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_analcalc succeeded, jobid=5951985 -2025-07-28 22:16:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_fcst_seg0 succeeded, jobid=5951986 -2025-07-28 22:22:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951985, in state QUEUED (PENDING) -2025-07-28 22:22:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951986, in state QUEUED (PENDING) -2025-07-28 22:22:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5951903, in state QUEUED (PENDING) -2025-07-28 22:22:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951904, in state QUEUED (PENDING) -2025-07-28 22:28:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951985, in state QUEUED (PENDING) -2025-07-28 22:28:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951986, in state QUEUED (PENDING) -2025-07-28 22:28:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5951903, in state QUEUED (PENDING) -2025-07-28 22:28:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951904, in state QUEUED (PENDING) -2025-07-28 22:34:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951985, in state QUEUED (PENDING) -2025-07-28 22:34:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951986, in state QUEUED (PENDING) -2025-07-28 22:34:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5951903, in state QUEUED (PENDING) -2025-07-28 22:34:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951904, in state QUEUED (PENDING) -2025-07-28 22:40:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951985, in state QUEUED (PENDING) -2025-07-28 22:40:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951986, in state QUEUED (PENDING) -2025-07-28 22:40:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5951903, in state QUEUED (PENDING) -2025-07-28 22:40:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951904, in state RUNNING (RUNNING) -2025-07-28 22:46:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951985, in state QUEUED (PENDING) -2025-07-28 22:46:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951986, in state QUEUED (PENDING) -2025-07-28 22:46:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5951903, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:46:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951904, in state SUCCEEDED (COMPLETED), ran for 124.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:46:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_analcalc -2025-07-28 22:46:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_fcst_seg0 -2025-07-28 22:46:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_verfozn -2025-07-28 22:46:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_verfrad -2025-07-28 22:46:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_analcalc succeeded, jobid=5952038 -2025-07-28 22:46:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_fcst_seg0 succeeded, jobid=5952039 -2025-07-28 22:46:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_verfozn succeeded, jobid=5952040 -2025-07-28 22:46:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_verfrad succeeded, jobid=5952041 -2025-07-28 22:52:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951985, in state QUEUED (PENDING) -2025-07-28 22:52:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951986, in state QUEUED (PENDING) -2025-07-28 22:52:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952038, in state QUEUED (PENDING) -2025-07-28 22:52:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952039, in state QUEUED (PENDING) -2025-07-28 22:52:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfozn, jobid=5952040, in state SUCCEEDED (COMPLETED), ran for 63.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:52:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5952041, in state RUNNING (RUNNING) -2025-07-28 22:58:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951985, in state QUEUED (PENDING) -2025-07-28 22:58:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951986, in state QUEUED (PENDING) -2025-07-28 22:58:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952038, in state QUEUED (PENDING) -2025-07-28 22:58:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952039, in state QUEUED (PENDING) -2025-07-28 22:58:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5952041, in state RUNNING (RUNNING) -2025-07-28 23:04:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951985, in state QUEUED (PENDING) -2025-07-28 23:04:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951986, in state QUEUED (PENDING) -2025-07-28 23:04:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952038, in state QUEUED (PENDING) -2025-07-28 23:04:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952039, in state QUEUED (PENDING) -2025-07-28 23:04:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5952041, in state SUCCEEDED (COMPLETED), ran for 660.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:10:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951985, in state QUEUED (PENDING) -2025-07-28 23:10:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951986, in state QUEUED (PENDING) -2025-07-28 23:10:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952038, in state QUEUED (PENDING) -2025-07-28 23:10:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952039, in state QUEUED (PENDING) -2025-07-28 23:16:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951985, in state QUEUED (PENDING) -2025-07-28 23:16:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951986, in state QUEUED (PENDING) -2025-07-28 23:16:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952038, in state QUEUED (PENDING) -2025-07-28 23:16:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952039, in state QUEUED (PENDING) -2025-07-28 23:22:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951985, in state QUEUED (PENDING) -2025-07-28 23:22:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951986, in state QUEUED (PENDING) -2025-07-28 23:22:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952038, in state QUEUED (PENDING) -2025-07-28 23:22:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952039, in state QUEUED (PENDING) -2025-07-28 23:28:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951985, in state QUEUED (PENDING) -2025-07-28 23:28:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951986, in state RUNNING (RUNNING) -2025-07-28 23:28:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952038, in state QUEUED (PENDING) -2025-07-28 23:28:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952039, in state QUEUED (PENDING) -2025-07-28 23:28:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f000-f002 -2025-07-28 23:28:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_wavepostsbs_f000-f002 is pending at druby://130.18.14.151:43153 -2025-07-28 23:34:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_wavepostsbs_f000-f002 is success, jobid=5952162 -2025-07-28 23:34:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951985, in state QUEUED (PENDING) -2025-07-28 23:34:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951986, in state RUNNING (RUNNING) -2025-07-28 23:34:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5952162, in state QUEUED (PENDING) -2025-07-28 23:34:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952038, in state QUEUED (PENDING) -2025-07-28 23:34:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952039, in state QUEUED (PENDING) -2025-07-28 23:34:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f006 -2025-07-28 23:34:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f012 -2025-07-28 23:34:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f006 -2025-07-28 23:34:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f012 -2025-07-28 23:34:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f018 -2025-07-28 23:34:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f024 -2025-07-28 23:34:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f000-f002 -2025-07-28 23:34:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f003-f005 -2025-07-28 23:34:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f006-f008 -2025-07-28 23:34:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f009-f011 -2025-07-28 23:34:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f012-f014 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f015-f017 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f018-f020 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f021-f023 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f003-f005 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f006-f008 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f009-f011 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f012-f014 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f015-f017 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f018-f020 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f021-f023, because maximum global task throttle of 25 will be violated. -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f006 succeeded, jobid=5952170 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f012 succeeded, jobid=5952171 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f006 succeeded, jobid=5952172 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f012 succeeded, jobid=5952173 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f018 succeeded, jobid=5952174 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f024 succeeded, jobid=5952175 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f000-f002 succeeded, jobid=5952176 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f003-f005 succeeded, jobid=5952177 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f006-f008 succeeded, jobid=5952178 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f009-f011 succeeded, jobid=5952179 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f012-f014 succeeded, jobid=5952180 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f015-f017 succeeded, jobid=5952181 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f018-f020 succeeded, jobid=5952182 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f021-f023 succeeded, jobid=5952183 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f003-f005 succeeded, jobid=5952184 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f006-f008 succeeded, jobid=5952185 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f009-f011 succeeded, jobid=5952186 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f012-f014 succeeded, jobid=5952187 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f015-f017 succeeded, jobid=5952188 -2025-07-28 23:34:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f018-f020 succeeded, jobid=5952189 -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951985, in state QUEUED (PENDING) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951986, in state RUNNING (RUNNING) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f006, jobid=5952170, in state SUCCEEDED (COMPLETED), ran for 22.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f012, jobid=5952171, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f006, jobid=5952172, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f012, jobid=5952173, in state SUCCEEDED (COMPLETED), ran for 20.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f018, jobid=5952174, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f024, jobid=5952175, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952176, in state QUEUED (PENDING) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952177, in state QUEUED (PENDING) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952178, in state QUEUED (PENDING) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952179, in state QUEUED (PENDING) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952180, in state QUEUED (PENDING) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952181, in state QUEUED (PENDING) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952182, in state QUEUED (PENDING) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952183, in state QUEUED (PENDING) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5952162, in state QUEUED (PENDING) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5952184, in state QUEUED (PENDING) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5952185, in state QUEUED (PENDING) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5952186, in state QUEUED (PENDING) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5952187, in state QUEUED (PENDING) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5952188, in state QUEUED (PENDING) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5952189, in state QUEUED (PENDING) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952038, in state QUEUED (PENDING) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952039, in state QUEUED (PENDING) -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f018 -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f024 -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f030 -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f036 -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f042 -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f030 -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f021-f023, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f018 succeeded, jobid=5952235 -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f024 succeeded, jobid=5952236 -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f030 succeeded, jobid=5952237 -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f036 succeeded, jobid=5952238 -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f042 succeeded, jobid=5952239 -2025-07-28 23:41:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f030 succeeded, jobid=5952240 -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951985, in state QUEUED (PENDING) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951986, in state RUNNING (RUNNING) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f018, jobid=5952235, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f024, jobid=5952236, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f030, jobid=5952237, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f036, jobid=5952238, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f042, jobid=5952239, in state SUCCEEDED (COMPLETED), ran for 22.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f030, jobid=5952240, in state SUCCEEDED (COMPLETED), ran for 22.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952176, in state QUEUED (PENDING) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952177, in state QUEUED (PENDING) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952178, in state QUEUED (PENDING) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952179, in state QUEUED (PENDING) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952180, in state QUEUED (PENDING) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952181, in state QUEUED (PENDING) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952182, in state QUEUED (PENDING) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952183, in state QUEUED (PENDING) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5952162, in state QUEUED (PENDING) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5952184, in state QUEUED (PENDING) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5952185, in state QUEUED (PENDING) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5952186, in state QUEUED (PENDING) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5952187, in state QUEUED (PENDING) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5952188, in state QUEUED (PENDING) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5952189, in state QUEUED (PENDING) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952038, in state QUEUED (PENDING) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952039, in state QUEUED (PENDING) -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f048 -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f054 -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f060 -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f066 -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f072 -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f078 -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f021-f023, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f048 succeeded, jobid=5952263 -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f054 succeeded, jobid=5952265 -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f060 succeeded, jobid=5952264 -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f066 succeeded, jobid=5952268 -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f072 succeeded, jobid=5952266 -2025-07-28 23:47:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f078 succeeded, jobid=5952267 -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951985, in state QUEUED (PENDING) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951986, in state SUCCEEDED (COMPLETED), ran for 1550.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f048, jobid=5952263, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f054, jobid=5952265, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f060, jobid=5952264, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f066, jobid=5952268, in state SUCCEEDED (COMPLETED), ran for 23.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f072, jobid=5952266, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f078, jobid=5952267, in state SUCCEEDED (COMPLETED), ran for 23.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952176, in state QUEUED (PENDING) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952177, in state QUEUED (PENDING) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952178, in state QUEUED (PENDING) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952179, in state QUEUED (PENDING) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952180, in state QUEUED (PENDING) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952181, in state QUEUED (PENDING) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952182, in state QUEUED (PENDING) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952183, in state QUEUED (PENDING) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5952162, in state QUEUED (PENDING) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5952184, in state QUEUED (PENDING) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5952185, in state QUEUED (PENDING) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5952186, in state QUEUED (PENDING) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5952187, in state QUEUED (PENDING) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5952188, in state QUEUED (PENDING) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5952189, in state QUEUED (PENDING) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952038, in state QUEUED (PENDING) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952039, in state QUEUED (PENDING) -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f084 -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f090 -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f096 -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f102 -2025-07-28 23:53:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f108 -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f114 -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f120 -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f021-f023, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f084 succeeded, jobid=5952274 -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f090 succeeded, jobid=5952275 -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f096 succeeded, jobid=5952276 -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f102 succeeded, jobid=5952277 -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f108 succeeded, jobid=5952278 -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f114 succeeded, jobid=5952279 -2025-07-28 23:53:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f120 succeeded, jobid=5952280 -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951985, in state SUCCEEDED (COMPLETED), ran for 43.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f084, jobid=5952274, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f090, jobid=5952275, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f096, jobid=5952276, in state SUCCEEDED (COMPLETED), ran for 20.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f102, jobid=5952277, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f108, jobid=5952278, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f114, jobid=5952279, in state SUCCEEDED (COMPLETED), ran for 23.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f120, jobid=5952280, in state SUCCEEDED (COMPLETED), ran for 23.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952176, in state QUEUED (PENDING) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952177, in state QUEUED (PENDING) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952178, in state QUEUED (PENDING) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952179, in state QUEUED (PENDING) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952180, in state QUEUED (PENDING) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952181, in state QUEUED (PENDING) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952182, in state QUEUED (PENDING) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952183, in state QUEUED (PENDING) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5952162, in state QUEUED (PENDING) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5952184, in state QUEUED (PENDING) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5952185, in state QUEUED (PENDING) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5952186, in state QUEUED (PENDING) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5952187, in state QUEUED (PENDING) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5952188, in state QUEUED (PENDING) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5952189, in state QUEUED (PENDING) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952038, in state QUEUED (PENDING) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952039, in state QUEUED (PENDING) -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmanlupp -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f036 -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f042 -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f048 -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f054 -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f060 -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f066 -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f072 -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f021-f023, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmanlupp succeeded, jobid=5952287 -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f036 succeeded, jobid=5952288 -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f042 succeeded, jobid=5952289 -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f048 succeeded, jobid=5952290 -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f054 succeeded, jobid=5952291 -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f060 succeeded, jobid=5952292 -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f066 succeeded, jobid=5952293 -2025-07-28 23:59:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f072 succeeded, jobid=5952294 -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952287, in state QUEUED (PENDING) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f036, jobid=5952288, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f042, jobid=5952289, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f048, jobid=5952290, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f054, jobid=5952291, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f060, jobid=5952292, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f066, jobid=5952293, in state SUCCEEDED (COMPLETED), ran for 22.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f072, jobid=5952294, in state SUCCEEDED (COMPLETED), ran for 22.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952176, in state QUEUED (PENDING) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952177, in state QUEUED (PENDING) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952178, in state QUEUED (PENDING) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952179, in state QUEUED (PENDING) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952180, in state QUEUED (PENDING) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952181, in state QUEUED (PENDING) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952182, in state QUEUED (PENDING) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952183, in state QUEUED (PENDING) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5952162, in state QUEUED (PENDING) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5952184, in state QUEUED (PENDING) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5952185, in state QUEUED (PENDING) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5952186, in state QUEUED (PENDING) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5952187, in state QUEUED (PENDING) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5952188, in state QUEUED (PENDING) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5952189, in state QUEUED (PENDING) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952038, in state SUCCEEDED (COMPLETED), ran for 31.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952039, in state SUCCEEDED (COMPLETED), ran for 296.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f078 -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f084 -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f090 -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f096 -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f102 -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f108 -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f114 -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f120 -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f024-f026 -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f021-f023, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmanlupp, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f000, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f003, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f006, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f009, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f000, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f001, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f002, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f003, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f004, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f005, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f006, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f007, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f008, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f009, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f078 succeeded, jobid=5952335 -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f084 succeeded, jobid=5952336 -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f090 succeeded, jobid=5952337 -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f096 succeeded, jobid=5952338 -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f102 succeeded, jobid=5952339 -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f108 succeeded, jobid=5952340 -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f114 succeeded, jobid=5952341 -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f120 succeeded, jobid=5952342 -2025-07-29 00:05:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f024-f026 succeeded, jobid=5952343 -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952287, in state QUEUED (PENDING) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f078, jobid=5952335, in state SUCCEEDED (COMPLETED), ran for 24.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f084, jobid=5952336, in state SUCCEEDED (COMPLETED), ran for 24.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f090, jobid=5952337, in state SUCCEEDED (COMPLETED), ran for 24.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f096, jobid=5952338, in state SUCCEEDED (COMPLETED), ran for 16.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f102, jobid=5952339, in state SUCCEEDED (COMPLETED), ran for 22.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f108, jobid=5952340, in state SUCCEEDED (COMPLETED), ran for 22.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f114, jobid=5952341, in state SUCCEEDED (COMPLETED), ran for 22.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f120, jobid=5952342, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952176, in state QUEUED (PENDING) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952177, in state QUEUED (PENDING) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952178, in state QUEUED (PENDING) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952179, in state QUEUED (PENDING) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952180, in state QUEUED (PENDING) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952181, in state QUEUED (PENDING) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952182, in state QUEUED (PENDING) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952183, in state QUEUED (PENDING) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952343, in state QUEUED (PENDING) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5952162, in state QUEUED (PENDING) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5952184, in state QUEUED (PENDING) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5952185, in state QUEUED (PENDING) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5952186, in state QUEUED (PENDING) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5952187, in state QUEUED (PENDING) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5952188, in state QUEUED (PENDING) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5952189, in state QUEUED (PENDING) -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f027-f029 -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f030-f032 -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f033-f035 -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f036-f038 -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f039-f041 -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f042-f044 -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f045-f047 -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f048-f054 -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f021-f023, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmanlupp, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f000, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f003, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f006, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f009, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f000, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f001, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f002, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f003, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f004, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f005, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f006, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f007, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f008, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f009, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f027-f029 succeeded, jobid=5952384 -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f030-f032 succeeded, jobid=5952385 -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f033-f035 succeeded, jobid=5952386 -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f036-f038 succeeded, jobid=5952387 -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f039-f041 succeeded, jobid=5952388 -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f042-f044 succeeded, jobid=5952389 -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f045-f047 succeeded, jobid=5952390 -2025-07-29 00:11:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f048-f054 succeeded, jobid=5952391 -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952287, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952176, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952177, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952178, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952179, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952180, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952181, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952182, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952183, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952343, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952384, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952385, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952386, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952387, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952388, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952389, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952390, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952391, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5952162, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5952184, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5952185, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5952186, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5952187, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5952188, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5952189, in state QUEUED (PENDING) -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f021-f023, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmanlupp, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f000, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f003, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f006, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f009, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f000, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f001, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f002, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f003, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f004, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f005, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f006, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f007, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f008, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f009, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952287, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952176, in state RUNNING (RUNNING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952177, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952178, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952179, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952180, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952181, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952182, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952183, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952343, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952384, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952385, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952386, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952387, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952388, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952389, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952390, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952391, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5952162, in state SUCCEEDED (COMPLETED), ran for 75.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5952184, in state SUCCEEDED (COMPLETED), ran for 76.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5952185, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5952186, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5952187, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5952188, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5952189, in state QUEUED (PENDING) -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f057-f063 -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f066-f072 -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f021-f023, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmanlupp, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f000, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f003, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f006, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f009, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f000, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f001, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f002, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f003, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f004, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f005, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f006, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f007, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f008, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f009, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f057-f063 succeeded, jobid=5952443 -2025-07-29 00:23:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f066-f072 succeeded, jobid=5952444 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952287, in state QUEUED (PENDING) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952176, in state SUCCEEDED (COMPLETED), ran for 265.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952177, in state SUCCEEDED (COMPLETED), ran for 275.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952178, in state SUCCEEDED (COMPLETED), ran for 271.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952179, in state SUCCEEDED (COMPLETED), ran for 274.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952180, in state SUCCEEDED (COMPLETED), ran for 273.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952181, in state SUCCEEDED (COMPLETED), ran for 271.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952182, in state SUCCEEDED (COMPLETED), ran for 272.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952183, in state SUCCEEDED (COMPLETED), ran for 272.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952343, in state QUEUED (PENDING) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952384, in state QUEUED (PENDING) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952385, in state QUEUED (PENDING) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952386, in state QUEUED (PENDING) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952387, in state QUEUED (PENDING) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952388, in state QUEUED (PENDING) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952389, in state QUEUED (PENDING) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952390, in state QUEUED (PENDING) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952391, in state QUEUED (PENDING) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952443, in state QUEUED (PENDING) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952444, in state QUEUED (PENDING) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5952185, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5952186, in state SUCCEEDED (COMPLETED), ran for 75.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5952187, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5952188, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5952189, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f075-f081 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f084-f090 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f093-f099 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f102-f108 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f111-f114 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f117-f120 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f021-f023 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f024-f026 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f027-f029 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f030-f032 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f033-f035 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f036-f038 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f039-f041 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmanlupp, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f000, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f003, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f006, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f009, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f000, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f001, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f002, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f003, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f004, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f005, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f006, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f007, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f008, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f009, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f075-f081 succeeded, jobid=5952472 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f084-f090 succeeded, jobid=5952473 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f093-f099 succeeded, jobid=5952474 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f102-f108 succeeded, jobid=5952475 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f111-f114 succeeded, jobid=5952476 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f117-f120 succeeded, jobid=5952477 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f021-f023 succeeded, jobid=5952478 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f024-f026 succeeded, jobid=5952479 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f027-f029 succeeded, jobid=5952480 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f030-f032 succeeded, jobid=5952481 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f033-f035 succeeded, jobid=5952483 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f036-f038 succeeded, jobid=5952482 -2025-07-29 00:29:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f039-f041 succeeded, jobid=5952484 -2025-07-29 00:35:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952287, in state SUCCEEDED (COMPLETED), ran for 43.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:35:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952343, in state SUCCEEDED (COMPLETED), ran for 283.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:35:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952384, in state RUNNING (RUNNING) -2025-07-29 00:35:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952385, in state RUNNING (RUNNING) -2025-07-29 00:35:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952386, in state RUNNING (RUNNING) -2025-07-29 00:35:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952387, in state RUNNING (RUNNING) -2025-07-29 00:35:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952388, in state RUNNING (RUNNING) -2025-07-29 00:35:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952389, in state RUNNING (RUNNING) -2025-07-29 00:35:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952390, in state RUNNING (RUNNING) -2025-07-29 00:35:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952391, in state RUNNING (RUNNING) -2025-07-29 00:35:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952443, in state QUEUED (PENDING) -2025-07-29 00:35:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952444, in state QUEUED (PENDING) -2025-07-29 00:35:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952472, in state QUEUED (PENDING) -2025-07-29 00:35:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952473, in state QUEUED (PENDING) -2025-07-29 00:35:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952474, in state QUEUED (PENDING) -2025-07-29 00:35:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952475, in state QUEUED (PENDING) -2025-07-29 00:35:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5952476, in state QUEUED (PENDING) -2025-07-29 00:35:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5952477, in state QUEUED (PENDING) -2025-07-29 00:35:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5952478, in state RUNNING (RUNNING) -2025-07-29 00:35:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5952479, in state RUNNING (RUNNING) -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5952480, in state RUNNING (RUNNING) -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5952481, in state RUNNING (RUNNING) -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f033-f035, jobid=5952483, in state RUNNING (RUNNING) -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f036-f038, jobid=5952482, in state RUNNING (RUNNING) -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f039-f041, jobid=5952484, in state RUNNING (RUNNING) -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmanlprod -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f042-f044 -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmanlupp, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f000, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f003, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f006, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f009, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f000, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f001, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f002, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f003, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f004, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f005, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f006, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f007, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f008, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f009, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmanlprod succeeded, jobid=5952553 -2025-07-29 00:35:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f042-f044 succeeded, jobid=5952554 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlprod, jobid=5952553, in state QUEUED (PENDING) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952384, in state SUCCEEDED (COMPLETED), ran for 271.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952385, in state SUCCEEDED (COMPLETED), ran for 270.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952386, in state SUCCEEDED (COMPLETED), ran for 272.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952387, in state SUCCEEDED (COMPLETED), ran for 269.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952388, in state SUCCEEDED (COMPLETED), ran for 275.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952389, in state SUCCEEDED (COMPLETED), ran for 270.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952390, in state SUCCEEDED (COMPLETED), ran for 272.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952391, in state SUCCEEDED (COMPLETED), ran for 320.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952443, in state RUNNING (RUNNING) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952444, in state RUNNING (RUNNING) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952472, in state RUNNING (RUNNING) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952473, in state RUNNING (RUNNING) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952474, in state RUNNING (RUNNING) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952475, in state RUNNING (RUNNING) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5952476, in state RUNNING (RUNNING) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5952477, in state RUNNING (RUNNING) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5952478, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5952479, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5952480, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5952481, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f033-f035, jobid=5952483, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f036-f038, jobid=5952482, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f039-f041, jobid=5952484, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f042-f044, jobid=5952554, in state QUEUED (PENDING) -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f045-f047 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f048-f054 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f057-f063 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f066-f072 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f075-f081 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f084-f090 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f093-f099 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f102-f108 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f111-f114 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f117-f120 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostbndpnt -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostbndpntbll -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostpnt -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmanlupp -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f000 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f003, because maximum global task throttle of 25 will be violated. -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f006, because maximum global task throttle of 25 will be violated. -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f009, because maximum global task throttle of 25 will be violated. -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f000, because maximum global task throttle of 25 will be violated. -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f001, because maximum global task throttle of 25 will be violated. -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f002, because maximum global task throttle of 25 will be violated. -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f003, because maximum global task throttle of 25 will be violated. -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f004, because maximum global task throttle of 25 will be violated. -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f005, because maximum global task throttle of 25 will be violated. -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f006, because maximum global task throttle of 25 will be violated. -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f007, because maximum global task throttle of 25 will be violated. -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f008, because maximum global task throttle of 25 will be violated. -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f009, because maximum global task throttle of 25 will be violated. -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f045-f047 succeeded, jobid=5952579 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f048-f054 succeeded, jobid=5952580 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f057-f063 succeeded, jobid=5952581 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f066-f072 succeeded, jobid=5952582 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f075-f081 succeeded, jobid=5952583 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f084-f090 succeeded, jobid=5952584 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f093-f099 succeeded, jobid=5952585 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f102-f108 succeeded, jobid=5952586 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f111-f114 succeeded, jobid=5952587 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f117-f120 succeeded, jobid=5952588 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostbndpnt succeeded, jobid=5952589 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostbndpntbll succeeded, jobid=5952590 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostpnt succeeded, jobid=5952591 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmanlupp succeeded, jobid=5952592 -2025-07-29 00:41:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f000 succeeded, jobid=5952593 -2025-07-29 00:47:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlprod, jobid=5952553, in state QUEUED (PENDING) -2025-07-29 00:47:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952443, in state SUCCEEDED (COMPLETED), ran for 325.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:47:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952444, in state SUCCEEDED (COMPLETED), ran for 320.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:47:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952472, in state SUCCEEDED (COMPLETED), ran for 325.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:47:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952473, in state SUCCEEDED (COMPLETED), ran for 325.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:47:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952474, in state SUCCEEDED (COMPLETED), ran for 314.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:47:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952475, in state SUCCEEDED (COMPLETED), ran for 333.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:47:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5952476, in state SUCCEEDED (COMPLETED), ran for 213.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:47:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5952477, in state SUCCEEDED (COMPLETED), ran for 222.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:47:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f042-f044, jobid=5952554, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:47:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f045-f047, jobid=5952579, in state QUEUED (PENDING) -2025-07-29 00:47:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f048-f054, jobid=5952580, in state QUEUED (PENDING) -2025-07-29 00:47:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f057-f063, jobid=5952581, in state QUEUED (PENDING) -2025-07-29 00:47:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f066-f072, jobid=5952582, in state QUEUED (PENDING) -2025-07-29 00:47:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f075-f081, jobid=5952583, in state QUEUED (PENDING) -2025-07-29 00:47:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f084-f090, jobid=5952584, in state QUEUED (PENDING) -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f093-f099, jobid=5952585, in state QUEUED (PENDING) -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f102-f108, jobid=5952586, in state QUEUED (PENDING) -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f111-f114, jobid=5952587, in state QUEUED (PENDING) -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f117-f120, jobid=5952588, in state QUEUED (PENDING) -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostbndpnt, jobid=5952589, in state SUCCEEDED (COMPLETED), ran for 164.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostbndpntbll, jobid=5952590, in state SUCCEEDED (COMPLETED), ran for 74.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostpnt, jobid=5952591, in state SUCCEEDED (COMPLETED), ran for 106.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952592, in state QUEUED (PENDING) -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952593, in state QUEUED (PENDING) -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_tracker -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_genesis -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f003 -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f006 -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f009 -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostsbs_f000 -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostsbs_f001 -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostsbs_f002 -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostsbs_f003 -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostsbs_f004 -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostsbs_f005 -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostsbs_f006 -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f007, because maximum global task throttle of 25 will be violated. -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f008, because maximum global task throttle of 25 will be violated. -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f009, because maximum global task throttle of 25 will be violated. -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_tracker succeeded, jobid=5952629 -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_genesis succeeded, jobid=5952630 -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f003 succeeded, jobid=5952631 -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f006 succeeded, jobid=5952632 -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f009 succeeded, jobid=5952633 -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostsbs_f000 succeeded, jobid=5952634 -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostsbs_f001 succeeded, jobid=5952635 -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostsbs_f002 succeeded, jobid=5952636 -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostsbs_f003 succeeded, jobid=5952637 -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostsbs_f004 succeeded, jobid=5952638 -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostsbs_f005 succeeded, jobid=5952639 -2025-07-29 00:47:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostsbs_f006 succeeded, jobid=5952640 -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlprod, jobid=5952553, in state QUEUED (PENDING) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_tracker, jobid=5952629, in state SUCCEEDED (COMPLETED), ran for 62.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_genesis, jobid=5952630, in state RUNNING (RUNNING) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f045-f047, jobid=5952579, in state SUCCEEDED (COMPLETED), ran for 71.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f048-f054, jobid=5952580, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f057-f063, jobid=5952581, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f066-f072, jobid=5952582, in state SUCCEEDED (COMPLETED), ran for 70.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f075-f081, jobid=5952583, in state QUEUED (PENDING) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f084-f090, jobid=5952584, in state QUEUED (PENDING) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f093-f099, jobid=5952585, in state QUEUED (PENDING) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f102-f108, jobid=5952586, in state QUEUED (PENDING) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f111-f114, jobid=5952587, in state QUEUED (PENDING) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f117-f120, jobid=5952588, in state QUEUED (PENDING) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952592, in state QUEUED (PENDING) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952593, in state QUEUED (PENDING) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952631, in state QUEUED (PENDING) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952632, in state QUEUED (PENDING) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952633, in state QUEUED (PENDING) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f000, jobid=5952634, in state QUEUED (PENDING) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f001, jobid=5952635, in state QUEUED (PENDING) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f002, jobid=5952636, in state QUEUED (PENDING) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f003, jobid=5952637, in state QUEUED (PENDING) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f004, jobid=5952638, in state QUEUED (PENDING) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f005, jobid=5952639, in state QUEUED (PENDING) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f006, jobid=5952640, in state QUEUED (PENDING) -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostsbs_f007 -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostsbs_f008 -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostsbs_f009 -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostbndpnt -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostbndpntbll -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostsbs_f007 succeeded, jobid=5952658 -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostsbs_f008 succeeded, jobid=5952659 -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostsbs_f009 succeeded, jobid=5952660 -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostbndpnt succeeded, jobid=5952661 -2025-07-29 00:53:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostbndpntbll succeeded, jobid=5952662 -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlprod, jobid=5952553, in state SUCCEEDED (COMPLETED), ran for 101.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_genesis, jobid=5952630, in state SUCCEEDED (COMPLETED), ran for 484.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f075-f081, jobid=5952583, in state QUEUED (PENDING) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f084-f090, jobid=5952584, in state QUEUED (PENDING) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f093-f099, jobid=5952585, in state QUEUED (PENDING) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f102-f108, jobid=5952586, in state QUEUED (PENDING) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f111-f114, jobid=5952587, in state QUEUED (PENDING) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f117-f120, jobid=5952588, in state QUEUED (PENDING) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952592, in state QUEUED (PENDING) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952593, in state QUEUED (PENDING) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952631, in state QUEUED (PENDING) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952632, in state QUEUED (PENDING) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952633, in state QUEUED (PENDING) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f000, jobid=5952634, in state QUEUED (PENDING) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f001, jobid=5952635, in state QUEUED (PENDING) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f002, jobid=5952636, in state QUEUED (PENDING) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f003, jobid=5952637, in state QUEUED (PENDING) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f004, jobid=5952638, in state QUEUED (PENDING) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f005, jobid=5952639, in state QUEUED (PENDING) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f006, jobid=5952640, in state QUEUED (PENDING) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f007, jobid=5952658, in state QUEUED (PENDING) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f008, jobid=5952659, in state QUEUED (PENDING) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f009, jobid=5952660, in state QUEUED (PENDING) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostbndpnt, jobid=5952661, in state SUCCEEDED (COMPLETED), ran for 31.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostbndpntbll, jobid=5952662, in state SUCCEEDED (COMPLETED), ran for 28.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostpnt -2025-07-29 00:59:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_wavepostpnt is pending at druby://130.18.14.151:34453 -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_wavepostpnt is success, jobid=5952680 -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f075-f081, jobid=5952583, in state SUCCEEDED (COMPLETED), ran for 75.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f084-f090, jobid=5952584, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f093-f099, jobid=5952585, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f102-f108, jobid=5952586, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f111-f114, jobid=5952587, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f117-f120, jobid=5952588, in state SUCCEEDED (COMPLETED), ran for 48.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952592, in state QUEUED (PENDING) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952593, in state QUEUED (PENDING) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952631, in state QUEUED (PENDING) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952632, in state QUEUED (PENDING) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952633, in state QUEUED (PENDING) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f000, jobid=5952634, in state SUCCEEDED (COMPLETED), ran for 27.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f001, jobid=5952635, in state SUCCEEDED (COMPLETED), ran for 27.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f002, jobid=5952636, in state SUCCEEDED (COMPLETED), ran for 27.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f003, jobid=5952637, in state QUEUED (PENDING) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f004, jobid=5952638, in state QUEUED (PENDING) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f005, jobid=5952639, in state QUEUED (PENDING) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f006, jobid=5952640, in state QUEUED (PENDING) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f007, jobid=5952658, in state QUEUED (PENDING) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f008, jobid=5952659, in state QUEUED (PENDING) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f009, jobid=5952660, in state QUEUED (PENDING) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostpnt, jobid=5952680, in state SUCCEEDED (COMPLETED), ran for 25.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_arch_vrfy -2025-07-29 01:05:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_arch_vrfy is pending at druby://130.18.14.151:40061 -2025-07-29 01:11:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_arch_vrfy is success, jobid=5952704 -2025-07-29 01:11:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_arch_vrfy, jobid=5952704, in state SUCCEEDED (COMPLETED), ran for 60.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:11:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952592, in state SUCCEEDED (COMPLETED), ran for 33.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:11:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952593, in state SUCCEEDED (COMPLETED), ran for 63.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:11:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952631, in state SUCCEEDED (COMPLETED), ran for 66.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:11:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952632, in state SUCCEEDED (COMPLETED), ran for 65.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:11:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952633, in state SUCCEEDED (COMPLETED), ran for 66.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:11:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f003, jobid=5952637, in state SUCCEEDED (COMPLETED), ran for 27.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:11:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f004, jobid=5952638, in state SUCCEEDED (COMPLETED), ran for 26.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:11:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f005, jobid=5952639, in state SUCCEEDED (COMPLETED), ran for 26.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:11:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f006, jobid=5952640, in state SUCCEEDED (COMPLETED), ran for 27.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:11:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f007, jobid=5952658, in state QUEUED (PENDING) -2025-07-29 01:11:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f008, jobid=5952659, in state QUEUED (PENDING) -2025-07-29 01:11:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f009, jobid=5952660, in state QUEUED (PENDING) -2025-07-29 01:11:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_cleanup -2025-07-29 01:11:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_fit2obs -2025-07-29 01:11:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_cleanup is pending at druby://130.18.14.151:35421 -2025-07-29 01:11:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_fit2obs is pending at druby://130.18.14.151:35421 -2025-07-29 01:17:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_cleanup is success, jobid=5952719 -2025-07-29 01:17:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_fit2obs is success, jobid=5952720 -2025-07-29 01:17:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_cleanup, jobid=5952719, in state SUCCEEDED (COMPLETED), ran for 21.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:17:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fit2obs, jobid=5952720, in state SUCCEEDED (COMPLETED), ran for 20.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:17:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f007, jobid=5952658, in state QUEUED (PENDING) -2025-07-29 01:17:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f008, jobid=5952659, in state QUEUED (PENDING) -2025-07-29 01:17:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f009, jobid=5952660, in state QUEUED (PENDING) -2025-07-29 01:17:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmanlprod -2025-07-29 01:17:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_atmanlprod is pending at druby://130.18.14.151:37697 -2025-07-29 01:23:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_atmanlprod is success, jobid=5952726 -2025-07-29 01:23:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlprod, jobid=5952726, in state SUCCEEDED (COMPLETED), ran for 62.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:23:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f007, jobid=5952658, in state SUCCEEDED (COMPLETED), ran for 26.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:23:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f008, jobid=5952659, in state SUCCEEDED (COMPLETED), ran for 27.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:23:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f009, jobid=5952660, in state SUCCEEDED (COMPLETED), ran for 26.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:23:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_arch_vrfy -2025-07-29 01:23:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_arch_vrfy is pending at druby://130.18.14.151:34461 -2025-07-29 01:30:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_arch_vrfy is success, jobid=5952738 -2025-07-29 01:30:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_arch_vrfy, jobid=5952738, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:30:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_cleanup -2025-07-29 01:30:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_cleanup is pending at druby://130.18.14.151:34065 -2025-07-29 01:36:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_cleanup is success, jobid=5952765 -2025-07-29 01:36:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_cleanup, jobid=5952765, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:36:04 -0500 :: hercules-login-1.hpc.msstate.edu :: This cycle is complete: Success diff --git a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C48mx500_hybAOWCDA_logs_2021032418.log b/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C48mx500_hybAOWCDA_logs_2021032418.log deleted file mode 100644 index 50dcfd63..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C48mx500_hybAOWCDA_logs_2021032418.log +++ /dev/null @@ -1,155 +0,0 @@ -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_stage_ic -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_stage_ic -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_stage_ic is pending at druby://130.18.14.151:45363 -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of enkfgdas_stage_ic is pending at druby://130.18.14.151:45363 -2025-07-28 14:48:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_stage_ic is success, jobid=5950264 -2025-07-28 14:48:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending enkfgdas_stage_ic is success, jobid=5950266 -2025-07-28 14:48:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_stage_ic, jobid=5950264, in state RUNNING (RUNNING) -2025-07-28 14:48:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_stage_ic, jobid=5950266, in state RUNNING (RUNNING) -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_stage_ic, jobid=5950264, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_stage_ic, jobid=5950266, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_fcst_seg0 -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_fcst_mem001 -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_fcst_mem002 -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_fcst_seg0 succeeded, jobid=5950282 -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_fcst_mem001 succeeded, jobid=5950283 -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of enkfgdas_fcst_mem002 is pending at druby://130.18.14.151:36939 -2025-07-28 15:00:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending enkfgdas_fcst_mem002 is success, jobid=5950284 -2025-07-28 15:00:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5950282, in state RUNNING (RUNNING) -2025-07-28 15:00:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem001, jobid=5950283, in state QUEUED (PENDING) -2025-07-28 15:00:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem002, jobid=5950284, in state QUEUED (PENDING) -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5950282, in state SUCCEEDED (COMPLETED), ran for 316.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem001, jobid=5950283, in state RUNNING (RUNNING) -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem002, jobid=5950284, in state RUNNING (RUNNING) -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f000 -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f001 -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f002 -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f003 -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f004 -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f005 -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f006 -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f007 -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f008 -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f009 -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f000 succeeded, jobid=5950316 -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f001 succeeded, jobid=5950317 -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f002 succeeded, jobid=5950318 -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f003 succeeded, jobid=5950319 -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f004 succeeded, jobid=5950320 -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f005 succeeded, jobid=5950321 -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f006 succeeded, jobid=5950322 -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f007 succeeded, jobid=5950323 -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f008 succeeded, jobid=5950324 -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f009 succeeded, jobid=5950325 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950316, in state QUEUED (PENDING) -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950317, in state QUEUED (PENDING) -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950318, in state QUEUED (PENDING) -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950319, in state QUEUED (PENDING) -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950320, in state QUEUED (PENDING) -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950321, in state QUEUED (PENDING) -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950322, in state QUEUED (PENDING) -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950323, in state QUEUED (PENDING) -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950324, in state QUEUED (PENDING) -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950325, in state QUEUED (PENDING) -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem001, jobid=5950283, in state SUCCEEDED (COMPLETED), ran for 322.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem002, jobid=5950284, in state SUCCEEDED (COMPLETED), ran for 289.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_echgres -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos000 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos001 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos002 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos003 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos004 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos005 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos006 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_echgres succeeded, jobid=5950363 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos000 succeeded, jobid=5950364 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos001 succeeded, jobid=5950365 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos002 succeeded, jobid=5950366 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos003 succeeded, jobid=5950367 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos004 succeeded, jobid=5950368 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos005 succeeded, jobid=5950369 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos006 succeeded, jobid=5950370 -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950316, in state SUCCEEDED (COMPLETED), ran for 63.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950317, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950318, in state SUCCEEDED (COMPLETED), ran for 66.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950319, in state SUCCEEDED (COMPLETED), ran for 58.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950320, in state SUCCEEDED (COMPLETED), ran for 61.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950321, in state SUCCEEDED (COMPLETED), ran for 58.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950322, in state SUCCEEDED (COMPLETED), ran for 58.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950323, in state SUCCEEDED (COMPLETED), ran for 57.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950324, in state RUNNING (RUNNING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950325, in state RUNNING (RUNNING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950363, in state SUCCEEDED (COMPLETED), ran for 31.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950364, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950365, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950366, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950367, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950368, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950369, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950370, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950324, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950325, in state SUCCEEDED (COMPLETED), ran for 64.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950364, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950365, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950366, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950367, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950368, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950369, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950370, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950364, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950365, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950366, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950367, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950368, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950369, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950370, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950364, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950365, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950366, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950367, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950368, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950369, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950370, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950364, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950365, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950366, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950367, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950368, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950369, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950370, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950364, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950365, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950366, in state SUCCEEDED (COMPLETED), ran for 37.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950367, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950368, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950369, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950370, in state QUEUED (PENDING) -2025-07-28 15:55:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950367, in state QUEUED (PENDING) -2025-07-28 15:55:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950368, in state QUEUED (PENDING) -2025-07-28 15:55:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950369, in state QUEUED (PENDING) -2025-07-28 15:55:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950370, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950367, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950368, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950369, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950370, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950367, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950368, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950369, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950370, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950367, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950368, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950369, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950370, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950367, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950368, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950369, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950370, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950367, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950368, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950369, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950370, in state QUEUED (PENDING) -2025-07-28 16:31:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950367, in state SUCCEEDED (COMPLETED), ran for 37.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:31:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950368, in state SUCCEEDED (COMPLETED), ran for 37.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:31:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950369, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:31:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950370, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:31:19 -0500 :: hercules-login-1.hpc.msstate.edu :: This cycle is complete: Success diff --git a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C48mx500_hybAOWCDA_logs_2021032500.log b/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C48mx500_hybAOWCDA_logs_2021032500.log deleted file mode 100644 index 33b04a93..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C48mx500_hybAOWCDA_logs_2021032500.log +++ /dev/null @@ -1,525 +0,0 @@ -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_prepoceanobs -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_prepoceanobs succeeded, jobid=5950326 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5950326, in state SUCCEEDED (COMPLETED), ran for 215.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marinebmatinit -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_marinebmatinit succeeded, jobid=5950371 -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950371, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950371, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_prep -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_prep is pending at druby://130.18.14.151:43315 -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_prep is success, jobid=5950445 -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950371, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950371, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950371, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950371, in state QUEUED (PENDING) -2025-07-28 15:55:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 15:55:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950371, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950371, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950371, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950371, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950371, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950371, in state QUEUED (PENDING) -2025-07-28 16:31:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 16:31:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950371, in state SUCCEEDED (COMPLETED), ran for 33.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:31:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marinebmat -2025-07-28 16:31:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_ocnanalecen -2025-07-28 16:31:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_marinebmat is pending at druby://130.18.14.151:40859 -2025-07-28 16:31:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_ocnanalecen is pending at druby://130.18.14.151:40859 -2025-07-28 16:37:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_marinebmat is success, jobid=5950656 -2025-07-28 16:37:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_ocnanalecen is success, jobid=5950657 -2025-07-28 16:37:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 16:37:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 16:37:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_ocnanalecen, jobid=5950657, in state SUCCEEDED (COMPLETED), ran for 63.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:43:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 16:43:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 16:49:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 16:49:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 16:55:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 16:55:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 17:01:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 17:01:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 17:07:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 17:07:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 17:13:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 17:13:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 17:19:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 17:19:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 17:25:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 17:25:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 17:31:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 17:31:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 17:37:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 17:37:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 17:43:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 17:43:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 17:49:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 17:49:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 17:55:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 17:55:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 18:01:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 18:01:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 18:07:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 18:07:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 18:14:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 18:14:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 18:20:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 18:20:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 18:26:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 18:26:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 18:32:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 18:32:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 18:38:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 18:38:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 18:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 18:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 18:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 18:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 18:56:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 18:56:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 19:02:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 19:02:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 19:08:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 19:08:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 19:14:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state QUEUED (PENDING) -2025-07-28 19:14:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 19:20:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950445, in state SUCCEEDED (COMPLETED), ran for 101.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:20:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 19:20:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_anal -2025-07-28 19:20:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_eobs -2025-07-28 19:20:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_anal is pending at druby://130.18.14.151:36903 -2025-07-28 19:20:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of enkfgdas_eobs is pending at druby://130.18.14.151:36903 -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_anal is success, jobid=5951103 -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending enkfgdas_eobs is success, jobid=5951104 -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state QUEUED (PENDING) -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 19:26:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eobs, jobid=5951104, in state QUEUED (PENDING) -2025-07-28 19:32:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state QUEUED (PENDING) -2025-07-28 19:32:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 19:32:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eobs, jobid=5951104, in state QUEUED (PENDING) -2025-07-28 19:38:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state QUEUED (PENDING) -2025-07-28 19:38:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 19:38:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eobs, jobid=5951104, in state QUEUED (PENDING) -2025-07-28 19:44:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state QUEUED (PENDING) -2025-07-28 19:44:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state QUEUED (PENDING) -2025-07-28 19:44:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eobs, jobid=5951104, in state QUEUED (PENDING) -2025-07-28 19:50:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state QUEUED (PENDING) -2025-07-28 19:50:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950656, in state SUCCEEDED (COMPLETED), ran for 80.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:50:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eobs, jobid=5951104, in state QUEUED (PENDING) -2025-07-28 19:50:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marineanlinit -2025-07-28 19:50:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marineanlletkf -2025-07-28 19:50:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_marineanlinit is pending at druby://130.18.14.151:42671 -2025-07-28 19:50:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_marineanlletkf is pending at druby://130.18.14.151:42671 -2025-07-28 19:56:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_marineanlinit is success, jobid=5951198 -2025-07-28 19:56:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_marineanlletkf is success, jobid=5951199 -2025-07-28 19:56:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state QUEUED (PENDING) -2025-07-28 19:56:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlinit, jobid=5951198, in state QUEUED (PENDING) -2025-07-28 19:56:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlletkf, jobid=5951199, in state QUEUED (PENDING) -2025-07-28 19:56:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eobs, jobid=5951104, in state QUEUED (PENDING) -2025-07-28 20:02:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state QUEUED (PENDING) -2025-07-28 20:02:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlinit, jobid=5951198, in state QUEUED (PENDING) -2025-07-28 20:02:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlletkf, jobid=5951199, in state QUEUED (PENDING) -2025-07-28 20:02:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eobs, jobid=5951104, in state QUEUED (PENDING) -2025-07-28 20:08:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state QUEUED (PENDING) -2025-07-28 20:08:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlinit, jobid=5951198, in state QUEUED (PENDING) -2025-07-28 20:08:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlletkf, jobid=5951199, in state QUEUED (PENDING) -2025-07-28 20:08:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eobs, jobid=5951104, in state QUEUED (PENDING) -2025-07-28 20:14:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state QUEUED (PENDING) -2025-07-28 20:14:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlinit, jobid=5951198, in state QUEUED (PENDING) -2025-07-28 20:14:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlletkf, jobid=5951199, in state QUEUED (PENDING) -2025-07-28 20:14:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eobs, jobid=5951104, in state QUEUED (PENDING) -2025-07-28 20:20:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state QUEUED (PENDING) -2025-07-28 20:20:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlinit, jobid=5951198, in state QUEUED (PENDING) -2025-07-28 20:20:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlletkf, jobid=5951199, in state QUEUED (PENDING) -2025-07-28 20:20:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eobs, jobid=5951104, in state QUEUED (PENDING) -2025-07-28 20:26:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state QUEUED (PENDING) -2025-07-28 20:26:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlinit, jobid=5951198, in state SUCCEEDED (COMPLETED), ran for 33.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 20:26:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlletkf, jobid=5951199, in state QUEUED (PENDING) -2025-07-28 20:26:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eobs, jobid=5951104, in state RUNNING (RUNNING) -2025-07-28 20:26:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marineanlvar -2025-07-28 20:26:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_marineanlvar is pending at druby://130.18.14.151:34343 -2025-07-28 20:32:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_marineanlvar is success, jobid=5951341 -2025-07-28 20:32:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state QUEUED (PENDING) -2025-07-28 20:32:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951341, in state QUEUED (PENDING) -2025-07-28 20:32:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlletkf, jobid=5951199, in state QUEUED (PENDING) -2025-07-28 20:32:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eobs, jobid=5951104, in state SUCCEEDED (COMPLETED), ran for 429.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 20:32:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_ediag -2025-07-28 20:32:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of enkfgdas_ediag is pending at druby://130.18.14.151:38093 -2025-07-28 20:39:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending enkfgdas_ediag is success, jobid=5951396 -2025-07-28 20:39:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state QUEUED (PENDING) -2025-07-28 20:39:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951341, in state QUEUED (PENDING) -2025-07-28 20:39:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlletkf, jobid=5951199, in state QUEUED (PENDING) -2025-07-28 20:39:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ediag, jobid=5951396, in state SUCCEEDED (COMPLETED), ran for 143.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 20:39:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_eupd -2025-07-28 20:39:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of enkfgdas_eupd is pending at druby://130.18.14.151:39029 -2025-07-28 20:45:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending enkfgdas_eupd is success, jobid=5951426 -2025-07-28 20:45:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state QUEUED (PENDING) -2025-07-28 20:45:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951341, in state QUEUED (PENDING) -2025-07-28 20:45:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlletkf, jobid=5951199, in state QUEUED (PENDING) -2025-07-28 20:45:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951426, in state QUEUED (PENDING) -2025-07-28 20:51:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state QUEUED (PENDING) -2025-07-28 20:51:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951341, in state QUEUED (PENDING) -2025-07-28 20:51:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlletkf, jobid=5951199, in state QUEUED (PENDING) -2025-07-28 20:51:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951426, in state QUEUED (PENDING) -2025-07-28 20:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state QUEUED (PENDING) -2025-07-28 20:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951341, in state QUEUED (PENDING) -2025-07-28 20:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlletkf, jobid=5951199, in state QUEUED (PENDING) -2025-07-28 20:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951426, in state QUEUED (PENDING) -2025-07-28 21:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state QUEUED (PENDING) -2025-07-28 21:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951341, in state QUEUED (PENDING) -2025-07-28 21:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlletkf, jobid=5951199, in state QUEUED (PENDING) -2025-07-28 21:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951426, in state QUEUED (PENDING) -2025-07-28 21:09:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state QUEUED (PENDING) -2025-07-28 21:09:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951341, in state QUEUED (PENDING) -2025-07-28 21:09:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlletkf, jobid=5951199, in state QUEUED (PENDING) -2025-07-28 21:09:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951426, in state QUEUED (PENDING) -2025-07-28 21:15:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state QUEUED (PENDING) -2025-07-28 21:15:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951341, in state QUEUED (PENDING) -2025-07-28 21:15:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlletkf, jobid=5951199, in state QUEUED (PENDING) -2025-07-28 21:15:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951426, in state QUEUED (PENDING) -2025-07-28 21:21:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state RUNNING (RUNNING) -2025-07-28 21:21:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951341, in state QUEUED (PENDING) -2025-07-28 21:21:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlletkf, jobid=5951199, in state QUEUED (PENDING) -2025-07-28 21:21:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951426, in state RUNNING (RUNNING) -2025-07-28 21:27:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state RUNNING (RUNNING) -2025-07-28 21:27:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951341, in state SUCCEEDED (COMPLETED), ran for 51.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlletkf, jobid=5951199, in state SUCCEEDED (COMPLETED), ran for 115.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951426, in state SUCCEEDED (COMPLETED), ran for 153.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marineanlchkpt -2025-07-28 21:27:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_marineanlchkpt is pending at druby://130.18.14.151:43559 -2025-07-28 21:33:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_marineanlchkpt is success, jobid=5951660 -2025-07-28 21:33:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state RUNNING (RUNNING) -2025-07-28 21:33:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlchkpt, jobid=5951660, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marineanlfinal -2025-07-28 21:33:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_marineanlfinal is pending at druby://130.18.14.151:41643 -2025-07-28 21:39:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_marineanlfinal is success, jobid=5951709 -2025-07-28 21:39:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951103, in state SUCCEEDED (COMPLETED), ran for 936.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:39:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlfinal, jobid=5951709, in state QUEUED (PENDING) -2025-07-28 21:39:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_sfcanl -2025-07-28 21:39:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_analdiag -2025-07-28 21:39:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_vminmon -2025-07-28 21:39:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_sfcanl succeeded, jobid=5951770 -2025-07-28 21:39:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_analdiag succeeded, jobid=5951771 -2025-07-28 21:39:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_vminmon is pending at druby://130.18.14.151:35331 -2025-07-28 21:45:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_vminmon is success, jobid=5951772 -2025-07-28 21:45:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlfinal, jobid=5951709, in state QUEUED (PENDING) -2025-07-28 21:45:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5951770, in state QUEUED (PENDING) -2025-07-28 21:45:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951771, in state QUEUED (PENDING) -2025-07-28 21:45:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_vminmon, jobid=5951772, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:51:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlfinal, jobid=5951709, in state SUCCEEDED (COMPLETED), ran for 33.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:51:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5951770, in state QUEUED (PENDING) -2025-07-28 21:51:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951771, in state QUEUED (PENDING) -2025-07-28 21:51:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_anlstat -2025-07-28 21:51:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_anlstat is pending at druby://130.18.14.151:39335 -2025-07-28 21:57:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_anlstat is success, jobid=5951865 -2025-07-28 21:57:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5951770, in state QUEUED (PENDING) -2025-07-28 21:57:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951771, in state QUEUED (PENDING) -2025-07-28 21:57:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anlstat, jobid=5951865, in state SUCCEEDED (COMPLETED), ran for 31.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:03:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5951770, in state QUEUED (PENDING) -2025-07-28 22:03:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951771, in state QUEUED (PENDING) -2025-07-28 22:09:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5951770, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:09:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951771, in state SUCCEEDED (COMPLETED), ran for 126.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:09:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_analcalc -2025-07-28 22:09:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_fcst_seg0 -2025-07-28 22:09:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_verfozn -2025-07-28 22:09:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_verfrad -2025-07-28 22:09:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_analcalc succeeded, jobid=5951946 -2025-07-28 22:09:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_fcst_seg0 succeeded, jobid=5951947 -2025-07-28 22:09:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_verfozn succeeded, jobid=5951948 -2025-07-28 22:09:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_verfrad succeeded, jobid=5951949 -2025-07-28 22:15:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5951946, in state QUEUED (PENDING) -2025-07-28 22:15:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951947, in state QUEUED (PENDING) -2025-07-28 22:15:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfozn, jobid=5951948, in state RUNNING (RUNNING) -2025-07-28 22:15:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5951949, in state RUNNING (RUNNING) -2025-07-28 22:21:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5951946, in state QUEUED (PENDING) -2025-07-28 22:21:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951947, in state QUEUED (PENDING) -2025-07-28 22:21:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfozn, jobid=5951948, in state SUCCEEDED (COMPLETED), ran for 78.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:21:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5951949, in state RUNNING (RUNNING) -2025-07-28 22:27:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5951946, in state QUEUED (PENDING) -2025-07-28 22:27:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951947, in state QUEUED (PENDING) -2025-07-28 22:27:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5951949, in state SUCCEEDED (COMPLETED), ran for 713.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:33:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5951946, in state QUEUED (PENDING) -2025-07-28 22:33:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951947, in state QUEUED (PENDING) -2025-07-28 22:39:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5951946, in state QUEUED (PENDING) -2025-07-28 22:39:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951947, in state QUEUED (PENDING) -2025-07-28 22:45:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5951946, in state QUEUED (PENDING) -2025-07-28 22:45:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951947, in state QUEUED (PENDING) -2025-07-28 22:51:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5951946, in state QUEUED (PENDING) -2025-07-28 22:51:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951947, in state QUEUED (PENDING) -2025-07-28 22:57:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5951946, in state QUEUED (PENDING) -2025-07-28 22:57:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951947, in state QUEUED (PENDING) -2025-07-28 23:04:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5951946, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:04:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951947, in state QUEUED (PENDING) -2025-07-28 23:04:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmanlupp -2025-07-28 23:04:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_ecen000 -2025-07-28 23:04:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_ecen001 -2025-07-28 23:04:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_ecen002 -2025-07-28 23:04:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_esfc -2025-07-28 23:04:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmanlupp succeeded, jobid=5952083 -2025-07-28 23:04:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_ecen000 succeeded, jobid=5952084 -2025-07-28 23:04:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_ecen001 succeeded, jobid=5952085 -2025-07-28 23:04:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_ecen002 succeeded, jobid=5952086 -2025-07-28 23:04:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_esfc succeeded, jobid=5952087 -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952083, in state QUEUED (PENDING) -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951947, in state SUCCEEDED (COMPLETED), ran for 288.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen000, jobid=5952084, in state QUEUED (PENDING) -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen001, jobid=5952085, in state QUEUED (PENDING) -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen002, jobid=5952086, in state QUEUED (PENDING) -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_esfc, jobid=5952087, in state QUEUED (PENDING) -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f000 -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f001 -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f002 -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f003 -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f004 -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f005 -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f006 -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f007 -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f008 -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f009 -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f000 succeeded, jobid=5952103 -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f001 succeeded, jobid=5952104 -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f002 succeeded, jobid=5952105 -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f003 succeeded, jobid=5952106 -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f004 succeeded, jobid=5952107 -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f005 succeeded, jobid=5952108 -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f006 succeeded, jobid=5952109 -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f007 succeeded, jobid=5952110 -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f008 succeeded, jobid=5952111 -2025-07-28 23:10:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f009 succeeded, jobid=5952112 -2025-07-28 23:16:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952083, in state QUEUED (PENDING) -2025-07-28 23:16:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952103, in state QUEUED (PENDING) -2025-07-28 23:16:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5952104, in state QUEUED (PENDING) -2025-07-28 23:16:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5952105, in state QUEUED (PENDING) -2025-07-28 23:16:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952106, in state QUEUED (PENDING) -2025-07-28 23:16:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5952107, in state QUEUED (PENDING) -2025-07-28 23:16:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5952108, in state QUEUED (PENDING) -2025-07-28 23:16:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952109, in state QUEUED (PENDING) -2025-07-28 23:16:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5952110, in state QUEUED (PENDING) -2025-07-28 23:16:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5952111, in state QUEUED (PENDING) -2025-07-28 23:16:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952112, in state QUEUED (PENDING) -2025-07-28 23:16:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen000, jobid=5952084, in state QUEUED (PENDING) -2025-07-28 23:16:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen001, jobid=5952085, in state QUEUED (PENDING) -2025-07-28 23:16:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen002, jobid=5952086, in state QUEUED (PENDING) -2025-07-28 23:16:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_esfc, jobid=5952087, in state QUEUED (PENDING) -2025-07-28 23:22:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952083, in state QUEUED (PENDING) -2025-07-28 23:22:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952103, in state QUEUED (PENDING) -2025-07-28 23:22:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5952104, in state QUEUED (PENDING) -2025-07-28 23:22:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5952105, in state QUEUED (PENDING) -2025-07-28 23:22:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952106, in state QUEUED (PENDING) -2025-07-28 23:22:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5952107, in state QUEUED (PENDING) -2025-07-28 23:22:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5952108, in state QUEUED (PENDING) -2025-07-28 23:22:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952109, in state QUEUED (PENDING) -2025-07-28 23:22:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5952110, in state QUEUED (PENDING) -2025-07-28 23:22:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5952111, in state QUEUED (PENDING) -2025-07-28 23:22:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952112, in state QUEUED (PENDING) -2025-07-28 23:22:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen000, jobid=5952084, in state QUEUED (PENDING) -2025-07-28 23:22:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen001, jobid=5952085, in state QUEUED (PENDING) -2025-07-28 23:22:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen002, jobid=5952086, in state QUEUED (PENDING) -2025-07-28 23:22:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_esfc, jobid=5952087, in state QUEUED (PENDING) -2025-07-28 23:28:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952083, in state QUEUED (PENDING) -2025-07-28 23:28:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952103, in state QUEUED (PENDING) -2025-07-28 23:28:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5952104, in state QUEUED (PENDING) -2025-07-28 23:28:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5952105, in state QUEUED (PENDING) -2025-07-28 23:28:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952106, in state QUEUED (PENDING) -2025-07-28 23:28:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5952107, in state QUEUED (PENDING) -2025-07-28 23:28:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5952108, in state QUEUED (PENDING) -2025-07-28 23:28:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952109, in state QUEUED (PENDING) -2025-07-28 23:28:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5952110, in state QUEUED (PENDING) -2025-07-28 23:28:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5952111, in state QUEUED (PENDING) -2025-07-28 23:28:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952112, in state QUEUED (PENDING) -2025-07-28 23:28:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen000, jobid=5952084, in state QUEUED (PENDING) -2025-07-28 23:28:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen001, jobid=5952085, in state QUEUED (PENDING) -2025-07-28 23:28:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen002, jobid=5952086, in state QUEUED (PENDING) -2025-07-28 23:28:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_esfc, jobid=5952087, in state QUEUED (PENDING) -2025-07-28 23:34:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952083, in state QUEUED (PENDING) -2025-07-28 23:34:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952103, in state QUEUED (PENDING) -2025-07-28 23:34:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5952104, in state QUEUED (PENDING) -2025-07-28 23:34:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5952105, in state QUEUED (PENDING) -2025-07-28 23:34:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952106, in state QUEUED (PENDING) -2025-07-28 23:34:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5952107, in state QUEUED (PENDING) -2025-07-28 23:34:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5952108, in state QUEUED (PENDING) -2025-07-28 23:34:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952109, in state QUEUED (PENDING) -2025-07-28 23:34:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5952110, in state QUEUED (PENDING) -2025-07-28 23:34:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5952111, in state QUEUED (PENDING) -2025-07-28 23:34:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952112, in state QUEUED (PENDING) -2025-07-28 23:34:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen000, jobid=5952084, in state QUEUED (PENDING) -2025-07-28 23:34:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen001, jobid=5952085, in state QUEUED (PENDING) -2025-07-28 23:34:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen002, jobid=5952086, in state QUEUED (PENDING) -2025-07-28 23:34:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_esfc, jobid=5952087, in state QUEUED (PENDING) -2025-07-28 23:40:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952083, in state QUEUED (PENDING) -2025-07-28 23:40:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952103, in state QUEUED (PENDING) -2025-07-28 23:40:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5952104, in state QUEUED (PENDING) -2025-07-28 23:40:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5952105, in state QUEUED (PENDING) -2025-07-28 23:40:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952106, in state QUEUED (PENDING) -2025-07-28 23:40:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5952107, in state QUEUED (PENDING) -2025-07-28 23:40:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5952108, in state QUEUED (PENDING) -2025-07-28 23:40:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952109, in state QUEUED (PENDING) -2025-07-28 23:40:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5952110, in state QUEUED (PENDING) -2025-07-28 23:40:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5952111, in state QUEUED (PENDING) -2025-07-28 23:40:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952112, in state QUEUED (PENDING) -2025-07-28 23:40:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen000, jobid=5952084, in state QUEUED (PENDING) -2025-07-28 23:40:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen001, jobid=5952085, in state QUEUED (PENDING) -2025-07-28 23:40:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen002, jobid=5952086, in state QUEUED (PENDING) -2025-07-28 23:40:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_esfc, jobid=5952087, in state QUEUED (PENDING) -2025-07-28 23:46:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952083, in state QUEUED (PENDING) -2025-07-28 23:46:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952103, in state QUEUED (PENDING) -2025-07-28 23:46:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5952104, in state QUEUED (PENDING) -2025-07-28 23:46:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5952105, in state QUEUED (PENDING) -2025-07-28 23:46:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952106, in state QUEUED (PENDING) -2025-07-28 23:46:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5952107, in state QUEUED (PENDING) -2025-07-28 23:46:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5952108, in state QUEUED (PENDING) -2025-07-28 23:46:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952109, in state QUEUED (PENDING) -2025-07-28 23:46:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5952110, in state QUEUED (PENDING) -2025-07-28 23:46:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5952111, in state QUEUED (PENDING) -2025-07-28 23:46:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952112, in state QUEUED (PENDING) -2025-07-28 23:46:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen000, jobid=5952084, in state QUEUED (PENDING) -2025-07-28 23:46:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen001, jobid=5952085, in state QUEUED (PENDING) -2025-07-28 23:46:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen002, jobid=5952086, in state QUEUED (PENDING) -2025-07-28 23:46:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_esfc, jobid=5952087, in state QUEUED (PENDING) -2025-07-28 23:52:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952083, in state QUEUED (PENDING) -2025-07-28 23:52:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952103, in state QUEUED (PENDING) -2025-07-28 23:52:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5952104, in state QUEUED (PENDING) -2025-07-28 23:52:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5952105, in state QUEUED (PENDING) -2025-07-28 23:52:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952106, in state QUEUED (PENDING) -2025-07-28 23:52:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5952107, in state QUEUED (PENDING) -2025-07-28 23:52:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5952108, in state QUEUED (PENDING) -2025-07-28 23:52:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952109, in state QUEUED (PENDING) -2025-07-28 23:52:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5952110, in state QUEUED (PENDING) -2025-07-28 23:52:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5952111, in state QUEUED (PENDING) -2025-07-28 23:52:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952112, in state QUEUED (PENDING) -2025-07-28 23:52:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen000, jobid=5952084, in state QUEUED (PENDING) -2025-07-28 23:52:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen001, jobid=5952085, in state QUEUED (PENDING) -2025-07-28 23:52:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen002, jobid=5952086, in state QUEUED (PENDING) -2025-07-28 23:52:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_esfc, jobid=5952087, in state QUEUED (PENDING) -2025-07-28 23:58:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952083, in state QUEUED (PENDING) -2025-07-28 23:58:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952103, in state QUEUED (PENDING) -2025-07-28 23:58:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5952104, in state QUEUED (PENDING) -2025-07-28 23:58:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5952105, in state QUEUED (PENDING) -2025-07-28 23:58:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952106, in state QUEUED (PENDING) -2025-07-28 23:58:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5952107, in state QUEUED (PENDING) -2025-07-28 23:58:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5952108, in state QUEUED (PENDING) -2025-07-28 23:58:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952109, in state QUEUED (PENDING) -2025-07-28 23:58:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5952110, in state QUEUED (PENDING) -2025-07-28 23:58:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5952111, in state QUEUED (PENDING) -2025-07-28 23:58:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952112, in state QUEUED (PENDING) -2025-07-28 23:58:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen000, jobid=5952084, in state QUEUED (PENDING) -2025-07-28 23:58:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen001, jobid=5952085, in state QUEUED (PENDING) -2025-07-28 23:58:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen002, jobid=5952086, in state QUEUED (PENDING) -2025-07-28 23:58:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_esfc, jobid=5952087, in state QUEUED (PENDING) -2025-07-29 00:04:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952083, in state QUEUED (PENDING) -2025-07-29 00:04:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952103, in state QUEUED (PENDING) -2025-07-29 00:04:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5952104, in state QUEUED (PENDING) -2025-07-29 00:04:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5952105, in state QUEUED (PENDING) -2025-07-29 00:04:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952106, in state QUEUED (PENDING) -2025-07-29 00:04:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5952107, in state QUEUED (PENDING) -2025-07-29 00:04:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5952108, in state QUEUED (PENDING) -2025-07-29 00:04:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952109, in state QUEUED (PENDING) -2025-07-29 00:04:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5952110, in state QUEUED (PENDING) -2025-07-29 00:04:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5952111, in state QUEUED (PENDING) -2025-07-29 00:04:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952112, in state QUEUED (PENDING) -2025-07-29 00:04:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen000, jobid=5952084, in state QUEUED (PENDING) -2025-07-29 00:04:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen001, jobid=5952085, in state QUEUED (PENDING) -2025-07-29 00:04:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen002, jobid=5952086, in state QUEUED (PENDING) -2025-07-29 00:04:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_esfc, jobid=5952087, in state QUEUED (PENDING) -2025-07-29 00:10:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952083, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:10:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952103, in state RUNNING (RUNNING) -2025-07-29 00:10:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5952104, in state RUNNING (RUNNING) -2025-07-29 00:10:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5952105, in state RUNNING (RUNNING) -2025-07-29 00:10:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952106, in state QUEUED (PENDING) -2025-07-29 00:10:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5952107, in state QUEUED (PENDING) -2025-07-29 00:10:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5952108, in state QUEUED (PENDING) -2025-07-29 00:10:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952109, in state QUEUED (PENDING) -2025-07-29 00:10:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5952110, in state QUEUED (PENDING) -2025-07-29 00:10:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5952111, in state QUEUED (PENDING) -2025-07-29 00:10:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952112, in state QUEUED (PENDING) -2025-07-29 00:10:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen000, jobid=5952084, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:10:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen001, jobid=5952085, in state SUCCEEDED (COMPLETED), ran for 43.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:10:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen002, jobid=5952086, in state SUCCEEDED (COMPLETED), ran for 43.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:10:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_esfc, jobid=5952087, in state SUCCEEDED (COMPLETED), ran for 121.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:10:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmanlprod -2025-07-29 00:10:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_fcst_mem001 -2025-07-29 00:10:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_fcst_mem002 -2025-07-29 00:10:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmanlprod succeeded, jobid=5952374 -2025-07-29 00:10:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of enkfgdas_fcst_mem001 is pending at druby://130.18.14.151:39857 -2025-07-29 00:10:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of enkfgdas_fcst_mem002 is pending at druby://130.18.14.151:39857 -2025-07-29 00:16:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending enkfgdas_fcst_mem001 is success, jobid=5952375 -2025-07-29 00:16:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending enkfgdas_fcst_mem002 is success, jobid=5952376 -2025-07-29 00:16:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlprod, jobid=5952374, in state QUEUED (PENDING) -2025-07-29 00:16:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952103, in state SUCCEEDED (COMPLETED), ran for 65.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:16:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5952104, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:16:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5952105, in state SUCCEEDED (COMPLETED), ran for 64.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:16:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952106, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:16:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5952107, in state SUCCEEDED (COMPLETED), ran for 66.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:16:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5952108, in state SUCCEEDED (COMPLETED), ran for 65.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:16:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952109, in state SUCCEEDED (COMPLETED), ran for 69.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:16:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5952110, in state SUCCEEDED (COMPLETED), ran for 68.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:16:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5952111, in state SUCCEEDED (COMPLETED), ran for 65.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:16:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952112, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:16:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem001, jobid=5952375, in state QUEUED (PENDING) -2025-07-29 00:16:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem002, jobid=5952376, in state QUEUED (PENDING) -2025-07-29 00:16:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_fit2obs -2025-07-29 00:16:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_fit2obs is pending at druby://130.18.14.151:37743 -2025-07-29 00:22:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_fit2obs is success, jobid=5952417 -2025-07-29 00:22:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlprod, jobid=5952374, in state QUEUED (PENDING) -2025-07-29 00:22:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fit2obs, jobid=5952417, in state SUCCEEDED (COMPLETED), ran for 24.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:22:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem001, jobid=5952375, in state QUEUED (PENDING) -2025-07-29 00:22:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem002, jobid=5952376, in state QUEUED (PENDING) -2025-07-29 00:28:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlprod, jobid=5952374, in state QUEUED (PENDING) -2025-07-29 00:28:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem001, jobid=5952375, in state RUNNING (RUNNING) -2025-07-29 00:28:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem002, jobid=5952376, in state RUNNING (RUNNING) -2025-07-29 00:34:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlprod, jobid=5952374, in state QUEUED (PENDING) -2025-07-29 00:34:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem001, jobid=5952375, in state SUCCEEDED (COMPLETED), ran for 292.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:34:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem002, jobid=5952376, in state SUCCEEDED (COMPLETED), ran for 295.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:34:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_echgres -2025-07-29 00:34:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos000 -2025-07-29 00:34:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos001 -2025-07-29 00:34:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos002 -2025-07-29 00:34:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos003 -2025-07-29 00:34:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos004 -2025-07-29 00:34:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos005 -2025-07-29 00:34:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos006 -2025-07-29 00:34:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_echgres succeeded, jobid=5952526 -2025-07-29 00:34:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos000 succeeded, jobid=5952527 -2025-07-29 00:34:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos001 succeeded, jobid=5952528 -2025-07-29 00:34:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos002 succeeded, jobid=5952529 -2025-07-29 00:34:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos003 succeeded, jobid=5952530 -2025-07-29 00:34:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos004 succeeded, jobid=5952531 -2025-07-29 00:34:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos005 succeeded, jobid=5952532 -2025-07-29 00:34:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos006 succeeded, jobid=5952533 -2025-07-29 00:40:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlprod, jobid=5952374, in state SUCCEEDED (COMPLETED), ran for 63.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:40:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5952526, in state SUCCEEDED (COMPLETED), ran for 24.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:40:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5952527, in state SUCCEEDED (COMPLETED), ran for 37.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:40:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5952528, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:40:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5952529, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:40:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5952530, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:40:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5952531, in state QUEUED (PENDING) -2025-07-29 00:40:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5952532, in state QUEUED (PENDING) -2025-07-29 00:40:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5952533, in state QUEUED (PENDING) -2025-07-29 00:40:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_arch_vrfy -2025-07-29 00:40:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_arch_vrfy is pending at druby://130.18.14.151:45427 -2025-07-29 00:46:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_arch_vrfy is success, jobid=5952568 -2025-07-29 00:46:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_arch_vrfy, jobid=5952568, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:46:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5952531, in state SUCCEEDED (COMPLETED), ran for 62.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:46:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5952532, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:46:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5952533, in state SUCCEEDED (COMPLETED), ran for 37.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:46:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_cleanup -2025-07-29 00:46:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_earc_vrfy -2025-07-29 00:46:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_cleanup is pending at druby://130.18.14.151:42953 -2025-07-29 00:46:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of enkfgdas_earc_vrfy is pending at druby://130.18.14.151:42953 -2025-07-29 00:52:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_cleanup is success, jobid=5952623 -2025-07-29 00:52:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending enkfgdas_earc_vrfy is success, jobid=5952624 -2025-07-29 00:52:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_cleanup, jobid=5952623, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:52:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_earc_vrfy, jobid=5952624, in state SUCCEEDED (COMPLETED), ran for 15.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:52:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_cleanup -2025-07-29 00:52:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of enkfgdas_cleanup is pending at druby://130.18.14.151:44371 -2025-07-29 00:58:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending enkfgdas_cleanup is success, jobid=5952652 -2025-07-29 00:58:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_cleanup, jobid=5952652, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:58:51 -0500 :: hercules-login-1.hpc.msstate.edu :: This cycle is complete: Success diff --git a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96C48_hybatmDA_logs_2021122018.log b/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96C48_hybatmDA_logs_2021122018.log deleted file mode 100644 index 74721903..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96C48_hybatmDA_logs_2021122018.log +++ /dev/null @@ -1,249 +0,0 @@ -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_stage_ic -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_stage_ic -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_stage_ic succeeded, jobid=5950269 -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_stage_ic succeeded, jobid=5950270 -2025-07-28 14:48:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_stage_ic, jobid=5950269, in state RUNNING (RUNNING) -2025-07-28 14:48:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_stage_ic, jobid=5950270, in state RUNNING (RUNNING) -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_stage_ic, jobid=5950269, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_stage_ic, jobid=5950270, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_fcst_seg0 -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_fcst_mem001 -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_fcst_mem002 -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_fcst_seg0 succeeded, jobid=5950289 -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_fcst_mem001 succeeded, jobid=5950290 -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_fcst_mem002 succeeded, jobid=5950291 -2025-07-28 15:00:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5950289, in state QUEUED (PENDING) -2025-07-28 15:00:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem001, jobid=5950290, in state QUEUED (PENDING) -2025-07-28 15:00:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem002, jobid=5950291, in state QUEUED (PENDING) -2025-07-28 15:06:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5950289, in state RUNNING (RUNNING) -2025-07-28 15:06:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem001, jobid=5950290, in state SUCCEEDED (COMPLETED), ran for 212.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:06:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem002, jobid=5950291, in state SUCCEEDED (COMPLETED), ran for 212.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:06:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos000 -2025-07-28 15:06:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos001 -2025-07-28 15:06:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos002 -2025-07-28 15:06:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos003 -2025-07-28 15:06:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos004 -2025-07-28 15:06:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos005 -2025-07-28 15:06:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos006 -2025-07-28 15:06:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos000 succeeded, jobid=5950327 -2025-07-28 15:06:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos001 succeeded, jobid=5950328 -2025-07-28 15:06:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos002 succeeded, jobid=5950329 -2025-07-28 15:06:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos003 succeeded, jobid=5950330 -2025-07-28 15:06:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos004 succeeded, jobid=5950331 -2025-07-28 15:06:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos005 succeeded, jobid=5950332 -2025-07-28 15:06:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos006 succeeded, jobid=5950333 -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5950289, in state SUCCEEDED (COMPLETED), ran for 348.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950327, in state SUCCEEDED (COMPLETED), ran for 32.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950328, in state SUCCEEDED (COMPLETED), ran for 32.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950329, in state QUEUED (PENDING) -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950330, in state QUEUED (PENDING) -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950331, in state QUEUED (PENDING) -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950332, in state QUEUED (PENDING) -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950333, in state QUEUED (PENDING) -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f000 -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f001 -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f002 -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f003 -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f004 -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f005 -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f006 -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f007 -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f008 -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f009 -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_echgres -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f000 succeeded, jobid=5950372 -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f001 succeeded, jobid=5950373 -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f002 succeeded, jobid=5950374 -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f003 succeeded, jobid=5950375 -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f004 succeeded, jobid=5950376 -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f005 succeeded, jobid=5950377 -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f006 succeeded, jobid=5950378 -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f007 succeeded, jobid=5950379 -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f008 succeeded, jobid=5950380 -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f009 succeeded, jobid=5950381 -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_echgres succeeded, jobid=5950382 -2025-07-28 15:18:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950372, in state QUEUED (PENDING) -2025-07-28 15:18:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950373, in state QUEUED (PENDING) -2025-07-28 15:18:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950374, in state QUEUED (PENDING) -2025-07-28 15:18:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950375, in state QUEUED (PENDING) -2025-07-28 15:18:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950376, in state QUEUED (PENDING) -2025-07-28 15:18:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950377, in state QUEUED (PENDING) -2025-07-28 15:18:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950378, in state QUEUED (PENDING) -2025-07-28 15:18:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950379, in state QUEUED (PENDING) -2025-07-28 15:18:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950380, in state QUEUED (PENDING) -2025-07-28 15:18:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 15:18:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950382, in state QUEUED (PENDING) -2025-07-28 15:18:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950329, in state SUCCEEDED (COMPLETED), ran for 37.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:18:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950330, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:18:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950331, in state SUCCEEDED (COMPLETED), ran for 32.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:18:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950332, in state SUCCEEDED (COMPLETED), ran for 32.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:18:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950333, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950372, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950373, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950374, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950375, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950376, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950377, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950378, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950379, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950380, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950382, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950333, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:30:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950372, in state QUEUED (PENDING) -2025-07-28 15:30:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950373, in state QUEUED (PENDING) -2025-07-28 15:30:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950374, in state QUEUED (PENDING) -2025-07-28 15:30:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950375, in state QUEUED (PENDING) -2025-07-28 15:30:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950376, in state QUEUED (PENDING) -2025-07-28 15:30:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950377, in state QUEUED (PENDING) -2025-07-28 15:30:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950378, in state QUEUED (PENDING) -2025-07-28 15:30:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950379, in state QUEUED (PENDING) -2025-07-28 15:30:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950380, in state QUEUED (PENDING) -2025-07-28 15:30:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 15:30:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950382, in state QUEUED (PENDING) -2025-07-28 15:36:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950372, in state QUEUED (PENDING) -2025-07-28 15:36:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950373, in state QUEUED (PENDING) -2025-07-28 15:36:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950374, in state QUEUED (PENDING) -2025-07-28 15:36:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950375, in state QUEUED (PENDING) -2025-07-28 15:36:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950376, in state QUEUED (PENDING) -2025-07-28 15:36:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950377, in state QUEUED (PENDING) -2025-07-28 15:36:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950378, in state QUEUED (PENDING) -2025-07-28 15:36:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950379, in state QUEUED (PENDING) -2025-07-28 15:36:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950380, in state QUEUED (PENDING) -2025-07-28 15:36:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 15:36:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950382, in state QUEUED (PENDING) -2025-07-28 15:42:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950372, in state QUEUED (PENDING) -2025-07-28 15:42:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950373, in state QUEUED (PENDING) -2025-07-28 15:42:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950374, in state QUEUED (PENDING) -2025-07-28 15:42:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950375, in state QUEUED (PENDING) -2025-07-28 15:42:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950376, in state QUEUED (PENDING) -2025-07-28 15:42:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950377, in state QUEUED (PENDING) -2025-07-28 15:42:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950378, in state QUEUED (PENDING) -2025-07-28 15:42:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950379, in state QUEUED (PENDING) -2025-07-28 15:42:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950380, in state QUEUED (PENDING) -2025-07-28 15:42:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 15:42:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950382, in state QUEUED (PENDING) -2025-07-28 15:49:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950372, in state QUEUED (PENDING) -2025-07-28 15:49:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950373, in state QUEUED (PENDING) -2025-07-28 15:49:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950374, in state QUEUED (PENDING) -2025-07-28 15:49:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950375, in state QUEUED (PENDING) -2025-07-28 15:49:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950376, in state QUEUED (PENDING) -2025-07-28 15:49:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950377, in state QUEUED (PENDING) -2025-07-28 15:49:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950378, in state QUEUED (PENDING) -2025-07-28 15:49:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950379, in state QUEUED (PENDING) -2025-07-28 15:49:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950380, in state QUEUED (PENDING) -2025-07-28 15:49:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 15:49:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950382, in state QUEUED (PENDING) -2025-07-28 15:55:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950372, in state QUEUED (PENDING) -2025-07-28 15:55:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950373, in state QUEUED (PENDING) -2025-07-28 15:55:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950374, in state QUEUED (PENDING) -2025-07-28 15:55:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950375, in state QUEUED (PENDING) -2025-07-28 15:55:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950376, in state QUEUED (PENDING) -2025-07-28 15:55:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950377, in state QUEUED (PENDING) -2025-07-28 15:55:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950378, in state QUEUED (PENDING) -2025-07-28 15:55:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950379, in state QUEUED (PENDING) -2025-07-28 15:55:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950380, in state QUEUED (PENDING) -2025-07-28 15:55:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 15:55:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950382, in state QUEUED (PENDING) -2025-07-28 16:01:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950372, in state QUEUED (PENDING) -2025-07-28 16:01:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950373, in state QUEUED (PENDING) -2025-07-28 16:01:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950374, in state QUEUED (PENDING) -2025-07-28 16:01:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950375, in state QUEUED (PENDING) -2025-07-28 16:01:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950376, in state QUEUED (PENDING) -2025-07-28 16:01:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950377, in state QUEUED (PENDING) -2025-07-28 16:01:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950378, in state QUEUED (PENDING) -2025-07-28 16:01:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950379, in state QUEUED (PENDING) -2025-07-28 16:01:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950380, in state QUEUED (PENDING) -2025-07-28 16:01:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 16:01:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950382, in state QUEUED (PENDING) -2025-07-28 16:07:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950372, in state QUEUED (PENDING) -2025-07-28 16:07:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950373, in state QUEUED (PENDING) -2025-07-28 16:07:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950374, in state QUEUED (PENDING) -2025-07-28 16:07:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950375, in state QUEUED (PENDING) -2025-07-28 16:07:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950376, in state QUEUED (PENDING) -2025-07-28 16:07:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950377, in state QUEUED (PENDING) -2025-07-28 16:07:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950378, in state QUEUED (PENDING) -2025-07-28 16:07:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950379, in state QUEUED (PENDING) -2025-07-28 16:07:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950380, in state QUEUED (PENDING) -2025-07-28 16:07:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 16:07:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950382, in state QUEUED (PENDING) -2025-07-28 16:13:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950372, in state QUEUED (PENDING) -2025-07-28 16:13:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950373, in state QUEUED (PENDING) -2025-07-28 16:13:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950374, in state QUEUED (PENDING) -2025-07-28 16:13:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950375, in state QUEUED (PENDING) -2025-07-28 16:13:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950376, in state QUEUED (PENDING) -2025-07-28 16:13:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950377, in state QUEUED (PENDING) -2025-07-28 16:13:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950378, in state QUEUED (PENDING) -2025-07-28 16:13:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950379, in state QUEUED (PENDING) -2025-07-28 16:13:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950380, in state QUEUED (PENDING) -2025-07-28 16:13:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 16:13:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950382, in state QUEUED (PENDING) -2025-07-28 16:19:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950372, in state QUEUED (PENDING) -2025-07-28 16:19:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950373, in state QUEUED (PENDING) -2025-07-28 16:19:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950374, in state QUEUED (PENDING) -2025-07-28 16:19:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950375, in state QUEUED (PENDING) -2025-07-28 16:19:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950376, in state QUEUED (PENDING) -2025-07-28 16:19:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950377, in state QUEUED (PENDING) -2025-07-28 16:19:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950378, in state QUEUED (PENDING) -2025-07-28 16:19:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950379, in state QUEUED (PENDING) -2025-07-28 16:19:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950380, in state QUEUED (PENDING) -2025-07-28 16:19:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 16:19:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950382, in state QUEUED (PENDING) -2025-07-28 16:25:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950372, in state QUEUED (PENDING) -2025-07-28 16:25:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950373, in state QUEUED (PENDING) -2025-07-28 16:25:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950374, in state QUEUED (PENDING) -2025-07-28 16:25:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950375, in state QUEUED (PENDING) -2025-07-28 16:25:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950376, in state QUEUED (PENDING) -2025-07-28 16:25:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950377, in state QUEUED (PENDING) -2025-07-28 16:25:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950378, in state QUEUED (PENDING) -2025-07-28 16:25:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950379, in state QUEUED (PENDING) -2025-07-28 16:25:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950380, in state QUEUED (PENDING) -2025-07-28 16:25:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 16:25:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950382, in state QUEUED (PENDING) -2025-07-28 16:31:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950372, in state SUCCEEDED (COMPLETED), ran for 66.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:31:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950373, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:31:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950374, in state SUCCEEDED (COMPLETED), ran for 68.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:31:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950375, in state SUCCEEDED (COMPLETED), ran for 69.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:31:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950376, in state SUCCEEDED (COMPLETED), ran for 69.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:31:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950377, in state QUEUED (PENDING) -2025-07-28 16:31:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950378, in state QUEUED (PENDING) -2025-07-28 16:31:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950379, in state QUEUED (PENDING) -2025-07-28 16:31:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950380, in state QUEUED (PENDING) -2025-07-28 16:31:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 16:31:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950382, in state QUEUED (PENDING) -2025-07-28 16:37:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950377, in state QUEUED (PENDING) -2025-07-28 16:37:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950378, in state QUEUED (PENDING) -2025-07-28 16:37:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950379, in state QUEUED (PENDING) -2025-07-28 16:37:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950380, in state QUEUED (PENDING) -2025-07-28 16:37:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 16:37:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950382, in state SUCCEEDED (COMPLETED), ran for 31.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:43:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950377, in state RUNNING (RUNNING) -2025-07-28 16:43:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950378, in state QUEUED (PENDING) -2025-07-28 16:43:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950379, in state QUEUED (PENDING) -2025-07-28 16:43:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950380, in state QUEUED (PENDING) -2025-07-28 16:43:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 16:49:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950377, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:49:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950378, in state QUEUED (PENDING) -2025-07-28 16:49:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950379, in state QUEUED (PENDING) -2025-07-28 16:49:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950380, in state QUEUED (PENDING) -2025-07-28 16:49:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 16:55:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950378, in state QUEUED (PENDING) -2025-07-28 16:55:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950379, in state QUEUED (PENDING) -2025-07-28 16:55:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950380, in state QUEUED (PENDING) -2025-07-28 16:55:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 17:01:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950378, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 17:01:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950379, in state QUEUED (PENDING) -2025-07-28 17:01:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950380, in state QUEUED (PENDING) -2025-07-28 17:01:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 17:07:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950379, in state SUCCEEDED (COMPLETED), ran for 69.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 17:07:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950380, in state SUCCEEDED (COMPLETED), ran for 68.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 17:07:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 17:13:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 17:19:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 17:25:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 17:31:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 17:37:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 17:43:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 17:49:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state QUEUED (PENDING) -2025-07-28 17:55:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950381, in state SUCCEEDED (COMPLETED), ran for 69.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 17:55:57 -0500 :: hercules-login-1.hpc.msstate.edu :: This cycle is complete: Success diff --git a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96C48_hybatmDA_logs_2021122100.log b/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96C48_hybatmDA_logs_2021122100.log deleted file mode 100644 index 5995050c..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96C48_hybatmDA_logs_2021122100.log +++ /dev/null @@ -1,556 +0,0 @@ -2025-07-28 17:55:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_prep -2025-07-28 17:55:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_prep -2025-07-28 17:55:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_prep succeeded, jobid=5950888 -2025-07-28 17:55:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_prep is pending at druby://130.18.14.151:32961 -2025-07-28 18:02:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_prep is success, jobid=5950889 -2025-07-28 18:02:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 18:02:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 18:08:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 18:08:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 18:14:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 18:14:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 18:20:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 18:20:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 18:26:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 18:26:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 18:32:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 18:32:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 18:38:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 18:38:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 18:44:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 18:44:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 18:50:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 18:50:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 18:56:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 18:56:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 19:02:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 19:02:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 19:08:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 19:08:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 19:14:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 19:14:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 19:20:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 19:20:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 19:26:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 19:26:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 19:32:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 19:32:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 19:38:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 19:38:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 19:44:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 19:44:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 19:50:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 19:50:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 19:56:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 19:56:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 20:03:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 20:03:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 20:09:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 20:09:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 20:15:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 20:15:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 20:21:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 20:21:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 20:27:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 20:27:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 20:33:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 20:33:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 20:39:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 20:39:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 20:45:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 20:45:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 20:51:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 20:51:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 20:57:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 20:57:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 21:03:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 21:03:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 21:09:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 21:09:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 21:15:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state QUEUED (PENDING) -2025-07-28 21:15:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state QUEUED (PENDING) -2025-07-28 21:21:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state RUNNING (RUNNING) -2025-07-28 21:21:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950889, in state SUCCEEDED (COMPLETED), ran for 139.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_anal -2025-07-28 21:21:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_eobs -2025-07-28 21:21:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_anal is pending at druby://130.18.14.151:37669 -2025-07-28 21:21:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of enkfgdas_eobs is pending at druby://130.18.14.151:37669 -2025-07-28 21:27:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_anal is success, jobid=5951610 -2025-07-28 21:27:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending enkfgdas_eobs is success, jobid=5951611 -2025-07-28 21:27:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950888, in state SUCCEEDED (COMPLETED), ran for 227.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951610, in state QUEUED (PENDING) -2025-07-28 21:27:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eobs, jobid=5951611, in state RUNNING (RUNNING) -2025-07-28 21:27:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_anal -2025-07-28 21:27:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_anal is pending at druby://130.18.14.151:42331 -2025-07-28 21:33:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_anal is success, jobid=5951667 -2025-07-28 21:33:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951667, in state QUEUED (PENDING) -2025-07-28 21:33:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951610, in state QUEUED (PENDING) -2025-07-28 21:33:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eobs, jobid=5951611, in state SUCCEEDED (COMPLETED), ran for 412.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_ediag -2025-07-28 21:33:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of enkfgdas_ediag is pending at druby://130.18.14.151:37809 -2025-07-28 21:39:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending enkfgdas_ediag is success, jobid=5951714 -2025-07-28 21:39:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951667, in state QUEUED (PENDING) -2025-07-28 21:39:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951610, in state RUNNING (RUNNING) -2025-07-28 21:39:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ediag, jobid=5951714, in state QUEUED (PENDING) -2025-07-28 21:45:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951667, in state QUEUED (PENDING) -2025-07-28 21:45:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951610, in state RUNNING (RUNNING) -2025-07-28 21:45:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ediag, jobid=5951714, in state QUEUED (PENDING) -2025-07-28 21:51:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951667, in state RUNNING (RUNNING) -2025-07-28 21:51:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951610, in state SUCCEEDED (COMPLETED), ran for 712.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:51:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ediag, jobid=5951714, in state SUCCEEDED (COMPLETED), ran for 143.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:51:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_analdiag -2025-07-28 21:51:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_vminmon -2025-07-28 21:51:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_eupd -2025-07-28 21:51:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_analdiag succeeded, jobid=5951878 -2025-07-28 21:51:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_vminmon succeeded, jobid=5951879 -2025-07-28 21:51:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_eupd succeeded, jobid=5951880 -2025-07-28 21:57:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951667, in state SUCCEEDED (COMPLETED), ran for 380.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:57:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951878, in state QUEUED (PENDING) -2025-07-28 21:57:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_vminmon, jobid=5951879, in state SUCCEEDED (COMPLETED), ran for 16.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:57:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951880, in state QUEUED (PENDING) -2025-07-28 21:57:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_sfcanl -2025-07-28 21:57:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_vminmon -2025-07-28 21:57:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_sfcanl succeeded, jobid=5951901 -2025-07-28 21:57:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_vminmon is pending at druby://130.18.14.151:40229 -2025-07-28 22:03:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_vminmon is success, jobid=5951902 -2025-07-28 22:03:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_sfcanl, jobid=5951901, in state QUEUED (PENDING) -2025-07-28 22:03:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_vminmon, jobid=5951902, in state SUCCEEDED (COMPLETED), ran for 16.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:03:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951878, in state QUEUED (PENDING) -2025-07-28 22:03:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951880, in state QUEUED (PENDING) -2025-07-28 22:10:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_sfcanl, jobid=5951901, in state QUEUED (PENDING) -2025-07-28 22:10:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951878, in state RUNNING (RUNNING) -2025-07-28 22:10:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951880, in state QUEUED (PENDING) -2025-07-28 22:16:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_sfcanl, jobid=5951901, in state QUEUED (PENDING) -2025-07-28 22:16:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951878, in state SUCCEEDED (COMPLETED), ran for 125.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:16:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951880, in state QUEUED (PENDING) -2025-07-28 22:16:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_verfozn -2025-07-28 22:16:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_verfrad -2025-07-28 22:16:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_verfozn is pending at druby://130.18.14.151:41833 -2025-07-28 22:16:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_verfrad is pending at druby://130.18.14.151:41833 -2025-07-28 22:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_verfozn is success, jobid=5951983 -2025-07-28 22:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_verfrad is success, jobid=5951984 -2025-07-28 22:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_sfcanl, jobid=5951901, in state QUEUED (PENDING) -2025-07-28 22:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfozn, jobid=5951983, in state SUCCEEDED (COMPLETED), ran for 71.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5951984, in state RUNNING (RUNNING) -2025-07-28 22:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951880, in state SUCCEEDED (COMPLETED), ran for 148.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_sfcanl -2025-07-28 22:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_sfcanl is pending at druby://130.18.14.151:38217 -2025-07-28 22:28:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_sfcanl is success, jobid=5952005 -2025-07-28 22:28:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_sfcanl, jobid=5951901, in state QUEUED (PENDING) -2025-07-28 22:28:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952005, in state QUEUED (PENDING) -2025-07-28 22:28:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5951984, in state RUNNING (RUNNING) -2025-07-28 22:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_sfcanl, jobid=5951901, in state QUEUED (PENDING) -2025-07-28 22:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952005, in state QUEUED (PENDING) -2025-07-28 22:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5951984, in state SUCCEEDED (COMPLETED), ran for 595.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:40:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_sfcanl, jobid=5951901, in state SUCCEEDED (COMPLETED), ran for 46.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:40:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952005, in state QUEUED (PENDING) -2025-07-28 22:40:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_analcalc -2025-07-28 22:40:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_fcst_seg0 -2025-07-28 22:40:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_analcalc succeeded, jobid=5952027 -2025-07-28 22:40:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_fcst_seg0 succeeded, jobid=5952028 -2025-07-28 22:46:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5952027, in state QUEUED (PENDING) -2025-07-28 22:46:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state QUEUED (PENDING) -2025-07-28 22:46:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952005, in state QUEUED (PENDING) -2025-07-28 22:52:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5952027, in state QUEUED (PENDING) -2025-07-28 22:52:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state QUEUED (PENDING) -2025-07-28 22:52:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952005, in state QUEUED (PENDING) -2025-07-28 22:58:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5952027, in state QUEUED (PENDING) -2025-07-28 22:58:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state QUEUED (PENDING) -2025-07-28 22:58:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952005, in state QUEUED (PENDING) -2025-07-28 23:04:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5952027, in state QUEUED (PENDING) -2025-07-28 23:04:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state QUEUED (PENDING) -2025-07-28 23:04:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952005, in state QUEUED (PENDING) -2025-07-28 23:10:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5952027, in state QUEUED (PENDING) -2025-07-28 23:10:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state QUEUED (PENDING) -2025-07-28 23:10:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952005, in state QUEUED (PENDING) -2025-07-28 23:16:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5952027, in state QUEUED (PENDING) -2025-07-28 23:16:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state QUEUED (PENDING) -2025-07-28 23:16:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952005, in state QUEUED (PENDING) -2025-07-28 23:22:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5952027, in state QUEUED (PENDING) -2025-07-28 23:22:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state QUEUED (PENDING) -2025-07-28 23:22:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952005, in state QUEUED (PENDING) -2025-07-28 23:28:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5952027, in state QUEUED (PENDING) -2025-07-28 23:28:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state QUEUED (PENDING) -2025-07-28 23:28:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952005, in state QUEUED (PENDING) -2025-07-28 23:34:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5952027, in state QUEUED (PENDING) -2025-07-28 23:34:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state QUEUED (PENDING) -2025-07-28 23:34:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952005, in state QUEUED (PENDING) -2025-07-28 23:40:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5952027, in state QUEUED (PENDING) -2025-07-28 23:40:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state QUEUED (PENDING) -2025-07-28 23:40:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952005, in state QUEUED (PENDING) -2025-07-28 23:46:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5952027, in state QUEUED (PENDING) -2025-07-28 23:46:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state QUEUED (PENDING) -2025-07-28 23:46:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952005, in state QUEUED (PENDING) -2025-07-28 23:52:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5952027, in state QUEUED (PENDING) -2025-07-28 23:52:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state QUEUED (PENDING) -2025-07-28 23:52:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952005, in state QUEUED (PENDING) -2025-07-28 23:58:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5952027, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:58:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state RUNNING (RUNNING) -2025-07-28 23:58:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952005, in state SUCCEEDED (COMPLETED), ran for 68.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:58:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmanlupp -2025-07-28 23:58:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_analcalc -2025-07-28 23:58:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_fcst_seg0 -2025-07-28 23:58:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmanlupp succeeded, jobid=5952284 -2025-07-28 23:58:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_analcalc succeeded, jobid=5952285 -2025-07-28 23:58:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_fcst_seg0 succeeded, jobid=5952286 -2025-07-29 00:05:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952284, in state QUEUED (PENDING) -2025-07-29 00:05:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state RUNNING (RUNNING) -2025-07-29 00:05:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952285, in state QUEUED (PENDING) -2025-07-29 00:05:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952286, in state QUEUED (PENDING) -2025-07-29 00:05:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f000-f002 -2025-07-29 00:05:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f003-f005 -2025-07-29 00:05:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f006-f008 -2025-07-29 00:05:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f009-f011 -2025-07-29 00:05:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f000-f002 succeeded, jobid=5952324 -2025-07-29 00:05:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f003-f005 succeeded, jobid=5952325 -2025-07-29 00:05:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f006-f008 succeeded, jobid=5952326 -2025-07-29 00:05:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f009-f011 succeeded, jobid=5952327 -2025-07-29 00:11:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952284, in state QUEUED (PENDING) -2025-07-29 00:11:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state RUNNING (RUNNING) -2025-07-29 00:11:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952324, in state QUEUED (PENDING) -2025-07-29 00:11:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952325, in state QUEUED (PENDING) -2025-07-29 00:11:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952326, in state QUEUED (PENDING) -2025-07-29 00:11:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952327, in state QUEUED (PENDING) -2025-07-29 00:11:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952285, in state QUEUED (PENDING) -2025-07-29 00:11:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952286, in state QUEUED (PENDING) -2025-07-29 00:11:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f012-f014 -2025-07-29 00:11:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f015-f017 -2025-07-29 00:11:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f018-f020 -2025-07-29 00:11:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f021-f023 -2025-07-29 00:11:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f024-f026 -2025-07-29 00:11:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f012-f014 succeeded, jobid=5952379 -2025-07-29 00:11:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f015-f017 succeeded, jobid=5952380 -2025-07-29 00:11:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f018-f020 succeeded, jobid=5952381 -2025-07-29 00:11:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f021-f023 succeeded, jobid=5952382 -2025-07-29 00:11:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f024-f026 succeeded, jobid=5952383 -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952284, in state QUEUED (PENDING) -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state RUNNING (RUNNING) -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952324, in state QUEUED (PENDING) -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952325, in state QUEUED (PENDING) -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952326, in state QUEUED (PENDING) -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952327, in state QUEUED (PENDING) -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952379, in state QUEUED (PENDING) -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952380, in state QUEUED (PENDING) -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952381, in state QUEUED (PENDING) -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952382, in state QUEUED (PENDING) -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952383, in state QUEUED (PENDING) -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952285, in state QUEUED (PENDING) -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952286, in state QUEUED (PENDING) -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f027-f029 -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f030-f032 -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f033-f035 -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f036-f038 -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f039-f041 -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f027-f029 succeeded, jobid=5952418 -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f030-f032 succeeded, jobid=5952419 -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f033-f035 succeeded, jobid=5952420 -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f036-f038 succeeded, jobid=5952421 -2025-07-29 00:17:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f039-f041 succeeded, jobid=5952422 -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952284, in state QUEUED (PENDING) -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state RUNNING (RUNNING) -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952324, in state QUEUED (PENDING) -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952325, in state QUEUED (PENDING) -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952326, in state QUEUED (PENDING) -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952327, in state QUEUED (PENDING) -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952379, in state QUEUED (PENDING) -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952380, in state QUEUED (PENDING) -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952381, in state QUEUED (PENDING) -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952382, in state QUEUED (PENDING) -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952383, in state QUEUED (PENDING) -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952418, in state QUEUED (PENDING) -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952419, in state QUEUED (PENDING) -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952420, in state QUEUED (PENDING) -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952421, in state QUEUED (PENDING) -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952422, in state QUEUED (PENDING) -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952285, in state QUEUED (PENDING) -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952286, in state QUEUED (PENDING) -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f042-f044 -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f045-f047 -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f048-f054 -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f042-f044 succeeded, jobid=5952440 -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f045-f047 succeeded, jobid=5952441 -2025-07-29 00:23:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f048-f054 succeeded, jobid=5952442 -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952284, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state RUNNING (RUNNING) -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952324, in state QUEUED (PENDING) -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952325, in state QUEUED (PENDING) -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952326, in state QUEUED (PENDING) -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952327, in state QUEUED (PENDING) -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952379, in state QUEUED (PENDING) -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952380, in state QUEUED (PENDING) -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952381, in state QUEUED (PENDING) -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952382, in state QUEUED (PENDING) -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952383, in state QUEUED (PENDING) -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952418, in state QUEUED (PENDING) -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952419, in state QUEUED (PENDING) -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952420, in state QUEUED (PENDING) -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952421, in state QUEUED (PENDING) -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952422, in state QUEUED (PENDING) -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952440, in state QUEUED (PENDING) -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952441, in state QUEUED (PENDING) -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952442, in state QUEUED (PENDING) -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952285, in state SUCCEEDED (COMPLETED), ran for 61.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952286, in state RUNNING (RUNNING) -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmanlprod -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f057-f063 -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmanlupp -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_ecen000 -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_ecen001 -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_ecen002 -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit enkfgdas_esfc, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmanlprod succeeded, jobid=5952466 -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f057-f063 succeeded, jobid=5952467 -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmanlupp succeeded, jobid=5952468 -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_ecen000 succeeded, jobid=5952469 -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_ecen001 succeeded, jobid=5952470 -2025-07-29 00:29:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_ecen002 succeeded, jobid=5952471 -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlprod, jobid=5952466, in state QUEUED (PENDING) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state RUNNING (RUNNING) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952324, in state SUCCEEDED (COMPLETED), ran for 273.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952325, in state SUCCEEDED (COMPLETED), ran for 278.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952326, in state SUCCEEDED (COMPLETED), ran for 276.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952327, in state SUCCEEDED (COMPLETED), ran for 273.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952379, in state RUNNING (RUNNING) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952380, in state RUNNING (RUNNING) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952381, in state RUNNING (RUNNING) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952382, in state RUNNING (RUNNING) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952383, in state RUNNING (RUNNING) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952418, in state RUNNING (RUNNING) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952419, in state RUNNING (RUNNING) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952420, in state RUNNING (RUNNING) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952421, in state RUNNING (RUNNING) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952422, in state RUNNING (RUNNING) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952440, in state QUEUED (PENDING) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952441, in state QUEUED (PENDING) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952442, in state QUEUED (PENDING) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952467, in state QUEUED (PENDING) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952468, in state SUCCEEDED (COMPLETED), ran for 44.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952286, in state SUCCEEDED (COMPLETED), ran for 456.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen000, jobid=5952469, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen001, jobid=5952470, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen002, jobid=5952471, in state RUNNING (RUNNING) -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f066-f072 -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f075-f081 -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmanlprod -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f000 -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f001 -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f002 -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f003 -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f004 -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f005, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f006, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f007, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f008, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f009, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit enkfgdas_esfc, because maximum global task throttle of 25 will be violated. -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f066-f072 succeeded, jobid=5952545 -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f075-f081 succeeded, jobid=5952546 -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmanlprod succeeded, jobid=5952547 -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f000 succeeded, jobid=5952548 -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f001 succeeded, jobid=5952549 -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f002 succeeded, jobid=5952550 -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f003 succeeded, jobid=5952551 -2025-07-29 00:35:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f004 succeeded, jobid=5952552 -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlprod, jobid=5952466, in state RUNNING (RUNNING) -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state RUNNING (RUNNING) -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952379, in state SUCCEEDED (COMPLETED), ran for 274.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952380, in state SUCCEEDED (COMPLETED), ran for 271.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952381, in state SUCCEEDED (COMPLETED), ran for 277.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952382, in state SUCCEEDED (COMPLETED), ran for 269.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952383, in state SUCCEEDED (COMPLETED), ran for 276.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952418, in state SUCCEEDED (COMPLETED), ran for 274.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952419, in state SUCCEEDED (COMPLETED), ran for 270.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952420, in state SUCCEEDED (COMPLETED), ran for 283.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952421, in state SUCCEEDED (COMPLETED), ran for 285.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952422, in state SUCCEEDED (COMPLETED), ran for 271.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952440, in state RUNNING (RUNNING) -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952441, in state RUNNING (RUNNING) -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952442, in state RUNNING (RUNNING) -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952467, in state RUNNING (RUNNING) -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952545, in state QUEUED (PENDING) -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952546, in state QUEUED (PENDING) -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlprod, jobid=5952547, in state QUEUED (PENDING) -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952548, in state QUEUED (PENDING) -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5952549, in state QUEUED (PENDING) -2025-07-29 00:41:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5952550, in state QUEUED (PENDING) -2025-07-29 00:41:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952551, in state QUEUED (PENDING) -2025-07-29 00:41:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5952552, in state QUEUED (PENDING) -2025-07-29 00:41:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen002, jobid=5952471, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:41:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f084-f090 -2025-07-29 00:41:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f093-f099 -2025-07-29 00:41:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f005 -2025-07-29 00:41:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f006 -2025-07-29 00:41:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f007 -2025-07-29 00:41:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f008 -2025-07-29 00:41:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f009 -2025-07-29 00:41:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_esfc -2025-07-29 00:41:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f084-f090 succeeded, jobid=5952571 -2025-07-29 00:41:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f093-f099 succeeded, jobid=5952572 -2025-07-29 00:41:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f005 succeeded, jobid=5952573 -2025-07-29 00:41:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f006 succeeded, jobid=5952574 -2025-07-29 00:41:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f007 succeeded, jobid=5952575 -2025-07-29 00:41:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f008 succeeded, jobid=5952576 -2025-07-29 00:41:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f009 succeeded, jobid=5952577 -2025-07-29 00:41:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_esfc succeeded, jobid=5952578 -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlprod, jobid=5952466, in state SUCCEEDED (COMPLETED), ran for 106.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state RUNNING (RUNNING) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952440, in state SUCCEEDED (COMPLETED), ran for 271.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952441, in state SUCCEEDED (COMPLETED), ran for 279.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952442, in state SUCCEEDED (COMPLETED), ran for 325.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952467, in state SUCCEEDED (COMPLETED), ran for 326.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952545, in state RUNNING (RUNNING) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952546, in state QUEUED (PENDING) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952571, in state QUEUED (PENDING) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952572, in state QUEUED (PENDING) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlprod, jobid=5952547, in state QUEUED (PENDING) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952548, in state QUEUED (PENDING) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5952549, in state QUEUED (PENDING) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5952550, in state QUEUED (PENDING) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952551, in state QUEUED (PENDING) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5952552, in state QUEUED (PENDING) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5952573, in state QUEUED (PENDING) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952574, in state QUEUED (PENDING) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5952575, in state QUEUED (PENDING) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5952576, in state QUEUED (PENDING) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952577, in state QUEUED (PENDING) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_esfc, jobid=5952578, in state QUEUED (PENDING) -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f102-f108 -2025-07-29 00:47:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_atmos_prod_f102-f108 is pending at druby://130.18.14.151:44675 -2025-07-29 00:53:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_atmos_prod_f102-f108 is success, jobid=5952628 -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5952028, in state SUCCEEDED (COMPLETED), ran for 3218.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952545, in state SUCCEEDED (COMPLETED), ran for 328.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952546, in state QUEUED (PENDING) -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952571, in state QUEUED (PENDING) -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952572, in state QUEUED (PENDING) -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952628, in state QUEUED (PENDING) -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlprod, jobid=5952547, in state QUEUED (PENDING) -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952548, in state QUEUED (PENDING) -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5952549, in state QUEUED (PENDING) -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5952550, in state QUEUED (PENDING) -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952551, in state QUEUED (PENDING) -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5952552, in state QUEUED (PENDING) -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5952573, in state QUEUED (PENDING) -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952574, in state QUEUED (PENDING) -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5952575, in state QUEUED (PENDING) -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5952576, in state QUEUED (PENDING) -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952577, in state QUEUED (PENDING) -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_esfc, jobid=5952578, in state SUCCEEDED (COMPLETED), ran for 143.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f111-f114 -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f117-f120 -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_fcst_mem001 -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_fcst_mem002 -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f111-f114 succeeded, jobid=5952654 -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f117-f120 succeeded, jobid=5952655 -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_fcst_mem001 succeeded, jobid=5952656 -2025-07-29 00:53:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_fcst_mem002 succeeded, jobid=5952657 -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952546, in state RUNNING (RUNNING) -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952571, in state RUNNING (RUNNING) -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952572, in state RUNNING (RUNNING) -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952628, in state QUEUED (PENDING) -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5952654, in state QUEUED (PENDING) -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5952655, in state QUEUED (PENDING) -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlprod, jobid=5952547, in state SUCCEEDED (COMPLETED), ran for 63.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952548, in state SUCCEEDED (COMPLETED), ran for 65.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5952549, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5952550, in state SUCCEEDED (COMPLETED), ran for 66.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952551, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5952552, in state SUCCEEDED (COMPLETED), ran for 76.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5952573, in state SUCCEEDED (COMPLETED), ran for 68.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952574, in state SUCCEEDED (COMPLETED), ran for 66.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5952575, in state SUCCEEDED (COMPLETED), ran for 64.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5952576, in state SUCCEEDED (COMPLETED), ran for 68.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952577, in state SUCCEEDED (COMPLETED), ran for 66.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem001, jobid=5952656, in state QUEUED (PENDING) -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem002, jobid=5952657, in state QUEUED (PENDING) -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_fit2obs -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_fit2obs is pending at druby://130.18.14.151:38847 -2025-07-29 01:05:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_fit2obs is success, jobid=5952678 -2025-07-29 01:05:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952546, in state SUCCEEDED (COMPLETED), ran for 321.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:05:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952571, in state SUCCEEDED (COMPLETED), ran for 325.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:05:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952572, in state SUCCEEDED (COMPLETED), ran for 333.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:05:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952628, in state QUEUED (PENDING) -2025-07-29 01:05:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5952654, in state QUEUED (PENDING) -2025-07-29 01:05:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5952655, in state QUEUED (PENDING) -2025-07-29 01:05:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fit2obs, jobid=5952678, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:05:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem001, jobid=5952656, in state QUEUED (PENDING) -2025-07-29 01:05:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem002, jobid=5952657, in state QUEUED (PENDING) -2025-07-29 01:05:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_arch_vrfy -2025-07-29 01:05:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_arch_vrfy is pending at druby://130.18.14.151:37477 -2025-07-29 01:11:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_arch_vrfy is success, jobid=5952703 -2025-07-29 01:11:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952628, in state RUNNING (RUNNING) -2025-07-29 01:11:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5952654, in state QUEUED (PENDING) -2025-07-29 01:11:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5952655, in state QUEUED (PENDING) -2025-07-29 01:11:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_arch_vrfy, jobid=5952703, in state SUCCEEDED (COMPLETED), ran for 34.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:11:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem001, jobid=5952656, in state QUEUED (PENDING) -2025-07-29 01:11:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem002, jobid=5952657, in state QUEUED (PENDING) -2025-07-29 01:11:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_cleanup -2025-07-29 01:11:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_cleanup is pending at druby://130.18.14.151:40839 -2025-07-29 01:17:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_cleanup is success, jobid=5952717 -2025-07-29 01:17:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952628, in state SUCCEEDED (COMPLETED), ran for 327.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:17:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5952654, in state RUNNING (RUNNING) -2025-07-29 01:17:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5952655, in state RUNNING (RUNNING) -2025-07-29 01:17:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_cleanup, jobid=5952717, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:17:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem001, jobid=5952656, in state RUNNING (RUNNING) -2025-07-29 01:17:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem002, jobid=5952657, in state QUEUED (PENDING) -2025-07-29 01:23:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5952654, in state SUCCEEDED (COMPLETED), ran for 215.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:23:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5952655, in state SUCCEEDED (COMPLETED), ran for 216.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:23:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem001, jobid=5952656, in state SUCCEEDED (COMPLETED), ran for 269.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:23:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem002, jobid=5952657, in state RUNNING (RUNNING) -2025-07-29 01:23:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_tracker -2025-07-29 01:23:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_genesis -2025-07-29 01:23:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_echgres -2025-07-29 01:23:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_tracker succeeded, jobid=5952735 -2025-07-29 01:23:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_genesis succeeded, jobid=5952736 -2025-07-29 01:23:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_echgres succeeded, jobid=5952737 -2025-07-29 01:29:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_tracker, jobid=5952735, in state SUCCEEDED (COMPLETED), ran for 68.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:29:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_genesis, jobid=5952736, in state RUNNING (RUNNING) -2025-07-29 01:29:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5952737, in state SUCCEEDED (COMPLETED), ran for 33.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:29:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem002, jobid=5952657, in state SUCCEEDED (COMPLETED), ran for 264.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:29:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos000 -2025-07-29 01:29:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos001 -2025-07-29 01:29:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos002 -2025-07-29 01:29:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos003 -2025-07-29 01:29:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos004 -2025-07-29 01:29:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos005 -2025-07-29 01:29:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos006 -2025-07-29 01:29:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos000 succeeded, jobid=5952758 -2025-07-29 01:29:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos001 succeeded, jobid=5952759 -2025-07-29 01:29:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos002 succeeded, jobid=5952760 -2025-07-29 01:29:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos003 succeeded, jobid=5952761 -2025-07-29 01:29:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos004 succeeded, jobid=5952762 -2025-07-29 01:29:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos005 succeeded, jobid=5952763 -2025-07-29 01:29:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos006 succeeded, jobid=5952764 -2025-07-29 01:35:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_genesis, jobid=5952736, in state SUCCEEDED (COMPLETED), ran for 517.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:35:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5952758, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:35:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5952759, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:35:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5952760, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:35:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5952761, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:35:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5952762, in state SUCCEEDED (COMPLETED), ran for 37.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:35:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5952763, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:35:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5952764, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:35:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_arch_vrfy -2025-07-29 01:35:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_earc_vrfy -2025-07-29 01:35:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_arch_vrfy succeeded, jobid=5952796 -2025-07-29 01:35:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_earc_vrfy succeeded, jobid=5952797 -2025-07-29 01:41:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_arch_vrfy, jobid=5952796, in state SUCCEEDED (COMPLETED), ran for 59.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:41:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_earc_vrfy, jobid=5952797, in state SUCCEEDED (COMPLETED), ran for 15.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:41:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_metpg2g1 -2025-07-29 01:41:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_cleanup -2025-07-29 01:41:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_metpg2g1 is pending at druby://130.18.14.151:35039 -2025-07-29 01:41:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of enkfgdas_cleanup is pending at druby://130.18.14.151:35039 -2025-07-29 01:47:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_metpg2g1 is success, jobid=5952816 -2025-07-29 01:47:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending enkfgdas_cleanup is success, jobid=5952817 -2025-07-29 01:47:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_metpg2g1, jobid=5952816, in state SUCCEEDED (COMPLETED), ran for 110.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:47:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_cleanup, jobid=5952817, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:47:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_metpg2o1 -2025-07-29 01:47:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_metpg2o1 is pending at druby://130.18.14.151:38343 -2025-07-29 01:53:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_metpg2o1 is success, jobid=5952853 -2025-07-29 01:53:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_metpg2o1, jobid=5952853, in state RUNNING (RUNNING) -2025-07-29 02:00:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_metpg2o1, jobid=5952853, in state SUCCEEDED (COMPLETED), ran for 358.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:00:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_metppcp1 -2025-07-29 02:00:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_metppcp1 succeeded, jobid=5952923 -2025-07-29 02:06:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_metppcp1, jobid=5952923, in state SUCCEEDED (COMPLETED), ran for 28.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:06:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_cleanup -2025-07-29 02:06:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_cleanup succeeded, jobid=5952999 -2025-07-29 02:12:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_cleanup, jobid=5952999, in state SUCCEEDED (COMPLETED), ran for 21.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:12:06 -0500 :: hercules-login-1.hpc.msstate.edu :: This cycle is complete: Success diff --git a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96C48_hybatmDA_logs_2021122106.log b/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96C48_hybatmDA_logs_2021122106.log deleted file mode 100644 index 1beb799a..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96C48_hybatmDA_logs_2021122106.log +++ /dev/null @@ -1,146 +0,0 @@ -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_prep -2025-07-29 00:59:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_prep is pending at druby://130.18.14.151:38847 -2025-07-29 01:05:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_prep is success, jobid=5952679 -2025-07-29 01:05:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5952679, in state QUEUED (PENDING) -2025-07-29 01:11:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5952679, in state QUEUED (PENDING) -2025-07-29 01:17:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5952679, in state QUEUED (PENDING) -2025-07-29 01:23:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5952679, in state SUCCEEDED (COMPLETED), ran for 115.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:35:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_anal -2025-07-29 01:35:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_eobs -2025-07-29 01:35:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_anal succeeded, jobid=5952798 -2025-07-29 01:35:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_eobs succeeded, jobid=5952799 -2025-07-29 01:41:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5952798, in state RUNNING (RUNNING) -2025-07-29 01:41:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eobs, jobid=5952799, in state RUNNING (RUNNING) -2025-07-29 01:47:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5952798, in state RUNNING (RUNNING) -2025-07-29 01:47:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eobs, jobid=5952799, in state SUCCEEDED (COMPLETED), ran for 415.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:47:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_ediag -2025-07-29 01:47:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of enkfgdas_ediag is pending at druby://130.18.14.151:38343 -2025-07-29 01:53:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending enkfgdas_ediag is success, jobid=5952854 -2025-07-29 01:53:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5952798, in state SUCCEEDED (COMPLETED), ran for 773.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:53:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ediag, jobid=5952854, in state SUCCEEDED (COMPLETED), ran for 142.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:53:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_analdiag -2025-07-29 01:53:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_vminmon -2025-07-29 01:53:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_eupd -2025-07-29 01:53:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_analdiag succeeded, jobid=5952891 -2025-07-29 01:53:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_vminmon succeeded, jobid=5952892 -2025-07-29 01:53:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of enkfgdas_eupd is pending at druby://130.18.14.151:43449 -2025-07-29 02:00:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending enkfgdas_eupd is success, jobid=5952893 -2025-07-29 02:00:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5952891, in state SUCCEEDED (COMPLETED), ran for 123.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:00:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_vminmon, jobid=5952892, in state SUCCEEDED (COMPLETED), ran for 15.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:00:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5952893, in state SUCCEEDED (COMPLETED), ran for 136.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:00:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_sfcanl -2025-07-29 02:00:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_verfozn -2025-07-29 02:00:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_verfrad -2025-07-29 02:00:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_sfcanl succeeded, jobid=5952924 -2025-07-29 02:00:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_verfozn succeeded, jobid=5952925 -2025-07-29 02:00:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_verfrad succeeded, jobid=5952926 -2025-07-29 02:06:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952924, in state SUCCEEDED (COMPLETED), ran for 64.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:06:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfozn, jobid=5952925, in state SUCCEEDED (COMPLETED), ran for 69.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:06:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5952926, in state RUNNING (RUNNING) -2025-07-29 02:06:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_analcalc -2025-07-29 02:06:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_fcst_seg0 -2025-07-29 02:06:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_analcalc succeeded, jobid=5953000 -2025-07-29 02:06:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_fcst_seg0 is pending at druby://130.18.14.151:36151 -2025-07-29 02:12:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_fcst_seg0 is success, jobid=5953001 -2025-07-29 02:12:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5953000, in state SUCCEEDED (COMPLETED), ran for 59.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:12:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5953001, in state RUNNING (RUNNING) -2025-07-29 02:12:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5952926, in state RUNNING (RUNNING) -2025-07-29 02:12:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_ecen000 -2025-07-29 02:12:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_ecen001 -2025-07-29 02:12:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_ecen002 -2025-07-29 02:12:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_esfc -2025-07-29 02:12:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_ecen000 succeeded, jobid=5953069 -2025-07-29 02:12:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_ecen001 succeeded, jobid=5953070 -2025-07-29 02:12:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_ecen002 succeeded, jobid=5953071 -2025-07-29 02:12:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_esfc succeeded, jobid=5953072 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5953001, in state SUCCEEDED (COMPLETED), ran for 421.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5952926, in state SUCCEEDED (COMPLETED), ran for 716.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen000, jobid=5953069, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen001, jobid=5953070, in state SUCCEEDED (COMPLETED), ran for 34.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen002, jobid=5953071, in state SUCCEEDED (COMPLETED), ran for 37.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_esfc, jobid=5953072, in state SUCCEEDED (COMPLETED), ran for 143.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmanlupp -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f000 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f001 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f002 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f003 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f004 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f005 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f006 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f007 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f008 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f009 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_fcst_mem001 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_fcst_mem002 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmanlupp succeeded, jobid=5953101 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f000 succeeded, jobid=5953102 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f001 succeeded, jobid=5953103 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f002 succeeded, jobid=5953104 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f003 succeeded, jobid=5953105 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f004 succeeded, jobid=5953106 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f005 succeeded, jobid=5953107 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f006 succeeded, jobid=5953108 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f007 succeeded, jobid=5953109 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f008 succeeded, jobid=5953110 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f009 succeeded, jobid=5953111 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_fcst_mem001 succeeded, jobid=5953112 -2025-07-29 02:18:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_fcst_mem002 succeeded, jobid=5953113 -2025-07-29 02:24:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5953101, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:24:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5953102, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:24:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5953103, in state SUCCEEDED (COMPLETED), ran for 61.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:24:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5953104, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:24:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5953105, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:24:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5953106, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:24:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5953107, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:24:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5953108, in state SUCCEEDED (COMPLETED), ran for 65.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:24:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5953109, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:24:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5953110, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:24:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5953111, in state SUCCEEDED (COMPLETED), ran for 60.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:24:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem001, jobid=5953112, in state RUNNING (RUNNING) -2025-07-29 02:24:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem002, jobid=5953113, in state SUCCEEDED (COMPLETED), ran for 247.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:24:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmanlprod -2025-07-29 02:24:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_fit2obs -2025-07-29 02:24:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_atmanlprod is pending at druby://130.18.14.151:40885 -2025-07-29 02:24:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_fit2obs is pending at druby://130.18.14.151:40885 -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_atmanlprod is success, jobid=5953154 -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_fit2obs is success, jobid=5953155 -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlprod, jobid=5953154, in state SUCCEEDED (COMPLETED), ran for 63.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fit2obs, jobid=5953155, in state SUCCEEDED (COMPLETED), ran for 15.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem001, jobid=5953112, in state SUCCEEDED (COMPLETED), ran for 269.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_arch_vrfy -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_echgres -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos000 -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos001 -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos002 -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos003 -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos004 -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos005 -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos006 -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_arch_vrfy succeeded, jobid=5953187 -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_echgres succeeded, jobid=5953188 -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos000 succeeded, jobid=5953189 -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos001 succeeded, jobid=5953190 -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos002 succeeded, jobid=5953191 -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos003 succeeded, jobid=5953192 -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos004 succeeded, jobid=5953193 -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos005 succeeded, jobid=5953194 -2025-07-29 02:30:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos006 succeeded, jobid=5953195 -2025-07-29 02:36:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_arch_vrfy, jobid=5953187, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:36:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5953188, in state SUCCEEDED (COMPLETED), ran for 26.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:36:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5953189, in state SUCCEEDED (COMPLETED), ran for 33.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:36:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5953190, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:36:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5953191, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:36:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5953192, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:36:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5953193, in state SUCCEEDED (COMPLETED), ran for 35.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:36:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5953194, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:36:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5953195, in state SUCCEEDED (COMPLETED), ran for 37.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:36:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_cleanup -2025-07-29 02:36:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_earc_vrfy -2025-07-29 02:36:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_cleanup is pending at druby://130.18.14.151:43779 -2025-07-29 02:36:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of enkfgdas_earc_vrfy is pending at druby://130.18.14.151:43779 -2025-07-29 02:42:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_cleanup is success, jobid=5953217 -2025-07-29 02:42:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending enkfgdas_earc_vrfy is success, jobid=5953218 -2025-07-29 02:42:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_cleanup, jobid=5953217, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:42:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_earc_vrfy, jobid=5953218, in state SUCCEEDED (COMPLETED), ran for 35.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:42:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_cleanup -2025-07-29 02:42:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of enkfgdas_cleanup is pending at druby://130.18.14.151:40577 diff --git a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96C48mx500_S2SW_cyc_gfs_logs_2021122012.log b/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96C48mx500_S2SW_cyc_gfs_logs_2021122012.log deleted file mode 100644 index c5d961f9..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96C48mx500_S2SW_cyc_gfs_logs_2021122012.log +++ /dev/null @@ -1,714 +0,0 @@ -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_waveinit -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_stage_ic -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_stage_ic -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_waveinit succeeded, jobid=5950276 -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_stage_ic succeeded, jobid=5950277 -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_stage_ic succeeded, jobid=5950278 -2025-07-28 14:48:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_waveinit, jobid=5950276, in state QUEUED (PENDING) -2025-07-28 14:48:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_stage_ic, jobid=5950277, in state RUNNING (RUNNING) -2025-07-28 14:48:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_stage_ic, jobid=5950278, in state RUNNING (RUNNING) -2025-07-28 14:54:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_waveinit, jobid=5950276, in state QUEUED (PENDING) -2025-07-28 14:54:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_stage_ic, jobid=5950277, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 14:54:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_stage_ic, jobid=5950278, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 14:54:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_fcst_seg0 -2025-07-28 14:54:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_fcst_mem001 -2025-07-28 14:54:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_fcst_mem002 -2025-07-28 14:54:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_fcst_seg0 succeeded, jobid=5950294 -2025-07-28 14:54:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_fcst_mem001 succeeded, jobid=5950295 -2025-07-28 14:54:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_fcst_mem002 succeeded, jobid=5950296 -2025-07-28 15:00:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_waveinit, jobid=5950276, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:00:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5950294, in state QUEUED (PENDING) -2025-07-28 15:00:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem001, jobid=5950295, in state QUEUED (PENDING) -2025-07-28 15:00:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem002, jobid=5950296, in state QUEUED (PENDING) -2025-07-28 15:06:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5950294, in state RUNNING (RUNNING) -2025-07-28 15:06:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem001, jobid=5950295, in state QUEUED (PENDING) -2025-07-28 15:06:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem002, jobid=5950296, in state QUEUED (PENDING) -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5950294, in state SUCCEEDED (COMPLETED), ran for 389.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem001, jobid=5950295, in state SUCCEEDED (COMPLETED), ran for 241.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem002, jobid=5950296, in state SUCCEEDED (COMPLETED), ran for 254.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f000 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f001 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f002 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f003 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f004 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f005 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f006 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f007 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f008 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f009 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_echgres -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos000 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos001 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos002 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos003 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos004 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos005 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos006 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f000 succeeded, jobid=5950387 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f001 succeeded, jobid=5950388 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f002 succeeded, jobid=5950389 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f003 succeeded, jobid=5950390 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f004 succeeded, jobid=5950391 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f005 succeeded, jobid=5950392 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f006 succeeded, jobid=5950393 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f007 succeeded, jobid=5950394 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f008 succeeded, jobid=5950395 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f009 succeeded, jobid=5950396 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_echgres succeeded, jobid=5950397 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos000 succeeded, jobid=5950398 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos001 succeeded, jobid=5950399 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos002 succeeded, jobid=5950400 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos003 succeeded, jobid=5950401 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos004 succeeded, jobid=5950402 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos005 succeeded, jobid=5950403 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos006 succeeded, jobid=5950404 -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950397, in state QUEUED (PENDING) -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950397, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950397, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950397, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950397, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950397, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950397, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950397, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950397, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950397, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950397, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950397, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 16:31:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 16:31:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 16:31:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 16:31:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 16:31:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 16:31:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 16:31:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 16:31:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 16:31:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 16:31:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 16:31:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950397, in state QUEUED (PENDING) -2025-07-28 16:31:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 16:31:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 16:31:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 16:31:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 16:31:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 16:31:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 16:31:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5950397, in state SUCCEEDED (COMPLETED), ran for 31.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 16:43:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 16:43:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 16:43:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 16:43:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 16:43:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 16:43:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 16:43:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 16:43:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 16:43:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 16:43:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 16:43:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 16:43:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 16:43:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 16:43:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 16:43:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 16:43:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 16:43:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 16:49:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 16:49:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 16:49:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 16:49:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 16:49:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 16:49:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 16:49:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 16:49:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 16:49:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 16:49:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 16:49:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 16:49:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 16:49:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 16:49:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 16:49:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 16:49:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 16:49:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 16:55:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 16:55:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 16:55:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 16:55:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 16:55:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 16:55:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 16:55:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 16:55:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 16:55:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 16:55:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 16:55:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 16:55:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 16:55:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 16:55:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 16:55:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 16:55:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 16:55:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 17:01:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 17:01:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 17:01:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 17:01:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 17:01:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 17:01:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 17:01:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 17:01:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 17:01:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 17:01:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 17:01:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 17:01:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 17:01:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 17:01:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 17:01:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 17:01:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 17:01:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 17:07:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 17:07:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 17:07:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 17:07:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 17:07:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 17:07:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 17:07:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 17:07:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 17:07:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 17:07:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 17:07:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 17:07:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 17:07:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 17:07:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 17:07:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 17:07:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 17:07:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 17:13:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 17:13:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 17:13:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 17:13:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 17:13:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 17:13:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 17:13:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 17:13:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 17:13:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 17:13:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 17:13:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 17:13:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 17:13:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 17:13:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 17:13:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 17:13:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 17:13:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 17:19:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 17:19:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 17:19:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 17:19:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 17:19:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 17:19:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 17:19:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 17:19:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 17:19:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 17:19:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 17:19:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 17:19:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 17:19:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 17:19:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 17:19:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 17:19:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 17:19:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 17:26:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 17:26:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 17:26:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 17:26:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 17:26:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 17:26:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 17:26:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 17:26:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 17:26:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 17:26:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 17:26:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 17:26:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 17:26:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 17:26:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 17:26:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 17:26:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 17:26:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 17:32:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 17:32:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 17:32:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 17:32:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 17:32:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 17:32:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 17:32:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 17:32:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 17:32:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 17:32:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 17:32:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 17:32:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 17:32:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 17:32:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 17:32:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 17:32:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 17:32:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 17:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 17:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 17:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 17:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 17:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 17:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 17:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 17:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 17:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 17:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 17:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 17:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 17:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 17:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 17:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 17:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 17:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 17:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 17:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 17:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 17:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 17:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 17:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 17:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 17:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 17:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 17:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 17:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 17:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 17:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 17:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 17:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 17:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 17:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 17:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 17:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 17:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 17:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 17:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 17:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 17:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 17:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 17:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 17:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 17:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 17:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 17:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 17:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 17:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 17:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 17:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 17:56:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 17:56:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 17:56:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 17:56:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 17:56:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 17:56:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 17:56:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 17:56:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 17:56:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 17:56:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 17:56:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 17:56:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 17:56:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 17:56:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 17:56:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 17:56:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 17:56:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 18:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 18:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 18:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 18:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 18:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 18:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 18:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 18:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 18:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 18:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 18:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 18:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 18:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 18:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 18:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 18:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 18:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 18:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 18:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 18:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 18:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 18:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 18:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 18:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 18:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 18:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 18:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 18:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 18:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 18:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 18:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 18:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 18:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 18:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 18:14:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 18:14:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 18:14:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 18:14:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 18:14:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 18:14:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 18:14:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 18:14:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 18:14:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 18:14:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 18:14:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 18:14:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 18:14:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 18:14:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 18:14:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 18:14:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 18:14:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 18:20:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 18:20:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 18:20:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 18:20:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 18:20:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 18:20:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 18:20:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 18:20:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 18:20:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 18:20:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 18:20:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 18:20:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 18:20:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 18:20:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 18:20:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 18:20:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 18:20:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 18:26:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 18:26:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 18:26:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 18:26:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 18:26:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 18:26:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 18:26:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 18:26:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 18:26:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 18:26:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 18:26:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 18:26:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 18:26:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 18:26:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 18:26:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 18:26:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 18:26:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 18:32:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state QUEUED (PENDING) -2025-07-28 18:32:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 18:32:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 18:32:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 18:32:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 18:32:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 18:32:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 18:32:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 18:32:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 18:32:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 18:32:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 18:32:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 18:32:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 18:32:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 18:32:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 18:32:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 18:32:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 18:38:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state RUNNING (RUNNING) -2025-07-28 18:38:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 18:38:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 18:38:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 18:38:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 18:38:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 18:38:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 18:38:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 18:38:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 18:38:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 18:38:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 18:38:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 18:38:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 18:38:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 18:38:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 18:38:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 18:38:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 18:44:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950387, in state SUCCEEDED (COMPLETED), ran for 63.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 18:44:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 18:44:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 18:44:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 18:44:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 18:44:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 18:44:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 18:44:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 18:44:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 18:44:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 18:44:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 18:44:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 18:44:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 18:44:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 18:44:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 18:44:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 18:44:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 18:50:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state QUEUED (PENDING) -2025-07-28 18:50:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state QUEUED (PENDING) -2025-07-28 18:50:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state QUEUED (PENDING) -2025-07-28 18:50:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state QUEUED (PENDING) -2025-07-28 18:50:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state QUEUED (PENDING) -2025-07-28 18:50:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state QUEUED (PENDING) -2025-07-28 18:50:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state QUEUED (PENDING) -2025-07-28 18:50:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state QUEUED (PENDING) -2025-07-28 18:50:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state QUEUED (PENDING) -2025-07-28 18:50:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state QUEUED (PENDING) -2025-07-28 18:50:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 18:50:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 18:50:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 18:50:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 18:50:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 18:50:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5950388, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5950389, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950390, in state SUCCEEDED (COMPLETED), ran for 68.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5950391, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5950392, in state SUCCEEDED (COMPLETED), ran for 68.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950393, in state SUCCEEDED (COMPLETED), ran for 69.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5950394, in state SUCCEEDED (COMPLETED), ran for 68.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5950395, in state SUCCEEDED (COMPLETED), ran for 69.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950396, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5950398, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state QUEUED (PENDING) -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state QUEUED (PENDING) -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state QUEUED (PENDING) -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state QUEUED (PENDING) -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 19:02:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5950399, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:02:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5950400, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:02:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5950401, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:02:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5950402, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:02:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state QUEUED (PENDING) -2025-07-28 19:02:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state QUEUED (PENDING) -2025-07-28 19:08:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5950403, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:08:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5950404, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:08:59 -0500 :: hercules-login-1.hpc.msstate.edu :: This cycle is complete: Success diff --git a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96C48mx500_S2SW_cyc_gfs_logs_2021122018.log b/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96C48mx500_S2SW_cyc_gfs_logs_2021122018.log deleted file mode 100644 index cdef672b..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96C48mx500_S2SW_cyc_gfs_logs_2021122018.log +++ /dev/null @@ -1,1422 +0,0 @@ -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_prepoceanobs -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_marinebmatinit -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_prepoceanobs -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marinebmatinit -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_prepoceanobs succeeded, jobid=5950405 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_marinebmatinit succeeded, jobid=5950406 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_prepoceanobs succeeded, jobid=5950407 -2025-07-28 15:12:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_marinebmatinit succeeded, jobid=5950408 -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prepoceanobs, jobid=5950405, in state QUEUED (PENDING) -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmatinit, jobid=5950406, in state QUEUED (PENDING) -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5950407, in state QUEUED (PENDING) -2025-07-28 15:18:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950408, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prepoceanobs, jobid=5950405, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmatinit, jobid=5950406, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5950407, in state QUEUED (PENDING) -2025-07-28 15:24:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950408, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prepoceanobs, jobid=5950405, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmatinit, jobid=5950406, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5950407, in state QUEUED (PENDING) -2025-07-28 15:30:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950408, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prepoceanobs, jobid=5950405, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmatinit, jobid=5950406, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5950407, in state QUEUED (PENDING) -2025-07-28 15:37:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950408, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prepoceanobs, jobid=5950405, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmatinit, jobid=5950406, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5950407, in state QUEUED (PENDING) -2025-07-28 15:43:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950408, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prepoceanobs, jobid=5950405, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmatinit, jobid=5950406, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5950407, in state QUEUED (PENDING) -2025-07-28 15:49:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950408, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prepoceanobs, jobid=5950405, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmatinit, jobid=5950406, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5950407, in state QUEUED (PENDING) -2025-07-28 15:55:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950408, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prepoceanobs, jobid=5950405, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmatinit, jobid=5950406, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5950407, in state QUEUED (PENDING) -2025-07-28 16:01:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950408, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prepoceanobs, jobid=5950405, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmatinit, jobid=5950406, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5950407, in state QUEUED (PENDING) -2025-07-28 16:07:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950408, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prepoceanobs, jobid=5950405, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmatinit, jobid=5950406, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5950407, in state QUEUED (PENDING) -2025-07-28 16:13:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950408, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prepoceanobs, jobid=5950405, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmatinit, jobid=5950406, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5950407, in state QUEUED (PENDING) -2025-07-28 16:19:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950408, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prepoceanobs, jobid=5950405, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmatinit, jobid=5950406, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5950407, in state QUEUED (PENDING) -2025-07-28 16:25:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950408, in state QUEUED (PENDING) -2025-07-28 16:31:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prepoceanobs, jobid=5950405, in state QUEUED (PENDING) -2025-07-28 16:31:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmatinit, jobid=5950406, in state QUEUED (PENDING) -2025-07-28 16:31:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5950407, in state QUEUED (PENDING) -2025-07-28 16:31:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950408, in state QUEUED (PENDING) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prepoceanobs, jobid=5950405, in state SUCCEEDED (COMPLETED), ran for 205.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmatinit, jobid=5950406, in state SUCCEEDED (COMPLETED), ran for 33.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5950407, in state SUCCEEDED (COMPLETED), ran for 206.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5950408, in state SUCCEEDED (COMPLETED), ran for 33.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_marinebmat -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocnanalecen -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marinebmat -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_ocnanalecen -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_marinebmat succeeded, jobid=5950700 -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocnanalecen succeeded, jobid=5950701 -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_marinebmat succeeded, jobid=5950702 -2025-07-28 16:37:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_ocnanalecen succeeded, jobid=5950703 -2025-07-28 16:43:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 16:43:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocnanalecen, jobid=5950701, in state QUEUED (PENDING) -2025-07-28 16:43:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 16:43:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_ocnanalecen, jobid=5950703, in state QUEUED (PENDING) -2025-07-28 16:49:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 16:49:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocnanalecen, jobid=5950701, in state QUEUED (PENDING) -2025-07-28 16:49:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 16:49:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_ocnanalecen, jobid=5950703, in state QUEUED (PENDING) -2025-07-28 16:55:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 16:55:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocnanalecen, jobid=5950701, in state SUCCEEDED (COMPLETED), ran for 61.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:55:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 16:55:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_ocnanalecen, jobid=5950703, in state SUCCEEDED (COMPLETED), ran for 61.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 17:01:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 17:01:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 17:07:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 17:07:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 17:13:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 17:13:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 17:19:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 17:19:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 17:26:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 17:26:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 17:32:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 17:32:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 17:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 17:38:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 17:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 17:44:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 17:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 17:50:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 17:56:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 17:56:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 18:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 18:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 18:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 18:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 18:14:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 18:14:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 18:20:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 18:20:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 18:26:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 18:26:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 18:32:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 18:32:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 18:38:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 18:38:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 18:44:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 18:44:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 18:50:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 18:50:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_prep -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_prep -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_prep succeeded, jobid=5950948 -2025-07-28 18:56:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_prep succeeded, jobid=5950949 -2025-07-28 19:02:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 19:02:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 19:02:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 19:02:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 19:08:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 19:08:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 19:08:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 19:08:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 19:15:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 19:15:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 19:15:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 19:15:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 19:21:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 19:21:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 19:21:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 19:21:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 19:27:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 19:27:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 19:27:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 19:27:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 19:33:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 19:33:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 19:33:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 19:33:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 19:39:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 19:39:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 19:39:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 19:39:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 19:45:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 19:45:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 19:45:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 19:45:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 19:51:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 19:51:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state QUEUED (PENDING) -2025-07-28 19:51:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 19:51:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 19:57:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 19:57:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5950700, in state SUCCEEDED (COMPLETED), ran for 42.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:57:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 19:57:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 19:57:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_marineanlinit -2025-07-28 19:57:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_marineanlinit succeeded, jobid=5951229 -2025-07-28 20:03:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 20:03:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlinit, jobid=5951229, in state QUEUED (PENDING) -2025-07-28 20:03:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 20:03:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 20:09:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 20:09:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlinit, jobid=5951229, in state QUEUED (PENDING) -2025-07-28 20:09:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 20:09:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 20:15:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 20:15:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlinit, jobid=5951229, in state QUEUED (PENDING) -2025-07-28 20:15:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 20:15:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 20:21:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 20:21:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlinit, jobid=5951229, in state QUEUED (PENDING) -2025-07-28 20:21:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 20:21:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state QUEUED (PENDING) -2025-07-28 20:27:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 20:27:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlinit, jobid=5951229, in state SUCCEEDED (COMPLETED), ran for 33.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 20:27:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 20:27:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5950702, in state SUCCEEDED (COMPLETED), ran for 77.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 20:27:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_marineanlvar -2025-07-28 20:27:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marineanlinit -2025-07-28 20:27:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_marineanlvar succeeded, jobid=5951359 -2025-07-28 20:27:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_marineanlinit succeeded, jobid=5951360 -2025-07-28 20:33:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 20:33:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlvar, jobid=5951359, in state QUEUED (PENDING) -2025-07-28 20:33:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 20:33:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlinit, jobid=5951360, in state QUEUED (PENDING) -2025-07-28 20:39:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 20:39:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlvar, jobid=5951359, in state QUEUED (PENDING) -2025-07-28 20:39:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 20:39:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlinit, jobid=5951360, in state SUCCEEDED (COMPLETED), ran for 34.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 20:39:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marineanlvar -2025-07-28 20:39:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_marineanlvar succeeded, jobid=5951430 -2025-07-28 20:45:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 20:45:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlvar, jobid=5951359, in state QUEUED (PENDING) -2025-07-28 20:45:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 20:45:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951430, in state QUEUED (PENDING) -2025-07-28 20:51:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 20:51:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlvar, jobid=5951359, in state QUEUED (PENDING) -2025-07-28 20:51:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 20:51:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951430, in state QUEUED (PENDING) -2025-07-28 20:57:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 20:57:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlvar, jobid=5951359, in state QUEUED (PENDING) -2025-07-28 20:57:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 20:57:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951430, in state QUEUED (PENDING) -2025-07-28 21:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 21:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlvar, jobid=5951359, in state QUEUED (PENDING) -2025-07-28 21:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 21:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951430, in state QUEUED (PENDING) -2025-07-28 21:10:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 21:10:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlvar, jobid=5951359, in state QUEUED (PENDING) -2025-07-28 21:10:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 21:10:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951430, in state QUEUED (PENDING) -2025-07-28 21:16:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state QUEUED (PENDING) -2025-07-28 21:16:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlvar, jobid=5951359, in state QUEUED (PENDING) -2025-07-28 21:16:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state QUEUED (PENDING) -2025-07-28 21:16:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951430, in state QUEUED (PENDING) -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950948, in state SUCCEEDED (COMPLETED), ran for 125.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlvar, jobid=5951359, in state QUEUED (PENDING) -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950949, in state SUCCEEDED (COMPLETED), ran for 136.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951430, in state QUEUED (PENDING) -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_anal -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_snowanl -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_waveinit -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgfs_eobs -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgfs_esnowanl -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_anal -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_snowanl -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_waveinit -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_eobs -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_esnowanl -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_anal succeeded, jobid=5951640 -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_snowanl succeeded, jobid=5951641 -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_waveinit succeeded, jobid=5951642 -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgfs_eobs succeeded, jobid=5951643 -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgfs_esnowanl succeeded, jobid=5951644 -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_anal succeeded, jobid=5951645 -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_snowanl succeeded, jobid=5951646 -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_waveinit succeeded, jobid=5951647 -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_eobs succeeded, jobid=5951648 -2025-07-28 21:22:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_esnowanl succeeded, jobid=5951649 -2025-07-28 21:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951640, in state QUEUED (PENDING) -2025-07-28 21:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlvar, jobid=5951359, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_snowanl, jobid=5951641, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_waveinit, jobid=5951642, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_eobs, jobid=5951643, in state RUNNING (RUNNING) -2025-07-28 21:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_esnowanl, jobid=5951644, in state SUCCEEDED (COMPLETED), ran for 70.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951645, in state QUEUED (PENDING) -2025-07-28 21:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5951430, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_snowanl, jobid=5951646, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_waveinit, jobid=5951647, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eobs, jobid=5951648, in state RUNNING (RUNNING) -2025-07-28 21:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_esnowanl, jobid=5951649, in state SUCCEEDED (COMPLETED), ran for 70.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_marineanlchkpt -2025-07-28 21:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marineanlchkpt -2025-07-28 21:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_marineanlchkpt succeeded, jobid=5951700 -2025-07-28 21:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_marineanlchkpt succeeded, jobid=5951701 -2025-07-28 21:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951640, in state QUEUED (PENDING) -2025-07-28 21:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlchkpt, jobid=5951700, in state SUCCEEDED (COMPLETED), ran for 48.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_eobs, jobid=5951643, in state SUCCEEDED (COMPLETED), ran for 363.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951645, in state QUEUED (PENDING) -2025-07-28 21:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlchkpt, jobid=5951701, in state SUCCEEDED (COMPLETED), ran for 48.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eobs, jobid=5951648, in state SUCCEEDED (COMPLETED), ran for 402.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_marineanlfinal -2025-07-28 21:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgfs_ediag -2025-07-28 21:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marineanlfinal -2025-07-28 21:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_ediag -2025-07-28 21:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_marineanlfinal succeeded, jobid=5951760 -2025-07-28 21:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgfs_ediag succeeded, jobid=5951761 -2025-07-28 21:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_marineanlfinal succeeded, jobid=5951762 -2025-07-28 21:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_ediag succeeded, jobid=5951763 -2025-07-28 21:40:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951640, in state RUNNING (RUNNING) -2025-07-28 21:40:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlfinal, jobid=5951760, in state QUEUED (PENDING) -2025-07-28 21:40:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ediag, jobid=5951761, in state QUEUED (PENDING) -2025-07-28 21:40:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951645, in state RUNNING (RUNNING) -2025-07-28 21:40:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlfinal, jobid=5951762, in state QUEUED (PENDING) -2025-07-28 21:40:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ediag, jobid=5951763, in state QUEUED (PENDING) -2025-07-28 21:46:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951640, in state SUCCEEDED (COMPLETED), ran for 436.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:46:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlfinal, jobid=5951760, in state QUEUED (PENDING) -2025-07-28 21:46:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ediag, jobid=5951761, in state QUEUED (PENDING) -2025-07-28 21:46:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951645, in state RUNNING (RUNNING) -2025-07-28 21:46:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlfinal, jobid=5951762, in state QUEUED (PENDING) -2025-07-28 21:46:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ediag, jobid=5951763, in state QUEUED (PENDING) -2025-07-28 21:46:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_sfcanl -2025-07-28 21:46:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_vminmon -2025-07-28 21:46:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_sfcanl succeeded, jobid=5951832 -2025-07-28 21:46:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_vminmon succeeded, jobid=5951833 -2025-07-28 21:52:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlfinal, jobid=5951760, in state QUEUED (PENDING) -2025-07-28 21:52:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_sfcanl, jobid=5951832, in state QUEUED (PENDING) -2025-07-28 21:52:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_vminmon, jobid=5951833, in state SUCCEEDED (COMPLETED), ran for 13.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:52:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ediag, jobid=5951761, in state QUEUED (PENDING) -2025-07-28 21:52:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951645, in state SUCCEEDED (COMPLETED), ran for 763.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:52:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlfinal, jobid=5951762, in state QUEUED (PENDING) -2025-07-28 21:52:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ediag, jobid=5951763, in state QUEUED (PENDING) -2025-07-28 21:52:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_analdiag -2025-07-28 21:52:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_vminmon -2025-07-28 21:52:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_analdiag succeeded, jobid=5951889 -2025-07-28 21:52:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_vminmon succeeded, jobid=5951890 -2025-07-28 21:58:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlfinal, jobid=5951760, in state QUEUED (PENDING) -2025-07-28 21:58:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_sfcanl, jobid=5951832, in state QUEUED (PENDING) -2025-07-28 21:58:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ediag, jobid=5951761, in state QUEUED (PENDING) -2025-07-28 21:58:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlfinal, jobid=5951762, in state QUEUED (PENDING) -2025-07-28 21:58:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951889, in state QUEUED (PENDING) -2025-07-28 21:58:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_vminmon, jobid=5951890, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:58:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ediag, jobid=5951763, in state QUEUED (PENDING) -2025-07-28 22:04:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlfinal, jobid=5951760, in state SUCCEEDED (COMPLETED), ran for 33.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:04:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_sfcanl, jobid=5951832, in state QUEUED (PENDING) -2025-07-28 22:04:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ediag, jobid=5951761, in state SUCCEEDED (COMPLETED), ran for 129.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:04:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlfinal, jobid=5951762, in state SUCCEEDED (COMPLETED), ran for 33.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:04:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951889, in state QUEUED (PENDING) -2025-07-28 22:04:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ediag, jobid=5951763, in state SUCCEEDED (COMPLETED), ran for 143.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:04:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_anlstat -2025-07-28 22:04:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgfs_eupd -2025-07-28 22:04:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_anlstat -2025-07-28 22:04:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_eupd -2025-07-28 22:04:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_anlstat succeeded, jobid=5951932 -2025-07-28 22:04:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgfs_eupd succeeded, jobid=5951933 -2025-07-28 22:04:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_anlstat succeeded, jobid=5951934 -2025-07-28 22:04:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_eupd succeeded, jobid=5951935 -2025-07-28 22:10:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_sfcanl, jobid=5951832, in state SUCCEEDED (COMPLETED), ran for 45.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anlstat, jobid=5951932, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_eupd, jobid=5951933, in state QUEUED (PENDING) -2025-07-28 22:10:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951889, in state RUNNING (RUNNING) -2025-07-28 22:10:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anlstat, jobid=5951934, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951935, in state QUEUED (PENDING) -2025-07-28 22:10:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_analcalc -2025-07-28 22:10:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_fcst_seg0 -2025-07-28 22:10:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_analcalc succeeded, jobid=5951963 -2025-07-28 22:10:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_fcst_seg0 succeeded, jobid=5951964 -2025-07-28 22:16:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951963, in state QUEUED (PENDING) -2025-07-28 22:16:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state QUEUED (PENDING) -2025-07-28 22:16:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_eupd, jobid=5951933, in state QUEUED (PENDING) -2025-07-28 22:16:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951889, in state SUCCEEDED (COMPLETED), ran for 123.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:16:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951935, in state QUEUED (PENDING) -2025-07-28 22:16:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_verfozn -2025-07-28 22:16:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_verfrad -2025-07-28 22:16:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_verfozn succeeded, jobid=5951992 -2025-07-28 22:16:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_verfrad succeeded, jobid=5951993 -2025-07-28 22:22:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951963, in state QUEUED (PENDING) -2025-07-28 22:22:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state QUEUED (PENDING) -2025-07-28 22:22:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_eupd, jobid=5951933, in state QUEUED (PENDING) -2025-07-28 22:22:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfozn, jobid=5951992, in state SUCCEEDED (COMPLETED), ran for 71.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:22:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5951993, in state RUNNING (RUNNING) -2025-07-28 22:22:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951935, in state QUEUED (PENDING) -2025-07-28 22:28:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951963, in state QUEUED (PENDING) -2025-07-28 22:28:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state QUEUED (PENDING) -2025-07-28 22:28:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_eupd, jobid=5951933, in state QUEUED (PENDING) -2025-07-28 22:28:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5951993, in state SUCCEEDED (COMPLETED), ran for 563.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:28:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951935, in state QUEUED (PENDING) -2025-07-28 22:34:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951963, in state QUEUED (PENDING) -2025-07-28 22:34:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state QUEUED (PENDING) -2025-07-28 22:34:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_eupd, jobid=5951933, in state QUEUED (PENDING) -2025-07-28 22:34:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951935, in state QUEUED (PENDING) -2025-07-28 22:40:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951963, in state QUEUED (PENDING) -2025-07-28 22:40:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state QUEUED (PENDING) -2025-07-28 22:40:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_eupd, jobid=5951933, in state QUEUED (PENDING) -2025-07-28 22:40:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951935, in state QUEUED (PENDING) -2025-07-28 22:47:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951963, in state QUEUED (PENDING) -2025-07-28 22:47:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state QUEUED (PENDING) -2025-07-28 22:47:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_eupd, jobid=5951933, in state QUEUED (PENDING) -2025-07-28 22:47:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951935, in state QUEUED (PENDING) -2025-07-28 22:53:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951963, in state QUEUED (PENDING) -2025-07-28 22:53:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state QUEUED (PENDING) -2025-07-28 22:53:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_eupd, jobid=5951933, in state QUEUED (PENDING) -2025-07-28 22:53:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951935, in state QUEUED (PENDING) -2025-07-28 22:59:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951963, in state QUEUED (PENDING) -2025-07-28 22:59:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state QUEUED (PENDING) -2025-07-28 22:59:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_eupd, jobid=5951933, in state QUEUED (PENDING) -2025-07-28 22:59:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951935, in state QUEUED (PENDING) -2025-07-28 23:05:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951963, in state QUEUED (PENDING) -2025-07-28 23:05:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state QUEUED (PENDING) -2025-07-28 23:05:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_eupd, jobid=5951933, in state SUCCEEDED (COMPLETED), ran for 118.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:05:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5951935, in state SUCCEEDED (COMPLETED), ran for 135.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:05:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_sfcanl -2025-07-28 23:05:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_sfcanl succeeded, jobid=5952099 -2025-07-28 23:11:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951963, in state SUCCEEDED (COMPLETED), ran for 60.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:11:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state QUEUED (PENDING) -2025-07-28 23:11:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952099, in state QUEUED (PENDING) -2025-07-28 23:11:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgfs_ecen000 -2025-07-28 23:11:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgfs_ecen001 -2025-07-28 23:11:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgfs_ecen002 -2025-07-28 23:11:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgfs_esfc -2025-07-28 23:11:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgfs_ecen000 succeeded, jobid=5952119 -2025-07-28 23:11:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgfs_ecen001 succeeded, jobid=5952120 -2025-07-28 23:11:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgfs_ecen002 succeeded, jobid=5952121 -2025-07-28 23:11:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgfs_esfc succeeded, jobid=5952122 -2025-07-28 23:17:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state QUEUED (PENDING) -2025-07-28 23:17:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen000, jobid=5952119, in state QUEUED (PENDING) -2025-07-28 23:17:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen001, jobid=5952120, in state QUEUED (PENDING) -2025-07-28 23:17:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen002, jobid=5952121, in state QUEUED (PENDING) -2025-07-28 23:17:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_esfc, jobid=5952122, in state QUEUED (PENDING) -2025-07-28 23:17:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952099, in state QUEUED (PENDING) -2025-07-28 23:17:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmanlupp -2025-07-28 23:17:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmanlupp succeeded, jobid=5952135 -2025-07-28 23:23:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952135, in state QUEUED (PENDING) -2025-07-28 23:23:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state QUEUED (PENDING) -2025-07-28 23:23:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen000, jobid=5952119, in state QUEUED (PENDING) -2025-07-28 23:23:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen001, jobid=5952120, in state QUEUED (PENDING) -2025-07-28 23:23:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen002, jobid=5952121, in state QUEUED (PENDING) -2025-07-28 23:23:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_esfc, jobid=5952122, in state QUEUED (PENDING) -2025-07-28 23:23:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952099, in state QUEUED (PENDING) -2025-07-28 23:29:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952135, in state QUEUED (PENDING) -2025-07-28 23:29:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state RUNNING (RUNNING) -2025-07-28 23:29:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen000, jobid=5952119, in state QUEUED (PENDING) -2025-07-28 23:29:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen001, jobid=5952120, in state QUEUED (PENDING) -2025-07-28 23:29:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen002, jobid=5952121, in state QUEUED (PENDING) -2025-07-28 23:29:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_esfc, jobid=5952122, in state QUEUED (PENDING) -2025-07-28 23:29:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952099, in state QUEUED (PENDING) -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952135, in state QUEUED (PENDING) -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state RUNNING (RUNNING) -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen000, jobid=5952119, in state QUEUED (PENDING) -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen001, jobid=5952120, in state QUEUED (PENDING) -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen002, jobid=5952121, in state QUEUED (PENDING) -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_esfc, jobid=5952122, in state QUEUED (PENDING) -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952099, in state QUEUED (PENDING) -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f006 -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f000-f002 -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f003-f005 -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f006-f008 -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f000-f002 -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f003-f005 -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f006-f008 -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f009-f011 -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f012-f014 -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f006 succeeded, jobid=5952211 -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f000-f002 succeeded, jobid=5952212 -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f003-f005 succeeded, jobid=5952213 -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f006-f008 succeeded, jobid=5952214 -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f000-f002 succeeded, jobid=5952215 -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f003-f005 succeeded, jobid=5952216 -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f006-f008 succeeded, jobid=5952217 -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f009-f011 succeeded, jobid=5952218 -2025-07-28 23:35:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f012-f014 succeeded, jobid=5952219 -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952135, in state QUEUED (PENDING) -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state RUNNING (RUNNING) -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f006, jobid=5952211, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952212, in state QUEUED (PENDING) -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952213, in state QUEUED (PENDING) -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952214, in state QUEUED (PENDING) -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5952215, in state QUEUED (PENDING) -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5952216, in state QUEUED (PENDING) -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5952217, in state QUEUED (PENDING) -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5952218, in state QUEUED (PENDING) -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5952219, in state QUEUED (PENDING) -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen000, jobid=5952119, in state QUEUED (PENDING) -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen001, jobid=5952120, in state QUEUED (PENDING) -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen002, jobid=5952121, in state QUEUED (PENDING) -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_esfc, jobid=5952122, in state QUEUED (PENDING) -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952099, in state QUEUED (PENDING) -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f006 -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f012 -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f018 -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f009-f011 -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f012-f014 -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f015-f017 -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f018-f020 -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f021-f023 -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f015-f017 -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f018-f020 -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f021-f023, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f006 succeeded, jobid=5952241 -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f012 succeeded, jobid=5952242 -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f018 succeeded, jobid=5952243 -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f009-f011 succeeded, jobid=5952244 -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f012-f014 succeeded, jobid=5952245 -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f015-f017 succeeded, jobid=5952246 -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f018-f020 succeeded, jobid=5952247 -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f021-f023 succeeded, jobid=5952248 -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f015-f017 succeeded, jobid=5952249 -2025-07-28 23:41:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f018-f020 succeeded, jobid=5952250 -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952135, in state QUEUED (PENDING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state RUNNING (RUNNING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f006, jobid=5952241, in state SUCCEEDED (COMPLETED), ran for 22.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f012, jobid=5952242, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f018, jobid=5952243, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952212, in state QUEUED (PENDING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952213, in state QUEUED (PENDING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952214, in state QUEUED (PENDING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952244, in state QUEUED (PENDING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952245, in state QUEUED (PENDING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952246, in state QUEUED (PENDING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952247, in state QUEUED (PENDING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952248, in state QUEUED (PENDING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5952215, in state QUEUED (PENDING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5952216, in state QUEUED (PENDING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5952217, in state QUEUED (PENDING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5952218, in state QUEUED (PENDING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5952219, in state QUEUED (PENDING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5952249, in state QUEUED (PENDING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5952250, in state QUEUED (PENDING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen000, jobid=5952119, in state QUEUED (PENDING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen001, jobid=5952120, in state QUEUED (PENDING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen002, jobid=5952121, in state QUEUED (PENDING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_esfc, jobid=5952122, in state QUEUED (PENDING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952099, in state QUEUED (PENDING) -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f012 -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f018 -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f024 -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f021-f023, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f012 succeeded, jobid=5952269 -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f018 succeeded, jobid=5952270 -2025-07-28 23:47:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f024 succeeded, jobid=5952271 -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952135, in state QUEUED (PENDING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state RUNNING (RUNNING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f012, jobid=5952269, in state SUCCEEDED (COMPLETED), ran for 24.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f018, jobid=5952270, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f024, jobid=5952271, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952212, in state QUEUED (PENDING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952213, in state QUEUED (PENDING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952214, in state QUEUED (PENDING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952244, in state QUEUED (PENDING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952245, in state QUEUED (PENDING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952246, in state QUEUED (PENDING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952247, in state QUEUED (PENDING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952248, in state QUEUED (PENDING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5952215, in state QUEUED (PENDING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5952216, in state QUEUED (PENDING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5952217, in state QUEUED (PENDING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5952218, in state QUEUED (PENDING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5952219, in state QUEUED (PENDING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5952249, in state QUEUED (PENDING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5952250, in state QUEUED (PENDING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen000, jobid=5952119, in state QUEUED (PENDING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen001, jobid=5952120, in state QUEUED (PENDING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen002, jobid=5952121, in state QUEUED (PENDING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_esfc, jobid=5952122, in state QUEUED (PENDING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952099, in state QUEUED (PENDING) -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f030 -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f036 -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f024 -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f021-f023, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f030 succeeded, jobid=5952281 -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f036 succeeded, jobid=5952282 -2025-07-28 23:53:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f024 succeeded, jobid=5952283 -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952135, in state QUEUED (PENDING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state RUNNING (RUNNING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f030, jobid=5952281, in state SUCCEEDED (COMPLETED), ran for 24.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f036, jobid=5952282, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f024, jobid=5952283, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952212, in state QUEUED (PENDING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952213, in state QUEUED (PENDING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952214, in state QUEUED (PENDING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952244, in state QUEUED (PENDING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952245, in state QUEUED (PENDING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952246, in state QUEUED (PENDING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952247, in state QUEUED (PENDING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952248, in state QUEUED (PENDING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5952215, in state QUEUED (PENDING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5952216, in state QUEUED (PENDING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5952217, in state QUEUED (PENDING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5952218, in state QUEUED (PENDING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5952219, in state QUEUED (PENDING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5952249, in state QUEUED (PENDING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5952250, in state QUEUED (PENDING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen000, jobid=5952119, in state QUEUED (PENDING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen001, jobid=5952120, in state QUEUED (PENDING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen002, jobid=5952121, in state QUEUED (PENDING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_esfc, jobid=5952122, in state QUEUED (PENDING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952099, in state QUEUED (PENDING) -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f042 -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f048 -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f054 -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f021-f023, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f042 succeeded, jobid=5952300 -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f048 succeeded, jobid=5952301 -2025-07-28 23:59:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f054 succeeded, jobid=5952302 -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952135, in state QUEUED (PENDING) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state RUNNING (RUNNING) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f042, jobid=5952300, in state SUCCEEDED (COMPLETED), ran for 23.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f048, jobid=5952301, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f054, jobid=5952302, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952212, in state QUEUED (PENDING) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952213, in state QUEUED (PENDING) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952214, in state QUEUED (PENDING) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952244, in state QUEUED (PENDING) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952245, in state QUEUED (PENDING) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952246, in state QUEUED (PENDING) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952247, in state QUEUED (PENDING) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952248, in state QUEUED (PENDING) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5952215, in state QUEUED (PENDING) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5952216, in state QUEUED (PENDING) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5952217, in state QUEUED (PENDING) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5952218, in state QUEUED (PENDING) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5952219, in state QUEUED (PENDING) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5952249, in state QUEUED (PENDING) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5952250, in state QUEUED (PENDING) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen000, jobid=5952119, in state SUCCEEDED (COMPLETED), ran for 43.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen001, jobid=5952120, in state QUEUED (PENDING) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen002, jobid=5952121, in state QUEUED (PENDING) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_esfc, jobid=5952122, in state QUEUED (PENDING) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952099, in state QUEUED (PENDING) -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f060 -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f066 -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f030 -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f036 -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f021-f023, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f060 succeeded, jobid=5952348 -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f066 succeeded, jobid=5952349 -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f030 succeeded, jobid=5952350 -2025-07-29 00:05:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f036 succeeded, jobid=5952351 -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952135, in state QUEUED (PENDING) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state RUNNING (RUNNING) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f060, jobid=5952348, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f066, jobid=5952349, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f030, jobid=5952350, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f036, jobid=5952351, in state SUCCEEDED (COMPLETED), ran for 53.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952212, in state QUEUED (PENDING) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952213, in state QUEUED (PENDING) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952214, in state QUEUED (PENDING) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952244, in state QUEUED (PENDING) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952245, in state QUEUED (PENDING) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952246, in state QUEUED (PENDING) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952247, in state QUEUED (PENDING) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952248, in state QUEUED (PENDING) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5952215, in state QUEUED (PENDING) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5952216, in state QUEUED (PENDING) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5952217, in state QUEUED (PENDING) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5952218, in state QUEUED (PENDING) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5952219, in state QUEUED (PENDING) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5952249, in state QUEUED (PENDING) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5952250, in state QUEUED (PENDING) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen001, jobid=5952120, in state SUCCEEDED (COMPLETED), ran for 43.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen002, jobid=5952121, in state SUCCEEDED (COMPLETED), ran for 43.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_esfc, jobid=5952122, in state SUCCEEDED (COMPLETED), ran for 123.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952099, in state SUCCEEDED (COMPLETED), ran for 62.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f072 -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f078 -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f042 -2025-07-29 00:11:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f048 -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f054 -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f060 -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f066 -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f072 -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f021-f023, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit enkfgfs_earc_vrfy, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_analcalc, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_fcst_seg0, because maximum global task throttle of 25 will be violated. -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f072 succeeded, jobid=5952408 -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f078 succeeded, jobid=5952409 -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f042 succeeded, jobid=5952410 -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f048 succeeded, jobid=5952411 -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f054 succeeded, jobid=5952412 -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f060 succeeded, jobid=5952413 -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f066 succeeded, jobid=5952414 -2025-07-29 00:11:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f072 succeeded, jobid=5952415 -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952135, in state QUEUED (PENDING) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state RUNNING (RUNNING) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f072, jobid=5952408, in state SUCCEEDED (COMPLETED), ran for 25.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f078, jobid=5952409, in state SUCCEEDED (COMPLETED), ran for 25.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f042, jobid=5952410, in state SUCCEEDED (COMPLETED), ran for 25.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f048, jobid=5952411, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f054, jobid=5952412, in state SUCCEEDED (COMPLETED), ran for 33.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f060, jobid=5952413, in state SUCCEEDED (COMPLETED), ran for 33.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f066, jobid=5952414, in state SUCCEEDED (COMPLETED), ran for 33.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f072, jobid=5952415, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952212, in state QUEUED (PENDING) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952213, in state QUEUED (PENDING) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952214, in state QUEUED (PENDING) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952244, in state QUEUED (PENDING) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952245, in state QUEUED (PENDING) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952246, in state QUEUED (PENDING) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952247, in state QUEUED (PENDING) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952248, in state QUEUED (PENDING) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5952215, in state QUEUED (PENDING) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5952216, in state QUEUED (PENDING) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5952217, in state QUEUED (PENDING) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5952218, in state QUEUED (PENDING) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5952219, in state QUEUED (PENDING) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5952249, in state QUEUED (PENDING) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5952250, in state QUEUED (PENDING) -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f084 -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f090 -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f096 -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f078 -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f084 -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f090 -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f096 -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f102 -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f021-f023, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit enkfgfs_earc_vrfy, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_analcalc, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_fcst_seg0, because maximum global task throttle of 25 will be violated. -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f084 succeeded, jobid=5952425 -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f090 succeeded, jobid=5952426 -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f096 succeeded, jobid=5952427 -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f078 succeeded, jobid=5952428 -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f084 succeeded, jobid=5952429 -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f090 succeeded, jobid=5952430 -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f096 succeeded, jobid=5952431 -2025-07-29 00:17:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f102 succeeded, jobid=5952432 -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5952135, in state SUCCEEDED (COMPLETED), ran for 44.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951964, in state SUCCEEDED (COMPLETED), ran for 3349.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f084, jobid=5952425, in state SUCCEEDED (COMPLETED), ran for 25.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f090, jobid=5952426, in state SUCCEEDED (COMPLETED), ran for 25.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f096, jobid=5952427, in state SUCCEEDED (COMPLETED), ran for 25.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f078, jobid=5952428, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f084, jobid=5952429, in state SUCCEEDED (COMPLETED), ran for 20.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f090, jobid=5952430, in state SUCCEEDED (COMPLETED), ran for 20.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f096, jobid=5952431, in state SUCCEEDED (COMPLETED), ran for 20.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f102, jobid=5952432, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952212, in state QUEUED (PENDING) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952213, in state QUEUED (PENDING) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952214, in state QUEUED (PENDING) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952244, in state QUEUED (PENDING) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952245, in state QUEUED (PENDING) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952246, in state QUEUED (PENDING) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952247, in state QUEUED (PENDING) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952248, in state QUEUED (PENDING) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5952215, in state QUEUED (PENDING) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5952216, in state QUEUED (PENDING) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5952217, in state QUEUED (PENDING) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5952218, in state QUEUED (PENDING) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5952219, in state QUEUED (PENDING) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5952249, in state QUEUED (PENDING) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5952250, in state QUEUED (PENDING) -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmanlprod -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f102 -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f108 -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f114 -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f120 -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f108 -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f114 -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f120 -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f024-f026 -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f027-f029 -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f021-f023, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f024-f026, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit enkfgfs_earc_vrfy, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_analcalc, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_fcst_seg0, because maximum global task throttle of 25 will be violated. -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmanlprod succeeded, jobid=5952447 -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f102 succeeded, jobid=5952448 -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f108 succeeded, jobid=5952449 -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f114 succeeded, jobid=5952450 -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f120 succeeded, jobid=5952451 -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f108 succeeded, jobid=5952452 -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f114 succeeded, jobid=5952453 -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f120 succeeded, jobid=5952454 -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f024-f026 succeeded, jobid=5952455 -2025-07-29 00:23:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f027-f029 succeeded, jobid=5952456 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlprod, jobid=5952447, in state QUEUED (PENDING) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f102, jobid=5952448, in state SUCCEEDED (COMPLETED), ran for 24.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f108, jobid=5952449, in state SUCCEEDED (COMPLETED), ran for 24.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f114, jobid=5952450, in state SUCCEEDED (COMPLETED), ran for 24.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f120, jobid=5952451, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f108, jobid=5952452, in state SUCCEEDED (COMPLETED), ran for 20.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f114, jobid=5952453, in state SUCCEEDED (COMPLETED), ran for 20.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f120, jobid=5952454, in state SUCCEEDED (COMPLETED), ran for 20.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952212, in state SUCCEEDED (COMPLETED), ran for 273.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952213, in state SUCCEEDED (COMPLETED), ran for 274.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952214, in state SUCCEEDED (COMPLETED), ran for 271.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952244, in state RUNNING (RUNNING) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952245, in state RUNNING (RUNNING) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952246, in state RUNNING (RUNNING) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952247, in state RUNNING (RUNNING) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952248, in state RUNNING (RUNNING) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952455, in state QUEUED (PENDING) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952456, in state QUEUED (PENDING) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5952215, in state SUCCEEDED (COMPLETED), ran for 78.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5952216, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5952217, in state SUCCEEDED (COMPLETED), ran for 74.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5952218, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5952219, in state SUCCEEDED (COMPLETED), ran for 75.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5952249, in state SUCCEEDED (COMPLETED), ran for 75.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5952250, in state SUCCEEDED (COMPLETED), ran for 75.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f030-f032 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f033-f035 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f036-f038 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f039-f041 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f042-f044 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f045-f047 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f048-f054 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f057-f063 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f066-f072 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f075-f081 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f084-f090 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f093-f099 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f102-f108 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f111-f114 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f117-f120 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f021-f023 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f024-f026 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f027-f029, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f030-f032, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit enkfgfs_earc_vrfy, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_analcalc, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_fcst_seg0, because maximum global task throttle of 25 will be violated. -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f030-f032 succeeded, jobid=5952501 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f033-f035 succeeded, jobid=5952502 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f036-f038 succeeded, jobid=5952503 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f039-f041 succeeded, jobid=5952504 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f042-f044 succeeded, jobid=5952505 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f045-f047 succeeded, jobid=5952506 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f048-f054 succeeded, jobid=5952507 -2025-07-29 00:29:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f057-f063 succeeded, jobid=5952508 -2025-07-29 00:30:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f066-f072 succeeded, jobid=5952509 -2025-07-29 00:30:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f075-f081 succeeded, jobid=5952510 -2025-07-29 00:30:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f084-f090 succeeded, jobid=5952511 -2025-07-29 00:30:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f093-f099 succeeded, jobid=5952512 -2025-07-29 00:30:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f102-f108 succeeded, jobid=5952513 -2025-07-29 00:30:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f111-f114 succeeded, jobid=5952514 -2025-07-29 00:30:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f117-f120 succeeded, jobid=5952515 -2025-07-29 00:30:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f021-f023 succeeded, jobid=5952516 -2025-07-29 00:30:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f024-f026 succeeded, jobid=5952517 -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlprod, jobid=5952447, in state QUEUED (PENDING) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952244, in state SUCCEEDED (COMPLETED), ran for 277.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952245, in state SUCCEEDED (COMPLETED), ran for 272.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952246, in state SUCCEEDED (COMPLETED), ran for 272.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952247, in state SUCCEEDED (COMPLETED), ran for 275.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952248, in state SUCCEEDED (COMPLETED), ran for 271.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952455, in state QUEUED (PENDING) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952456, in state QUEUED (PENDING) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952501, in state QUEUED (PENDING) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952502, in state QUEUED (PENDING) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952503, in state QUEUED (PENDING) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952504, in state QUEUED (PENDING) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952505, in state QUEUED (PENDING) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952506, in state QUEUED (PENDING) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952507, in state QUEUED (PENDING) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952508, in state QUEUED (PENDING) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952509, in state QUEUED (PENDING) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952510, in state QUEUED (PENDING) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952511, in state QUEUED (PENDING) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952512, in state QUEUED (PENDING) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952513, in state QUEUED (PENDING) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5952514, in state QUEUED (PENDING) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5952515, in state QUEUED (PENDING) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5952516, in state QUEUED (PENDING) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5952517, in state QUEUED (PENDING) -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f027-f029 -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f030-f032 -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f033-f035 -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f036-f038 -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f039-f041 -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit enkfgfs_earc_vrfy, because maximum global task throttle of 25 will be violated. -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_analcalc, because maximum global task throttle of 25 will be violated. -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_fcst_seg0, because maximum global task throttle of 25 will be violated. -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f027-f029 succeeded, jobid=5952556 -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f030-f032 succeeded, jobid=5952557 -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f033-f035 succeeded, jobid=5952558 -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f036-f038 succeeded, jobid=5952559 -2025-07-29 00:36:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f039-f041 succeeded, jobid=5952560 -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlprod, jobid=5952447, in state SUCCEEDED (COMPLETED), ran for 105.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952455, in state RUNNING (RUNNING) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952456, in state RUNNING (RUNNING) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952501, in state QUEUED (PENDING) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952502, in state QUEUED (PENDING) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952503, in state QUEUED (PENDING) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952504, in state QUEUED (PENDING) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952505, in state QUEUED (PENDING) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952506, in state QUEUED (PENDING) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952507, in state QUEUED (PENDING) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952508, in state QUEUED (PENDING) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952509, in state QUEUED (PENDING) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952510, in state QUEUED (PENDING) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952511, in state QUEUED (PENDING) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952512, in state QUEUED (PENDING) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952513, in state QUEUED (PENDING) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5952514, in state QUEUED (PENDING) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5952515, in state QUEUED (PENDING) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5952516, in state SUCCEEDED (COMPLETED), ran for 74.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5952517, in state SUCCEEDED (COMPLETED), ran for 74.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5952556, in state QUEUED (PENDING) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5952557, in state QUEUED (PENDING) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f033-f035, jobid=5952558, in state QUEUED (PENDING) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f036-f038, jobid=5952559, in state QUEUED (PENDING) -2025-07-29 00:42:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f039-f041, jobid=5952560, in state QUEUED (PENDING) -2025-07-29 00:42:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f042-f044 -2025-07-29 00:42:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f045-f047 -2025-07-29 00:42:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f048-f054 -2025-07-29 00:42:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-29 00:42:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-29 00:42:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-29 00:42:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-29 00:42:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-29 00:42:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-29 00:42:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-29 00:42:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-29 00:42:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:42:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-29 00:42:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:42:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit enkfgfs_earc_vrfy, because maximum global task throttle of 25 will be violated. -2025-07-29 00:42:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_analcalc, because maximum global task throttle of 25 will be violated. -2025-07-29 00:42:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_fcst_seg0, because maximum global task throttle of 25 will be violated. -2025-07-29 00:42:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f042-f044 succeeded, jobid=5952595 -2025-07-29 00:42:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_wavepostsbs_f045-f047 is pending at druby://130.18.14.151:41195 -2025-07-29 00:42:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_wavepostsbs_f048-f054 is pending at druby://130.18.14.151:41195 -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_wavepostsbs_f045-f047 is success, jobid=5952596 -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_wavepostsbs_f048-f054 is success, jobid=5952597 -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952455, in state SUCCEEDED (COMPLETED), ran for 275.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952456, in state SUCCEEDED (COMPLETED), ran for 270.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952501, in state RUNNING (RUNNING) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952502, in state RUNNING (RUNNING) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952503, in state RUNNING (RUNNING) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952504, in state RUNNING (RUNNING) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952505, in state RUNNING (RUNNING) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952506, in state RUNNING (RUNNING) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952507, in state RUNNING (RUNNING) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952508, in state RUNNING (RUNNING) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952509, in state RUNNING (RUNNING) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952510, in state RUNNING (RUNNING) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952511, in state RUNNING (RUNNING) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952512, in state RUNNING (RUNNING) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952513, in state RUNNING (RUNNING) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5952514, in state RUNNING (RUNNING) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5952515, in state RUNNING (RUNNING) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5952556, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5952557, in state SUCCEEDED (COMPLETED), ran for 70.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f033-f035, jobid=5952558, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f036-f038, jobid=5952559, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f039-f041, jobid=5952560, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f042-f044, jobid=5952595, in state QUEUED (PENDING) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f045-f047, jobid=5952596, in state QUEUED (PENDING) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f048-f054, jobid=5952597, in state QUEUED (PENDING) -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f057-f063 -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f066-f072 -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f075-f081 -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f084-f090 -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f093-f099 -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f102-f108 -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f111-f114 -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit enkfgfs_earc_vrfy, because maximum global task throttle of 25 will be violated. -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_analcalc, because maximum global task throttle of 25 will be violated. -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_fcst_seg0, because maximum global task throttle of 25 will be violated. -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f057-f063 succeeded, jobid=5952642 -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f066-f072 succeeded, jobid=5952643 -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f075-f081 succeeded, jobid=5952644 -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f084-f090 succeeded, jobid=5952645 -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f093-f099 succeeded, jobid=5952646 -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f102-f108 succeeded, jobid=5952647 -2025-07-29 00:48:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f111-f114 succeeded, jobid=5952648 -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952501, in state SUCCEEDED (COMPLETED), ran for 280.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952502, in state SUCCEEDED (COMPLETED), ran for 276.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952503, in state SUCCEEDED (COMPLETED), ran for 271.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952504, in state SUCCEEDED (COMPLETED), ran for 273.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952505, in state SUCCEEDED (COMPLETED), ran for 271.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952506, in state SUCCEEDED (COMPLETED), ran for 275.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952507, in state SUCCEEDED (COMPLETED), ran for 322.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952508, in state SUCCEEDED (COMPLETED), ran for 316.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952509, in state SUCCEEDED (COMPLETED), ran for 333.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952510, in state SUCCEEDED (COMPLETED), ran for 337.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952511, in state SUCCEEDED (COMPLETED), ran for 323.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952512, in state SUCCEEDED (COMPLETED), ran for 320.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952513, in state SUCCEEDED (COMPLETED), ran for 330.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5952514, in state SUCCEEDED (COMPLETED), ran for 213.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5952515, in state SUCCEEDED (COMPLETED), ran for 218.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f042-f044, jobid=5952595, in state QUEUED (PENDING) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f045-f047, jobid=5952596, in state QUEUED (PENDING) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f048-f054, jobid=5952597, in state QUEUED (PENDING) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f057-f063, jobid=5952642, in state QUEUED (PENDING) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f066-f072, jobid=5952643, in state QUEUED (PENDING) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f075-f081, jobid=5952644, in state QUEUED (PENDING) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f084-f090, jobid=5952645, in state QUEUED (PENDING) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f093-f099, jobid=5952646, in state QUEUED (PENDING) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f102-f108, jobid=5952647, in state QUEUED (PENDING) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f111-f114, jobid=5952648, in state QUEUED (PENDING) -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_tracker -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_genesis -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f117-f120 -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostbndpnt -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostbndpntbll -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostpnt -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgfs_earc_vrfy -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_analcalc -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_fcst_seg0 -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_tracker succeeded, jobid=5952664 -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_genesis succeeded, jobid=5952665 -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f117-f120 succeeded, jobid=5952666 -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostbndpnt succeeded, jobid=5952667 -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostbndpntbll succeeded, jobid=5952668 -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostpnt succeeded, jobid=5952669 -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgfs_earc_vrfy succeeded, jobid=5952670 -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_analcalc succeeded, jobid=5952671 -2025-07-29 00:54:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_fcst_seg0 succeeded, jobid=5952672 -2025-07-29 01:00:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_tracker, jobid=5952664, in state SUCCEEDED (COMPLETED), ran for 71.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:00:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_genesis, jobid=5952665, in state RUNNING (RUNNING) -2025-07-29 01:00:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f042-f044, jobid=5952595, in state RUNNING (RUNNING) -2025-07-29 01:00:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f045-f047, jobid=5952596, in state RUNNING (RUNNING) -2025-07-29 01:00:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f048-f054, jobid=5952597, in state RUNNING (RUNNING) -2025-07-29 01:00:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f057-f063, jobid=5952642, in state QUEUED (PENDING) -2025-07-29 01:00:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f066-f072, jobid=5952643, in state QUEUED (PENDING) -2025-07-29 01:00:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f075-f081, jobid=5952644, in state QUEUED (PENDING) -2025-07-29 01:00:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f084-f090, jobid=5952645, in state QUEUED (PENDING) -2025-07-29 01:00:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f093-f099, jobid=5952646, in state QUEUED (PENDING) -2025-07-29 01:00:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f102-f108, jobid=5952647, in state QUEUED (PENDING) -2025-07-29 01:00:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f111-f114, jobid=5952648, in state QUEUED (PENDING) -2025-07-29 01:00:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f117-f120, jobid=5952666, in state QUEUED (PENDING) -2025-07-29 01:00:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostbndpnt, jobid=5952667, in state SUCCEEDED (COMPLETED), ran for 165.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:00:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostbndpntbll, jobid=5952668, in state SUCCEEDED (COMPLETED), ran for 80.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:00:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostpnt, jobid=5952669, in state RUNNING (RUNNING) -2025-07-29 01:00:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_earc_vrfy, jobid=5952670, in state SUCCEEDED (COMPLETED), ran for 15.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:00:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952671, in state QUEUED (PENDING) -2025-07-29 01:00:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952672, in state QUEUED (PENDING) -2025-07-29 01:00:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgfs_cleanup -2025-07-29 01:00:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgfs_cleanup succeeded, jobid=5952683 -2025-07-29 01:06:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_genesis, jobid=5952665, in state SUCCEEDED (COMPLETED), ran for 496.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:06:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f042-f044, jobid=5952595, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:06:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f045-f047, jobid=5952596, in state SUCCEEDED (COMPLETED), ran for 77.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:06:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f048-f054, jobid=5952597, in state SUCCEEDED (COMPLETED), ran for 74.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:06:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f057-f063, jobid=5952642, in state QUEUED (PENDING) -2025-07-29 01:06:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f066-f072, jobid=5952643, in state QUEUED (PENDING) -2025-07-29 01:06:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f075-f081, jobid=5952644, in state QUEUED (PENDING) -2025-07-29 01:06:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f084-f090, jobid=5952645, in state QUEUED (PENDING) -2025-07-29 01:06:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f093-f099, jobid=5952646, in state QUEUED (PENDING) -2025-07-29 01:06:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f102-f108, jobid=5952647, in state QUEUED (PENDING) -2025-07-29 01:06:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f111-f114, jobid=5952648, in state QUEUED (PENDING) -2025-07-29 01:06:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f117-f120, jobid=5952666, in state QUEUED (PENDING) -2025-07-29 01:06:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostpnt, jobid=5952669, in state SUCCEEDED (COMPLETED), ran for 98.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:06:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_cleanup, jobid=5952683, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:06:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952671, in state QUEUED (PENDING) -2025-07-29 01:06:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952672, in state QUEUED (PENDING) -2025-07-29 01:12:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f057-f063, jobid=5952642, in state SUCCEEDED (COMPLETED), ran for 75.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:12:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f066-f072, jobid=5952643, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:12:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f075-f081, jobid=5952644, in state QUEUED (PENDING) -2025-07-29 01:12:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f084-f090, jobid=5952645, in state QUEUED (PENDING) -2025-07-29 01:12:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f093-f099, jobid=5952646, in state QUEUED (PENDING) -2025-07-29 01:12:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f102-f108, jobid=5952647, in state QUEUED (PENDING) -2025-07-29 01:12:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f111-f114, jobid=5952648, in state QUEUED (PENDING) -2025-07-29 01:12:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f117-f120, jobid=5952666, in state QUEUED (PENDING) -2025-07-29 01:12:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952671, in state QUEUED (PENDING) -2025-07-29 01:12:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952672, in state QUEUED (PENDING) -2025-07-29 01:18:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f075-f081, jobid=5952644, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:18:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f084-f090, jobid=5952645, in state SUCCEEDED (COMPLETED), ran for 71.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:18:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f093-f099, jobid=5952646, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:18:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f102-f108, jobid=5952647, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:18:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f111-f114, jobid=5952648, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:18:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f117-f120, jobid=5952666, in state QUEUED (PENDING) -2025-07-29 01:18:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952671, in state SUCCEEDED (COMPLETED), ran for 71.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:18:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952672, in state QUEUED (PENDING) -2025-07-29 01:18:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmanlupp -2025-07-29 01:18:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_ecen000 -2025-07-29 01:18:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_ecen001 -2025-07-29 01:18:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_ecen002 -2025-07-29 01:18:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_esfc -2025-07-29 01:18:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmanlupp succeeded, jobid=5952727 -2025-07-29 01:18:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_ecen000 succeeded, jobid=5952728 -2025-07-29 01:18:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_ecen001 succeeded, jobid=5952729 -2025-07-29 01:18:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_ecen002 succeeded, jobid=5952730 -2025-07-29 01:18:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_esfc succeeded, jobid=5952731 -2025-07-29 01:24:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f117-f120, jobid=5952666, in state SUCCEEDED (COMPLETED), ran for 48.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:24:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952727, in state SUCCEEDED (COMPLETED), ran for 45.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:24:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952672, in state RUNNING (RUNNING) -2025-07-29 01:24:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen000, jobid=5952728, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:24:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen001, jobid=5952729, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:24:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen002, jobid=5952730, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:24:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_esfc, jobid=5952731, in state QUEUED (PENDING) -2025-07-29 01:24:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_arch_vrfy -2025-07-29 01:24:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmanlprod -2025-07-29 01:24:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostsbs_f000 -2025-07-29 01:24:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostsbs_f001 -2025-07-29 01:24:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostsbs_f002 -2025-07-29 01:24:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostsbs_f003 -2025-07-29 01:24:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostsbs_f004 -2025-07-29 01:24:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_arch_vrfy succeeded, jobid=5952740 -2025-07-29 01:24:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmanlprod succeeded, jobid=5952741 -2025-07-29 01:24:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostsbs_f000 succeeded, jobid=5952742 -2025-07-29 01:24:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostsbs_f001 succeeded, jobid=5952743 -2025-07-29 01:24:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostsbs_f002 succeeded, jobid=5952744 -2025-07-29 01:24:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostsbs_f003 succeeded, jobid=5952745 -2025-07-29 01:24:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostsbs_f004 succeeded, jobid=5952746 -2025-07-29 01:30:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_arch_vrfy, jobid=5952740, in state SUCCEEDED (COMPLETED), ran for 52.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:30:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlprod, jobid=5952741, in state SUCCEEDED (COMPLETED), ran for 63.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:30:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952672, in state SUCCEEDED (COMPLETED), ran for 469.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:30:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f000, jobid=5952742, in state SUCCEEDED (COMPLETED), ran for 25.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:30:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f001, jobid=5952743, in state SUCCEEDED (COMPLETED), ran for 25.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:30:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f002, jobid=5952744, in state SUCCEEDED (COMPLETED), ran for 24.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:30:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f003, jobid=5952745, in state SUCCEEDED (COMPLETED), ran for 25.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:30:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f004, jobid=5952746, in state SUCCEEDED (COMPLETED), ran for 25.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:30:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_esfc, jobid=5952731, in state SUCCEEDED (COMPLETED), ran for 143.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:30:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_metpg2g1 -2025-07-29 01:30:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f000 -2025-07-29 01:30:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f001 -2025-07-29 01:30:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f002 -2025-07-29 01:30:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f003 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f004 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f005 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f006 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f007 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f008 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f009 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostsbs_f005 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostsbs_f006 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostsbs_f007 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostsbs_f008 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostsbs_f009 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostbndpnt -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostbndpntbll -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_wavepostpnt -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_fcst_mem001 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_fcst_mem002 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_metpg2g1 succeeded, jobid=5952767 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f000 succeeded, jobid=5952768 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f001 succeeded, jobid=5952769 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f002 succeeded, jobid=5952770 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f003 succeeded, jobid=5952771 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f004 succeeded, jobid=5952772 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f005 succeeded, jobid=5952773 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f006 succeeded, jobid=5952774 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f007 succeeded, jobid=5952775 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f008 succeeded, jobid=5952776 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f009 succeeded, jobid=5952777 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostsbs_f005 succeeded, jobid=5952778 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostsbs_f006 succeeded, jobid=5952779 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostsbs_f007 succeeded, jobid=5952780 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostsbs_f008 succeeded, jobid=5952781 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostsbs_f009 succeeded, jobid=5952782 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostbndpnt succeeded, jobid=5952783 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostbndpntbll succeeded, jobid=5952784 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_wavepostpnt succeeded, jobid=5952785 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_fcst_mem001 succeeded, jobid=5952786 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_fcst_mem002 succeeded, jobid=5952787 -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_metpg2g1, jobid=5952767, in state RUNNING (RUNNING) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952768, in state SUCCEEDED (COMPLETED), ran for 66.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5952769, in state SUCCEEDED (COMPLETED), ran for 68.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5952770, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952771, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f004, jobid=5952772, in state SUCCEEDED (COMPLETED), ran for 66.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f005, jobid=5952773, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952774, in state SUCCEEDED (COMPLETED), ran for 65.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f007, jobid=5952775, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f008, jobid=5952776, in state SUCCEEDED (COMPLETED), ran for 68.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952777, in state SUCCEEDED (COMPLETED), ran for 68.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f005, jobid=5952778, in state SUCCEEDED (COMPLETED), ran for 27.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f006, jobid=5952779, in state SUCCEEDED (COMPLETED), ran for 27.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f007, jobid=5952780, in state SUCCEEDED (COMPLETED), ran for 27.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f008, jobid=5952781, in state SUCCEEDED (COMPLETED), ran for 27.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostsbs_f009, jobid=5952782, in state SUCCEEDED (COMPLETED), ran for 27.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostbndpnt, jobid=5952783, in state SUCCEEDED (COMPLETED), ran for 33.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostbndpntbll, jobid=5952784, in state SUCCEEDED (COMPLETED), ran for 30.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_wavepostpnt, jobid=5952785, in state SUCCEEDED (COMPLETED), ran for 26.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem001, jobid=5952786, in state RUNNING (RUNNING) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem002, jobid=5952787, in state RUNNING (RUNNING) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_fit2obs -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_fit2obs succeeded, jobid=5952802 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_metpg2g1, jobid=5952767, in state SUCCEEDED (COMPLETED), ran for 109.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fit2obs, jobid=5952802, in state SUCCEEDED (COMPLETED), ran for 16.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem001, jobid=5952786, in state SUCCEEDED (COMPLETED), ran for 285.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_fcst_mem002, jobid=5952787, in state SUCCEEDED (COMPLETED), ran for 285.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_metpg2o1 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_arch_vrfy -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_echgres -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos000 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos001 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos002 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos003 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos004 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos005 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_epos006 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_metpg2o1 succeeded, jobid=5952822 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_arch_vrfy succeeded, jobid=5952823 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_echgres succeeded, jobid=5952824 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos000 succeeded, jobid=5952825 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos001 succeeded, jobid=5952826 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos002 succeeded, jobid=5952827 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos003 succeeded, jobid=5952828 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos004 succeeded, jobid=5952829 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos005 succeeded, jobid=5952830 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_epos006 succeeded, jobid=5952831 -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_metpg2o1, jobid=5952822, in state SUCCEEDED (COMPLETED), ran for 128.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_arch_vrfy, jobid=5952823, in state SUCCEEDED (COMPLETED), ran for 20.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_echgres, jobid=5952824, in state SUCCEEDED (COMPLETED), ran for 29.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos000, jobid=5952825, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos001, jobid=5952826, in state SUCCEEDED (COMPLETED), ran for 33.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos002, jobid=5952827, in state SUCCEEDED (COMPLETED), ran for 33.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos003, jobid=5952828, in state SUCCEEDED (COMPLETED), ran for 37.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos004, jobid=5952829, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos005, jobid=5952830, in state SUCCEEDED (COMPLETED), ran for 37.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_epos006, jobid=5952831, in state SUCCEEDED (COMPLETED), ran for 37.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_metppcp1 -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_cleanup -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_earc_vrfy -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_metppcp1 succeeded, jobid=5952863 -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_cleanup succeeded, jobid=5952864 -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_earc_vrfy succeeded, jobid=5952865 -2025-07-29 01:54:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_metppcp1, jobid=5952863, in state SUCCEEDED (COMPLETED), ran for 28.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:54:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_cleanup, jobid=5952864, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:54:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_earc_vrfy, jobid=5952865, in state SUCCEEDED (COMPLETED), ran for 15.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:54:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_cleanup -2025-07-29 01:54:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_cleanup -2025-07-29 01:54:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_cleanup succeeded, jobid=5952902 -2025-07-29 01:54:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_cleanup succeeded, jobid=5952903 -2025-07-29 02:00:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_cleanup, jobid=5952902, in state SUCCEEDED (COMPLETED), ran for 21.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:00:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_cleanup, jobid=5952903, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:00:54 -0500 :: hercules-login-1.hpc.msstate.edu :: This cycle is complete: Success diff --git a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96C48mx500_S2SW_cyc_gfs_logs_2021122100.log b/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96C48mx500_S2SW_cyc_gfs_logs_2021122100.log deleted file mode 100644 index 7cdfe3d3..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96C48mx500_S2SW_cyc_gfs_logs_2021122100.log +++ /dev/null @@ -1,349 +0,0 @@ -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_prepoceanobs -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_prepoceanobs -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_prepoceanobs succeeded, jobid=5952788 -2025-07-29 01:30:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_prepoceanobs succeeded, jobid=5952789 -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prepoceanobs, jobid=5952788, in state RUNNING (RUNNING) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5952789, in state RUNNING (RUNNING) -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_prep -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_prep -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_prep succeeded, jobid=5952803 -2025-07-29 01:36:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_prep succeeded, jobid=5952804 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5952803, in state SUCCEEDED (COMPLETED), ran for 135.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prepoceanobs, jobid=5952788, in state SUCCEEDED (COMPLETED), ran for 212.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5952804, in state SUCCEEDED (COMPLETED), ran for 141.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prepoceanobs, jobid=5952789, in state SUCCEEDED (COMPLETED), ran for 213.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_marinebmatinit -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_snowanl -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_waveinit -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marinebmatinit -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_snowanl -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_waveinit -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_marinebmatinit succeeded, jobid=5952832 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_snowanl succeeded, jobid=5952833 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_waveinit succeeded, jobid=5952834 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_marinebmatinit succeeded, jobid=5952835 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_snowanl succeeded, jobid=5952836 -2025-07-29 01:42:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_waveinit succeeded, jobid=5952837 -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmatinit, jobid=5952832, in state SUCCEEDED (COMPLETED), ran for 32.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_snowanl, jobid=5952833, in state SUCCEEDED (COMPLETED), ran for 90.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_waveinit, jobid=5952834, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmatinit, jobid=5952835, in state SUCCEEDED (COMPLETED), ran for 32.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_snowanl, jobid=5952836, in state SUCCEEDED (COMPLETED), ran for 90.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_waveinit, jobid=5952837, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_anal -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_marinebmat -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocnanalecen -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgfs_eobs -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgfs_esnowanl -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_anal -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marinebmat -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_ocnanalecen -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_eobs -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_esnowanl -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_anal succeeded, jobid=5952866 -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_marinebmat succeeded, jobid=5952867 -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocnanalecen succeeded, jobid=5952868 -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgfs_eobs succeeded, jobid=5952869 -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgfs_esnowanl succeeded, jobid=5952870 -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_anal succeeded, jobid=5952871 -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_marinebmat succeeded, jobid=5952872 -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_ocnanalecen succeeded, jobid=5952873 -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_eobs succeeded, jobid=5952874 -2025-07-29 01:48:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_esnowanl succeeded, jobid=5952875 -2025-07-29 01:54:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5952866, in state RUNNING (RUNNING) -2025-07-29 01:54:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marinebmat, jobid=5952867, in state SUCCEEDED (COMPLETED), ran for 57.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:54:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocnanalecen, jobid=5952868, in state SUCCEEDED (COMPLETED), ran for 56.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:54:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_eobs, jobid=5952869, in state RUNNING (RUNNING) -2025-07-29 01:54:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_esnowanl, jobid=5952870, in state SUCCEEDED (COMPLETED), ran for 69.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:54:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5952871, in state RUNNING (RUNNING) -2025-07-29 01:54:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marinebmat, jobid=5952872, in state SUCCEEDED (COMPLETED), ran for 57.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:54:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_ocnanalecen, jobid=5952873, in state SUCCEEDED (COMPLETED), ran for 56.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:54:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eobs, jobid=5952874, in state RUNNING (RUNNING) -2025-07-29 01:54:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_esnowanl, jobid=5952875, in state SUCCEEDED (COMPLETED), ran for 69.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:54:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_marineanlinit -2025-07-29 01:54:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marineanlinit -2025-07-29 01:54:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_marineanlinit succeeded, jobid=5952904 -2025-07-29 01:54:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_marineanlinit succeeded, jobid=5952905 -2025-07-29 02:00:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5952866, in state SUCCEEDED (COMPLETED), ran for 522.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:00:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlinit, jobid=5952904, in state SUCCEEDED (COMPLETED), ran for 33.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:00:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_eobs, jobid=5952869, in state SUCCEEDED (COMPLETED), ran for 360.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:00:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5952871, in state RUNNING (RUNNING) -2025-07-29 02:00:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlinit, jobid=5952905, in state SUCCEEDED (COMPLETED), ran for 33.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:00:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eobs, jobid=5952874, in state SUCCEEDED (COMPLETED), ran for 418.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:00:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_marineanlvar -2025-07-29 02:00:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_sfcanl -2025-07-29 02:00:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_vminmon -2025-07-29 02:00:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgfs_ediag -2025-07-29 02:00:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marineanlvar -2025-07-29 02:00:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_ediag -2025-07-29 02:00:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_marineanlvar succeeded, jobid=5952939 -2025-07-29 02:00:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_sfcanl succeeded, jobid=5952940 -2025-07-29 02:00:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_vminmon succeeded, jobid=5952941 -2025-07-29 02:00:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgfs_ediag succeeded, jobid=5952942 -2025-07-29 02:00:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_marineanlvar succeeded, jobid=5952943 -2025-07-29 02:00:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_ediag succeeded, jobid=5952944 -2025-07-29 02:06:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlvar, jobid=5952939, in state SUCCEEDED (COMPLETED), ran for 45.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:06:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_sfcanl, jobid=5952940, in state SUCCEEDED (COMPLETED), ran for 46.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:06:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_vminmon, jobid=5952941, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:06:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ediag, jobid=5952942, in state SUCCEEDED (COMPLETED), ran for 122.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:06:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5952871, in state SUCCEEDED (COMPLETED), ran for 859.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:06:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlvar, jobid=5952943, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:06:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ediag, jobid=5952944, in state RUNNING (RUNNING) -2025-07-29 02:06:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_analcalc -2025-07-29 02:06:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_marineanlchkpt -2025-07-29 02:06:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgfs_eupd -2025-07-29 02:06:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marineanlchkpt -2025-07-29 02:06:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_analdiag -2025-07-29 02:06:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_vminmon -2025-07-29 02:06:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_analcalc succeeded, jobid=5953002 -2025-07-29 02:06:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_marineanlchkpt succeeded, jobid=5953003 -2025-07-29 02:06:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgfs_eupd succeeded, jobid=5953004 -2025-07-29 02:06:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_marineanlchkpt succeeded, jobid=5953005 -2025-07-29 02:06:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_analdiag succeeded, jobid=5953006 -2025-07-29 02:06:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_vminmon succeeded, jobid=5953007 -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5953002, in state SUCCEEDED (COMPLETED), ran for 61.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlchkpt, jobid=5953003, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_eupd, jobid=5953004, in state SUCCEEDED (COMPLETED), ran for 125.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlchkpt, jobid=5953005, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5953006, in state SUCCEEDED (COMPLETED), ran for 122.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_vminmon, jobid=5953007, in state SUCCEEDED (COMPLETED), ran for 16.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ediag, jobid=5952944, in state SUCCEEDED (COMPLETED), ran for 149.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_marineanlfinal -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmanlupp -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgfs_ecen000 -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgfs_ecen001 -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgfs_ecen002 -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgfs_esfc -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_marineanlfinal -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_verfozn -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_verfrad -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_eupd -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_marineanlfinal succeeded, jobid=5953073 -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmanlupp succeeded, jobid=5953074 -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgfs_ecen000 succeeded, jobid=5953075 -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgfs_ecen001 succeeded, jobid=5953076 -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgfs_ecen002 succeeded, jobid=5953077 -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgfs_esfc succeeded, jobid=5953078 -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_marineanlfinal succeeded, jobid=5953079 -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_verfozn succeeded, jobid=5953080 -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_verfrad succeeded, jobid=5953081 -2025-07-29 02:13:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_eupd succeeded, jobid=5953082 -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_marineanlfinal, jobid=5953073, in state SUCCEEDED (COMPLETED), ran for 33.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5953074, in state SUCCEEDED (COMPLETED), ran for 45.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen000, jobid=5953075, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen001, jobid=5953076, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_ecen002, jobid=5953077, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_esfc, jobid=5953078, in state SUCCEEDED (COMPLETED), ran for 122.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_marineanlfinal, jobid=5953079, in state SUCCEEDED (COMPLETED), ran for 34.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfozn, jobid=5953080, in state SUCCEEDED (COMPLETED), ran for 70.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5953081, in state RUNNING (RUNNING) -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_eupd, jobid=5953082, in state SUCCEEDED (COMPLETED), ran for 140.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmanlprod -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_fcst_seg0 -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_anlstat -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgfs_earc_vrfy -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_sfcanl -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_anlstat -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmanlprod succeeded, jobid=5953114 -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_fcst_seg0 succeeded, jobid=5953115 -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_anlstat succeeded, jobid=5953116 -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgfs_earc_vrfy succeeded, jobid=5953117 -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_sfcanl succeeded, jobid=5953118 -2025-07-29 02:19:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_anlstat succeeded, jobid=5953119 -2025-07-29 02:25:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlprod, jobid=5953114, in state SUCCEEDED (COMPLETED), ran for 104.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:25:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5953115, in state RUNNING (RUNNING) -2025-07-29 02:25:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anlstat, jobid=5953116, in state SUCCEEDED (COMPLETED), ran for 42.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:25:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_earc_vrfy, jobid=5953117, in state SUCCEEDED (COMPLETED), ran for 15.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:25:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5953118, in state SUCCEEDED (COMPLETED), ran for 55.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:25:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5953081, in state RUNNING (RUNNING) -2025-07-29 02:25:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anlstat, jobid=5953119, in state SUCCEEDED (COMPLETED), ran for 42.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:25:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f000-f002 -2025-07-29 02:25:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgfs_cleanup -2025-07-29 02:25:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_analcalc -2025-07-29 02:25:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_fcst_seg0 -2025-07-29 02:25:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f000-f002 succeeded, jobid=5953160 -2025-07-29 02:25:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgfs_cleanup succeeded, jobid=5953161 -2025-07-29 02:25:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_analcalc succeeded, jobid=5953162 -2025-07-29 02:25:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_fcst_seg0 succeeded, jobid=5953163 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5953115, in state RUNNING (RUNNING) -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5953160, in state SUCCEEDED (COMPLETED), ran for 63.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgfs_cleanup, jobid=5953161, in state SUCCEEDED (COMPLETED), ran for 21.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5953162, in state SUCCEEDED (COMPLETED), ran for 52.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5953163, in state RUNNING (RUNNING) -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5953081, in state SUCCEEDED (COMPLETED), ran for 836.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f006 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f012 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f000-f002 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f003-f005 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f006-f008 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f009-f011 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f003-f005 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f006-f008 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f009-f011 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f012-f014 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f015-f017 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_ecen000 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_ecen001 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_ecen002 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting enkfgdas_esfc -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f006 succeeded, jobid=5953196 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f012 succeeded, jobid=5953197 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f000-f002 succeeded, jobid=5953198 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f003-f005 succeeded, jobid=5953199 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f006-f008 succeeded, jobid=5953200 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f009-f011 succeeded, jobid=5953201 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f003-f005 succeeded, jobid=5953202 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f006-f008 succeeded, jobid=5953203 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f009-f011 succeeded, jobid=5953204 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f012-f014 succeeded, jobid=5953205 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f015-f017 succeeded, jobid=5953206 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_ecen000 succeeded, jobid=5953207 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_ecen001 succeeded, jobid=5953208 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_ecen002 succeeded, jobid=5953209 -2025-07-29 02:31:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of enkfgdas_esfc succeeded, jobid=5953210 -2025-07-29 02:37:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5953115, in state RUNNING (RUNNING) -2025-07-29 02:37:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f006, jobid=5953196, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:37:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f012, jobid=5953197, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:37:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5953198, in state RUNNING (RUNNING) -2025-07-29 02:37:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5953199, in state RUNNING (RUNNING) -2025-07-29 02:37:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5953200, in state RUNNING (RUNNING) -2025-07-29 02:37:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5953201, in state RUNNING (RUNNING) -2025-07-29 02:37:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5953202, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:37:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5953203, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5953204, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5953205, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5953206, in state SUCCEEDED (COMPLETED), ran for 74.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5953163, in state RUNNING (RUNNING) -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen000, jobid=5953207, in state SUCCEEDED (COMPLETED), ran for 34.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen001, jobid=5953208, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_ecen002, jobid=5953209, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_esfc, jobid=5953210, in state RUNNING (RUNNING) -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f006 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f012 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f018 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f024 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f012-f014 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f015-f017 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f018-f020 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f021-f023 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f024-f026 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f018-f020 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f021-f023 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f024-f026 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f027-f029 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f030-f032 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmanlupp -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f000 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f001 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f002 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f003, because maximum global task throttle of 25 will be violated. -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f004, because maximum global task throttle of 25 will be violated. -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f000, because maximum global task throttle of 25 will be violated. -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f001, because maximum global task throttle of 25 will be violated. -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f002, because maximum global task throttle of 25 will be violated. -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f003, because maximum global task throttle of 25 will be violated. -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f004, because maximum global task throttle of 25 will be violated. -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f005, because maximum global task throttle of 25 will be violated. -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f006, because maximum global task throttle of 25 will be violated. -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f007, because maximum global task throttle of 25 will be violated. -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f008, because maximum global task throttle of 25 will be violated. -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f009, because maximum global task throttle of 25 will be violated. -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f006 succeeded, jobid=5953219 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f012 succeeded, jobid=5953220 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f018 succeeded, jobid=5953221 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f024 succeeded, jobid=5953222 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f012-f014 succeeded, jobid=5953223 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f015-f017 succeeded, jobid=5953224 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f018-f020 succeeded, jobid=5953225 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f021-f023 succeeded, jobid=5953226 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f024-f026 succeeded, jobid=5953227 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f018-f020 succeeded, jobid=5953228 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f021-f023 succeeded, jobid=5953229 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f024-f026 succeeded, jobid=5953230 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f027-f029 succeeded, jobid=5953231 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f030-f032 succeeded, jobid=5953232 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmanlupp succeeded, jobid=5953233 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f000 succeeded, jobid=5953234 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f001 succeeded, jobid=5953235 -2025-07-29 02:37:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f002 succeeded, jobid=5953236 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5953115, in state RUNNING (RUNNING) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f006, jobid=5953219, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f012, jobid=5953220, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f018, jobid=5953221, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f024, jobid=5953222, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5953198, in state SUCCEEDED (COMPLETED), ran for 271.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5953199, in state SUCCEEDED (COMPLETED), ran for 265.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5953200, in state SUCCEEDED (COMPLETED), ran for 272.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5953201, in state SUCCEEDED (COMPLETED), ran for 274.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5953223, in state RUNNING (RUNNING) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5953224, in state RUNNING (RUNNING) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5953225, in state RUNNING (RUNNING) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5953226, in state RUNNING (RUNNING) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5953227, in state RUNNING (RUNNING) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5953228, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5953229, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5953230, in state SUCCEEDED (COMPLETED), ran for 71.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5953231, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5953232, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5953233, in state SUCCEEDED (COMPLETED), ran for 27.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5953163, in state SUCCEEDED (COMPLETED), ran for 475.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5953234, in state SUCCEEDED (COMPLETED), ran for 65.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f001, jobid=5953235, in state SUCCEEDED (COMPLETED), ran for 66.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f002, jobid=5953236, in state SUCCEEDED (COMPLETED), ran for 66.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task enkfgdas_esfc, jobid=5953210, in state SUCCEEDED (COMPLETED), ran for 143.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f018 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f024 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f030 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f036 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f027-f029 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f030-f032 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f033-f035 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f036-f038 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f033-f035 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f036-f038 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f039-f041 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f042-f044 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmanlprod -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f003 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f004 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f005 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f006 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f007 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f008 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f009, because maximum global task throttle of 25 will be violated. -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f000, because maximum global task throttle of 25 will be violated. -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f001, because maximum global task throttle of 25 will be violated. -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f002, because maximum global task throttle of 25 will be violated. -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f003, because maximum global task throttle of 25 will be violated. -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f004, because maximum global task throttle of 25 will be violated. -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f005, because maximum global task throttle of 25 will be violated. -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f006, because maximum global task throttle of 25 will be violated. -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f007, because maximum global task throttle of 25 will be violated. -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f008, because maximum global task throttle of 25 will be violated. -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostsbs_f009, because maximum global task throttle of 25 will be violated. -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit enkfgdas_fcst_mem001, because maximum global task throttle of 25 will be violated. -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit enkfgdas_fcst_mem002, because maximum global task throttle of 25 will be violated. -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f018 succeeded, jobid=5953257 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f024 succeeded, jobid=5953258 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f030 succeeded, jobid=5953259 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f036 succeeded, jobid=5953260 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f027-f029 succeeded, jobid=5953261 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f030-f032 succeeded, jobid=5953262 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f033-f035 succeeded, jobid=5953263 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f036-f038 succeeded, jobid=5953264 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f033-f035 succeeded, jobid=5953265 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f036-f038 succeeded, jobid=5953266 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f039-f041 succeeded, jobid=5953267 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f042-f044 succeeded, jobid=5953268 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmanlprod succeeded, jobid=5953269 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f003 succeeded, jobid=5953270 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f004 succeeded, jobid=5953271 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f005 succeeded, jobid=5953272 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f006 succeeded, jobid=5953273 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f007 succeeded, jobid=5953274 -2025-07-29 02:43:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f008 succeeded, jobid=5953275 diff --git a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96_atm3DVar_logs_2021122018.log b/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96_atm3DVar_logs_2021122018.log deleted file mode 100644 index 0684a05d..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96_atm3DVar_logs_2021122018.log +++ /dev/null @@ -1,41 +0,0 @@ -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_stage_ic -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_stage_ic is pending at druby://130.18.14.151:37759 -2025-07-28 14:48:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_stage_ic is success, jobid=5950263 -2025-07-28 14:48:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_stage_ic, jobid=5950263, in state RUNNING (RUNNING) -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_stage_ic, jobid=5950263, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_fcst_seg0 -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_fcst_seg0 is pending at druby://130.18.14.151:38819 -2025-07-28 15:00:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_fcst_seg0 is success, jobid=5950285 -2025-07-28 15:00:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5950285, in state QUEUED (PENDING) -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5950285, in state RUNNING (RUNNING) -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5950285, in state SUCCEEDED (COMPLETED), ran for 344.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f000 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f003 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f006 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f009 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f000 succeeded, jobid=5950359 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f003 succeeded, jobid=5950360 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f006 succeeded, jobid=5950361 -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f009 succeeded, jobid=5950362 -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950359, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950360, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950361, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950362, in state QUEUED (PENDING) -2025-07-28 15:24:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950359, in state QUEUED (PENDING) -2025-07-28 15:24:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950360, in state QUEUED (PENDING) -2025-07-28 15:24:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950361, in state QUEUED (PENDING) -2025-07-28 15:24:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950362, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950359, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950360, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950361, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950362, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950359, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950360, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950361, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950362, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5950359, in state SUCCEEDED (COMPLETED), ran for 63.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5950360, in state SUCCEEDED (COMPLETED), ran for 66.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5950361, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950362, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5950362, in state SUCCEEDED (COMPLETED), ran for 66.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: This cycle is complete: Success diff --git a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96_atm3DVar_logs_2021122100.log b/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96_atm3DVar_logs_2021122100.log deleted file mode 100644 index cfb16254..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96_atm3DVar_logs_2021122100.log +++ /dev/null @@ -1,714 +0,0 @@ -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_prep -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_prep -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_prep succeeded, jobid=5950559 -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_prep is pending at druby://130.18.14.151:33161 -2025-07-28 15:55:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_prep is success, jobid=5950561 -2025-07-28 15:55:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 15:55:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 16:13:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 16:25:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 16:31:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 16:31:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 16:37:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 16:37:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 16:43:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 16:43:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 16:49:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 16:49:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 16:55:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 16:55:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 17:01:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 17:01:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 17:07:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 17:07:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 17:13:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 17:13:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 17:19:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 17:19:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 17:25:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 17:25:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 17:31:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 17:31:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 17:37:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 17:37:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 17:43:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 17:43:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 17:49:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 17:49:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 17:55:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 17:55:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 18:01:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 18:01:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 18:07:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 18:07:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 18:14:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 18:14:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 18:20:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 18:20:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 18:26:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 18:26:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 18:32:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 18:32:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 18:38:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 18:38:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 18:44:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 18:44:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 18:50:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 18:50:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 18:56:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 18:56:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 19:02:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 19:02:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 19:08:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 19:08:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 19:14:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state QUEUED (PENDING) -2025-07-28 19:14:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 19:20:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state RUNNING (RUNNING) -2025-07-28 19:20:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 19:26:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_prep, jobid=5950559, in state SUCCEEDED (COMPLETED), ran for 130.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:26:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 19:26:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_anal -2025-07-28 19:26:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_anal is pending at druby://130.18.14.151:45575 -2025-07-28 19:32:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_anal is success, jobid=5951167 -2025-07-28 19:32:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951167, in state QUEUED (PENDING) -2025-07-28 19:32:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 19:38:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951167, in state QUEUED (PENDING) -2025-07-28 19:38:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 19:44:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951167, in state QUEUED (PENDING) -2025-07-28 19:44:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state QUEUED (PENDING) -2025-07-28 19:50:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951167, in state QUEUED (PENDING) -2025-07-28 19:50:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state RUNNING (RUNNING) -2025-07-28 19:56:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951167, in state QUEUED (PENDING) -2025-07-28 19:56:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5950561, in state SUCCEEDED (COMPLETED), ran for 136.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:56:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_anal -2025-07-28 19:56:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_anal is pending at druby://130.18.14.151:37721 -2025-07-28 20:02:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_anal is success, jobid=5951227 -2025-07-28 20:02:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951167, in state QUEUED (PENDING) -2025-07-28 20:02:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951227, in state QUEUED (PENDING) -2025-07-28 20:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951167, in state QUEUED (PENDING) -2025-07-28 20:08:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951227, in state QUEUED (PENDING) -2025-07-28 20:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951167, in state QUEUED (PENDING) -2025-07-28 20:14:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951227, in state QUEUED (PENDING) -2025-07-28 20:20:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951167, in state QUEUED (PENDING) -2025-07-28 20:20:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951227, in state QUEUED (PENDING) -2025-07-28 20:26:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951167, in state QUEUED (PENDING) -2025-07-28 20:26:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951227, in state QUEUED (PENDING) -2025-07-28 20:32:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951167, in state QUEUED (PENDING) -2025-07-28 20:32:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951227, in state QUEUED (PENDING) -2025-07-28 20:39:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951167, in state QUEUED (PENDING) -2025-07-28 20:39:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951227, in state QUEUED (PENDING) -2025-07-28 20:45:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951167, in state QUEUED (PENDING) -2025-07-28 20:45:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951227, in state QUEUED (PENDING) -2025-07-28 20:51:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951167, in state QUEUED (PENDING) -2025-07-28 20:51:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951227, in state QUEUED (PENDING) -2025-07-28 20:57:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951167, in state QUEUED (PENDING) -2025-07-28 20:57:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951227, in state QUEUED (PENDING) -2025-07-28 21:03:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951167, in state QUEUED (PENDING) -2025-07-28 21:03:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951227, in state QUEUED (PENDING) -2025-07-28 21:09:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951167, in state QUEUED (PENDING) -2025-07-28 21:09:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951227, in state QUEUED (PENDING) -2025-07-28 21:15:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951167, in state QUEUED (PENDING) -2025-07-28 21:15:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951227, in state QUEUED (PENDING) -2025-07-28 21:21:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951167, in state RUNNING (RUNNING) -2025-07-28 21:21:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951227, in state RUNNING (RUNNING) -2025-07-28 21:27:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951167, in state RUNNING (RUNNING) -2025-07-28 21:27:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951227, in state RUNNING (RUNNING) -2025-07-28 21:33:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_anal, jobid=5951167, in state SUCCEEDED (COMPLETED), ran for 708.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951227, in state RUNNING (RUNNING) -2025-07-28 21:33:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_sfcanl -2025-07-28 21:33:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_vminmon -2025-07-28 21:33:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_sfcanl succeeded, jobid=5951710 -2025-07-28 21:33:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_vminmon is pending at druby://130.18.14.151:41059 -2025-07-28 21:39:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_vminmon is success, jobid=5951711 -2025-07-28 21:39:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_sfcanl, jobid=5951710, in state QUEUED (PENDING) -2025-07-28 21:39:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_vminmon, jobid=5951711, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:39:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5951227, in state SUCCEEDED (COMPLETED), ran for 1031.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:39:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_sfcanl -2025-07-28 21:39:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_analdiag -2025-07-28 21:39:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_vminmon -2025-07-28 21:39:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_sfcanl succeeded, jobid=5951773 -2025-07-28 21:39:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_analdiag succeeded, jobid=5951774 -2025-07-28 21:39:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_vminmon succeeded, jobid=5951775 -2025-07-28 21:45:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_sfcanl, jobid=5951710, in state QUEUED (PENDING) -2025-07-28 21:45:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5951773, in state QUEUED (PENDING) -2025-07-28 21:45:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951774, in state QUEUED (PENDING) -2025-07-28 21:45:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_vminmon, jobid=5951775, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:51:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_sfcanl, jobid=5951710, in state SUCCEEDED (COMPLETED), ran for 48.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:51:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5951773, in state QUEUED (PENDING) -2025-07-28 21:51:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951774, in state QUEUED (PENDING) -2025-07-28 21:51:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_analcalc -2025-07-28 21:51:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_fcst_seg0 -2025-07-28 21:51:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_analcalc is pending at druby://130.18.14.151:40679 -2025-07-28 21:51:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_fcst_seg0 is pending at druby://130.18.14.151:40679 -2025-07-28 21:57:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_analcalc is success, jobid=5951866 -2025-07-28 21:57:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_fcst_seg0 is success, jobid=5951867 -2025-07-28 21:57:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951866, in state QUEUED (PENDING) -2025-07-28 21:57:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951867, in state QUEUED (PENDING) -2025-07-28 21:57:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5951773, in state QUEUED (PENDING) -2025-07-28 21:57:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951774, in state QUEUED (PENDING) -2025-07-28 22:03:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951866, in state QUEUED (PENDING) -2025-07-28 22:03:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951867, in state QUEUED (PENDING) -2025-07-28 22:03:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5951773, in state QUEUED (PENDING) -2025-07-28 22:03:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951774, in state QUEUED (PENDING) -2025-07-28 22:09:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951866, in state RUNNING (RUNNING) -2025-07-28 22:09:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951867, in state QUEUED (PENDING) -2025-07-28 22:09:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5951773, in state SUCCEEDED (COMPLETED), ran for 44.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:09:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5951774, in state SUCCEEDED (COMPLETED), ran for 124.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:09:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_analcalc -2025-07-28 22:09:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_fcst_seg0 -2025-07-28 22:09:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_verfozn -2025-07-28 22:09:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_verfrad -2025-07-28 22:09:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_analcalc succeeded, jobid=5951950 -2025-07-28 22:09:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_fcst_seg0 succeeded, jobid=5951951 -2025-07-28 22:09:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_verfozn succeeded, jobid=5951952 -2025-07-28 22:09:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_verfrad succeeded, jobid=5951953 -2025-07-28 22:15:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_analcalc, jobid=5951866, in state SUCCEEDED (COMPLETED), ran for 47.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:15:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951867, in state QUEUED (PENDING) -2025-07-28 22:15:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5951950, in state QUEUED (PENDING) -2025-07-28 22:15:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951951, in state QUEUED (PENDING) -2025-07-28 22:15:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfozn, jobid=5951952, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:15:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5951953, in state RUNNING (RUNNING) -2025-07-28 22:15:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmanlupp -2025-07-28 22:15:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_atmanlupp is pending at druby://130.18.14.151:44767 -2025-07-28 22:21:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_atmanlupp is success, jobid=5951981 -2025-07-28 22:21:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5951981, in state QUEUED (PENDING) -2025-07-28 22:21:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951867, in state QUEUED (PENDING) -2025-07-28 22:21:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5951950, in state QUEUED (PENDING) -2025-07-28 22:21:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951951, in state QUEUED (PENDING) -2025-07-28 22:21:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5951953, in state RUNNING (RUNNING) -2025-07-28 22:27:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5951981, in state QUEUED (PENDING) -2025-07-28 22:27:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951867, in state QUEUED (PENDING) -2025-07-28 22:27:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5951950, in state QUEUED (PENDING) -2025-07-28 22:27:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951951, in state QUEUED (PENDING) -2025-07-28 22:27:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5951953, in state SUCCEEDED (COMPLETED), ran for 599.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:33:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5951981, in state QUEUED (PENDING) -2025-07-28 22:33:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951867, in state RUNNING (RUNNING) -2025-07-28 22:33:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5951950, in state QUEUED (PENDING) -2025-07-28 22:33:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951951, in state QUEUED (PENDING) -2025-07-28 22:39:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5951981, in state QUEUED (PENDING) -2025-07-28 22:39:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951867, in state RUNNING (RUNNING) -2025-07-28 22:39:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5951950, in state QUEUED (PENDING) -2025-07-28 22:39:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951951, in state QUEUED (PENDING) -2025-07-28 22:39:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f000-f002 -2025-07-28 22:39:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f003-f005 -2025-07-28 22:39:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f006-f008 -2025-07-28 22:39:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f009-f011 -2025-07-28 22:39:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f000-f002 succeeded, jobid=5952022 -2025-07-28 22:39:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f003-f005 succeeded, jobid=5952023 -2025-07-28 22:39:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f006-f008 succeeded, jobid=5952024 -2025-07-28 22:39:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f009-f011 succeeded, jobid=5952025 -2025-07-28 22:45:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5951981, in state QUEUED (PENDING) -2025-07-28 22:45:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951867, in state RUNNING (RUNNING) -2025-07-28 22:45:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952022, in state QUEUED (PENDING) -2025-07-28 22:45:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952023, in state QUEUED (PENDING) -2025-07-28 22:45:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952024, in state QUEUED (PENDING) -2025-07-28 22:45:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952025, in state QUEUED (PENDING) -2025-07-28 22:45:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5951950, in state QUEUED (PENDING) -2025-07-28 22:45:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951951, in state QUEUED (PENDING) -2025-07-28 22:45:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f012-f014 -2025-07-28 22:45:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f015-f017 -2025-07-28 22:45:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f018-f020 -2025-07-28 22:45:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f021-f023 -2025-07-28 22:45:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f024-f026 -2025-07-28 22:45:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f012-f014 succeeded, jobid=5952032 -2025-07-28 22:45:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f015-f017 succeeded, jobid=5952033 -2025-07-28 22:45:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f018-f020 succeeded, jobid=5952034 -2025-07-28 22:45:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f021-f023 succeeded, jobid=5952036 -2025-07-28 22:45:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f024-f026 succeeded, jobid=5952037 -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5951981, in state QUEUED (PENDING) -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951867, in state RUNNING (RUNNING) -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952022, in state QUEUED (PENDING) -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952023, in state QUEUED (PENDING) -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952024, in state QUEUED (PENDING) -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952025, in state QUEUED (PENDING) -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952032, in state QUEUED (PENDING) -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952033, in state QUEUED (PENDING) -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952034, in state QUEUED (PENDING) -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952036, in state QUEUED (PENDING) -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952037, in state QUEUED (PENDING) -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5951950, in state QUEUED (PENDING) -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951951, in state QUEUED (PENDING) -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f027-f029 -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f030-f032 -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f033-f035 -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f036-f038 -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f039-f041 -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f027-f029 succeeded, jobid=5952047 -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f030-f032 succeeded, jobid=5952048 -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f033-f035 succeeded, jobid=5952049 -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f036-f038 succeeded, jobid=5952050 -2025-07-28 22:51:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f039-f041 succeeded, jobid=5952051 -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5951981, in state QUEUED (PENDING) -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951867, in state RUNNING (RUNNING) -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952022, in state QUEUED (PENDING) -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952023, in state QUEUED (PENDING) -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952024, in state QUEUED (PENDING) -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952025, in state QUEUED (PENDING) -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952032, in state QUEUED (PENDING) -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952033, in state QUEUED (PENDING) -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952034, in state QUEUED (PENDING) -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952036, in state QUEUED (PENDING) -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952037, in state QUEUED (PENDING) -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952047, in state QUEUED (PENDING) -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952048, in state QUEUED (PENDING) -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952049, in state QUEUED (PENDING) -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952050, in state QUEUED (PENDING) -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952051, in state QUEUED (PENDING) -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5951950, in state QUEUED (PENDING) -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951951, in state QUEUED (PENDING) -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f042-f044 -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f045-f047 -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f048-f054 -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f042-f044 succeeded, jobid=5952061 -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f045-f047 succeeded, jobid=5952062 -2025-07-28 22:58:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f048-f054 succeeded, jobid=5952063 -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5951981, in state QUEUED (PENDING) -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951867, in state RUNNING (RUNNING) -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952022, in state QUEUED (PENDING) -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952023, in state QUEUED (PENDING) -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952024, in state QUEUED (PENDING) -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952025, in state QUEUED (PENDING) -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952032, in state QUEUED (PENDING) -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952033, in state QUEUED (PENDING) -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952034, in state QUEUED (PENDING) -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952036, in state QUEUED (PENDING) -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952037, in state QUEUED (PENDING) -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952047, in state QUEUED (PENDING) -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952048, in state QUEUED (PENDING) -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952049, in state QUEUED (PENDING) -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952050, in state QUEUED (PENDING) -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952051, in state QUEUED (PENDING) -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952061, in state QUEUED (PENDING) -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952062, in state QUEUED (PENDING) -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952063, in state QUEUED (PENDING) -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5951950, in state QUEUED (PENDING) -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951951, in state QUEUED (PENDING) -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f057-f063 -2025-07-28 23:04:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_atmos_prod_f057-f063 is pending at druby://130.18.14.151:42559 -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_atmos_prod_f057-f063 is success, jobid=5952088 -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5951981, in state QUEUED (PENDING) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951867, in state RUNNING (RUNNING) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952022, in state QUEUED (PENDING) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952023, in state QUEUED (PENDING) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952024, in state QUEUED (PENDING) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952025, in state QUEUED (PENDING) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952032, in state QUEUED (PENDING) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952033, in state QUEUED (PENDING) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952034, in state QUEUED (PENDING) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952036, in state QUEUED (PENDING) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952037, in state QUEUED (PENDING) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952047, in state QUEUED (PENDING) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952048, in state QUEUED (PENDING) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952049, in state QUEUED (PENDING) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952050, in state QUEUED (PENDING) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952051, in state QUEUED (PENDING) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952061, in state QUEUED (PENDING) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952062, in state QUEUED (PENDING) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952063, in state QUEUED (PENDING) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952088, in state QUEUED (PENDING) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5951950, in state SUCCEEDED (COMPLETED), ran for 45.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951951, in state RUNNING (RUNNING) -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f066-f072 -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f075-f081 -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmanlupp -2025-07-28 23:10:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f066-f072 succeeded, jobid=5952113 -2025-07-28 23:10:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f075-f081 succeeded, jobid=5952114 -2025-07-28 23:10:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmanlupp succeeded, jobid=5952115 -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5951981, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951867, in state RUNNING (RUNNING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952022, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952023, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952024, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952025, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952032, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952033, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952034, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952036, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952037, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952047, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952048, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952049, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952050, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952051, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952061, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952062, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952063, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952088, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952113, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952114, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952115, in state QUEUED (PENDING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951951, in state RUNNING (RUNNING) -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f084-f090 -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f000, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f003, because maximum global task throttle of 25 will be violated. -2025-07-28 23:16:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_atmos_prod_f084-f090 is pending at druby://130.18.14.151:36201 -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_atmos_prod_f084-f090 is success, jobid=5952134 -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5951981, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951867, in state RUNNING (RUNNING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952022, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952023, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952024, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952025, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952032, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952033, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952034, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952036, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952037, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952047, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952048, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952049, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952050, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952051, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952061, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952062, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952063, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952088, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952113, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952114, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952134, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952115, in state QUEUED (PENDING) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5951951, in state SUCCEEDED (COMPLETED), ran for 438.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f093-f099 -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f000, because maximum global task throttle of 25 will be violated. -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f003, because maximum global task throttle of 25 will be violated. -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f006, because maximum global task throttle of 25 will be violated. -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f009, because maximum global task throttle of 25 will be violated. -2025-07-28 23:22:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f093-f099 succeeded, jobid=5952145 -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5951981, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5951867, in state SUCCEEDED (COMPLETED), ran for 3214.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952022, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952023, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952024, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952025, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952032, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952033, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952034, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952036, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952037, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952047, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952048, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952049, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952050, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952051, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952061, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952062, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952063, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952088, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952113, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952114, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952134, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952145, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952115, in state QUEUED (PENDING) -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f102-f108 -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f000, because maximum global task throttle of 25 will be violated. -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f003, because maximum global task throttle of 25 will be violated. -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f006, because maximum global task throttle of 25 will be violated. -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f009, because maximum global task throttle of 25 will be violated. -2025-07-28 23:28:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f102-f108 succeeded, jobid=5952157 -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlupp, jobid=5951981, in state SUCCEEDED (COMPLETED), ran for 44.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952022, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952023, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952024, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952025, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952032, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952033, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952034, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952036, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952037, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952047, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952048, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952049, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952050, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952051, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952061, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952062, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952063, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952088, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952113, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952114, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952134, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952145, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952157, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952115, in state QUEUED (PENDING) -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmanlprod -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f000, because maximum global task throttle of 25 will be violated. -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f003, because maximum global task throttle of 25 will be violated. -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f006, because maximum global task throttle of 25 will be violated. -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f009, because maximum global task throttle of 25 will be violated. -2025-07-28 23:34:16 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_atmanlprod is pending at druby://130.18.14.151:38913 -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_atmanlprod is success, jobid=5952169 -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlprod, jobid=5952169, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952022, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952023, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952024, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952025, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952032, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952033, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952034, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952036, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952037, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952047, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952048, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952049, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952050, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952051, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952061, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952062, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952063, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952088, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952113, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952114, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952134, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952145, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952157, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952115, in state QUEUED (PENDING) -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f000, because maximum global task throttle of 25 will be violated. -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f003, because maximum global task throttle of 25 will be violated. -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f006, because maximum global task throttle of 25 will be violated. -2025-07-28 23:40:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f009, because maximum global task throttle of 25 will be violated. -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlprod, jobid=5952169, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952022, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952023, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952024, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952025, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952032, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952033, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952034, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952036, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952037, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952047, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952048, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952049, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952050, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952051, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952061, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952062, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952063, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952088, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952113, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952114, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952134, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952145, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952157, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952115, in state QUEUED (PENDING) -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f000, because maximum global task throttle of 25 will be violated. -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f003, because maximum global task throttle of 25 will be violated. -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f006, because maximum global task throttle of 25 will be violated. -2025-07-28 23:46:21 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f009, because maximum global task throttle of 25 will be violated. -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlprod, jobid=5952169, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952022, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952023, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952024, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952025, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952032, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952033, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952034, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952036, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952037, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952047, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952048, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952049, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952050, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952051, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952061, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952062, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952063, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952088, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952113, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952114, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952134, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952145, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952157, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952115, in state QUEUED (PENDING) -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f000, because maximum global task throttle of 25 will be violated. -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f003, because maximum global task throttle of 25 will be violated. -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f006, because maximum global task throttle of 25 will be violated. -2025-07-28 23:52:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f009, because maximum global task throttle of 25 will be violated. -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlprod, jobid=5952169, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952022, in state RUNNING (RUNNING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952023, in state RUNNING (RUNNING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952024, in state RUNNING (RUNNING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952025, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952032, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952033, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952034, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952036, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952037, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952047, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952048, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952049, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952050, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952051, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952061, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952062, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952063, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952088, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952113, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952114, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952134, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952145, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952157, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952115, in state QUEUED (PENDING) -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f000, because maximum global task throttle of 25 will be violated. -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f003, because maximum global task throttle of 25 will be violated. -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f006, because maximum global task throttle of 25 will be violated. -2025-07-28 23:58:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gdas_atmos_prod_f009, because maximum global task throttle of 25 will be violated. -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlprod, jobid=5952169, in state QUEUED (PENDING) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5952022, in state SUCCEEDED (COMPLETED), ran for 267.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5952023, in state SUCCEEDED (COMPLETED), ran for 279.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5952024, in state SUCCEEDED (COMPLETED), ran for 279.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5952025, in state SUCCEEDED (COMPLETED), ran for 278.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5952032, in state SUCCEEDED (COMPLETED), ran for 281.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5952033, in state SUCCEEDED (COMPLETED), ran for 278.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5952034, in state SUCCEEDED (COMPLETED), ran for 272.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5952036, in state SUCCEEDED (COMPLETED), ran for 273.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5952037, in state SUCCEEDED (COMPLETED), ran for 275.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5952047, in state SUCCEEDED (COMPLETED), ran for 273.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5952048, in state SUCCEEDED (COMPLETED), ran for 272.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5952049, in state SUCCEEDED (COMPLETED), ran for 269.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5952050, in state SUCCEEDED (COMPLETED), ran for 271.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5952051, in state SUCCEEDED (COMPLETED), ran for 273.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5952061, in state SUCCEEDED (COMPLETED), ran for 277.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5952062, in state SUCCEEDED (COMPLETED), ran for 274.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952063, in state RUNNING (RUNNING) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952088, in state QUEUED (PENDING) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952113, in state QUEUED (PENDING) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952114, in state QUEUED (PENDING) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952134, in state QUEUED (PENDING) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952145, in state QUEUED (PENDING) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952157, in state QUEUED (PENDING) -2025-07-29 00:04:29 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952115, in state QUEUED (PENDING) -2025-07-29 00:04:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f111-f114 -2025-07-29 00:04:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f117-f120 -2025-07-29 00:04:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f000 -2025-07-29 00:04:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f003 -2025-07-29 00:04:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f006 -2025-07-29 00:04:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f009 -2025-07-29 00:04:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f111-f114 succeeded, jobid=5952318 -2025-07-29 00:04:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f117-f120 succeeded, jobid=5952319 -2025-07-29 00:04:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f000 succeeded, jobid=5952320 -2025-07-29 00:04:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f003 succeeded, jobid=5952321 -2025-07-29 00:04:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f006 succeeded, jobid=5952322 -2025-07-29 00:04:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f009 succeeded, jobid=5952323 -2025-07-29 00:10:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlprod, jobid=5952169, in state QUEUED (PENDING) -2025-07-29 00:10:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5952063, in state SUCCEEDED (COMPLETED), ran for 319.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:10:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5952088, in state SUCCEEDED (COMPLETED), ran for 316.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:10:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952113, in state QUEUED (PENDING) -2025-07-29 00:10:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952114, in state QUEUED (PENDING) -2025-07-29 00:10:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952134, in state QUEUED (PENDING) -2025-07-29 00:10:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952145, in state QUEUED (PENDING) -2025-07-29 00:10:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952157, in state QUEUED (PENDING) -2025-07-29 00:10:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5952318, in state QUEUED (PENDING) -2025-07-29 00:10:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5952319, in state QUEUED (PENDING) -2025-07-29 00:10:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952115, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:10:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952320, in state QUEUED (PENDING) -2025-07-29 00:10:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952321, in state QUEUED (PENDING) -2025-07-29 00:10:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952322, in state QUEUED (PENDING) -2025-07-29 00:10:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952323, in state QUEUED (PENDING) -2025-07-29 00:10:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmanlprod -2025-07-29 00:10:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_atmanlprod is pending at druby://130.18.14.151:45731 -2025-07-29 00:16:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_atmanlprod is success, jobid=5952377 -2025-07-29 00:16:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlprod, jobid=5952169, in state QUEUED (PENDING) -2025-07-29 00:16:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952113, in state RUNNING (RUNNING) -2025-07-29 00:16:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952114, in state RUNNING (RUNNING) -2025-07-29 00:16:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952134, in state QUEUED (PENDING) -2025-07-29 00:16:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952145, in state QUEUED (PENDING) -2025-07-29 00:16:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952157, in state QUEUED (PENDING) -2025-07-29 00:16:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5952318, in state QUEUED (PENDING) -2025-07-29 00:16:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5952319, in state QUEUED (PENDING) -2025-07-29 00:16:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlprod, jobid=5952377, in state QUEUED (PENDING) -2025-07-29 00:16:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952320, in state QUEUED (PENDING) -2025-07-29 00:16:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952321, in state QUEUED (PENDING) -2025-07-29 00:16:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952322, in state QUEUED (PENDING) -2025-07-29 00:16:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952323, in state QUEUED (PENDING) -2025-07-29 00:22:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmanlprod, jobid=5952169, in state SUCCEEDED (COMPLETED), ran for 107.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:22:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5952113, in state SUCCEEDED (COMPLETED), ran for 316.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:22:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5952114, in state SUCCEEDED (COMPLETED), ran for 317.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:22:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952134, in state RUNNING (RUNNING) -2025-07-29 00:22:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952145, in state RUNNING (RUNNING) -2025-07-29 00:22:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952157, in state RUNNING (RUNNING) -2025-07-29 00:22:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5952318, in state QUEUED (PENDING) -2025-07-29 00:22:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5952319, in state QUEUED (PENDING) -2025-07-29 00:22:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlprod, jobid=5952377, in state QUEUED (PENDING) -2025-07-29 00:22:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952320, in state QUEUED (PENDING) -2025-07-29 00:22:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952321, in state QUEUED (PENDING) -2025-07-29 00:22:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952322, in state QUEUED (PENDING) -2025-07-29 00:22:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952323, in state QUEUED (PENDING) -2025-07-29 00:28:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5952134, in state SUCCEEDED (COMPLETED), ran for 326.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:28:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5952145, in state SUCCEEDED (COMPLETED), ran for 324.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:28:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5952157, in state SUCCEEDED (COMPLETED), ran for 331.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:28:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5952318, in state QUEUED (PENDING) -2025-07-29 00:28:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5952319, in state QUEUED (PENDING) -2025-07-29 00:28:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlprod, jobid=5952377, in state QUEUED (PENDING) -2025-07-29 00:28:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952320, in state QUEUED (PENDING) -2025-07-29 00:28:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952321, in state QUEUED (PENDING) -2025-07-29 00:28:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952322, in state QUEUED (PENDING) -2025-07-29 00:28:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952323, in state QUEUED (PENDING) -2025-07-29 00:34:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5952318, in state SUCCEEDED (COMPLETED), ran for 218.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:34:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5952319, in state SUCCEEDED (COMPLETED), ran for 216.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:34:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlprod, jobid=5952377, in state QUEUED (PENDING) -2025-07-29 00:34:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952320, in state SUCCEEDED (COMPLETED), ran for 65.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:34:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952321, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:34:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952322, in state SUCCEEDED (COMPLETED), ran for 67.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:34:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952323, in state SUCCEEDED (COMPLETED), ran for 68.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:34:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_tracker -2025-07-29 00:34:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_genesis -2025-07-29 00:34:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_fit2obs -2025-07-29 00:34:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_tracker succeeded, jobid=5952534 -2025-07-29 00:34:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_genesis succeeded, jobid=5952535 -2025-07-29 00:34:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_fit2obs succeeded, jobid=5952536 -2025-07-29 00:40:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_tracker, jobid=5952534, in state SUCCEEDED (COMPLETED), ran for 65.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:40:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_genesis, jobid=5952535, in state RUNNING (RUNNING) -2025-07-29 00:40:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlprod, jobid=5952377, in state SUCCEEDED (COMPLETED), ran for 64.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:40:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fit2obs, jobid=5952536, in state SUCCEEDED (COMPLETED), ran for 22.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:40:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_arch_vrfy -2025-07-29 00:40:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_arch_vrfy is pending at druby://130.18.14.151:45597 -2025-07-29 00:46:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_arch_vrfy is success, jobid=5952569 -2025-07-29 00:46:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_genesis, jobid=5952535, in state SUCCEEDED (COMPLETED), ran for 504.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:46:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_arch_vrfy, jobid=5952569, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:46:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_arch_vrfy -2025-07-29 00:46:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_cleanup -2025-07-29 00:46:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_arch_vrfy succeeded, jobid=5952625 -2025-07-29 00:46:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_cleanup succeeded, jobid=5952626 -2025-07-29 00:52:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_arch_vrfy, jobid=5952625, in state SUCCEEDED (COMPLETED), ran for 75.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:52:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_cleanup, jobid=5952626, in state SUCCEEDED (COMPLETED), ran for 24.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:52:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_metpg2g1 -2025-07-29 00:52:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_metpg2g1 is pending at druby://130.18.14.151:46259 -2025-07-29 00:58:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_metpg2g1 is success, jobid=5952653 -2025-07-29 00:58:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_metpg2g1, jobid=5952653, in state SUCCEEDED (COMPLETED), ran for 115.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:58:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_metpg2o1 -2025-07-29 00:58:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_metpg2o1 is pending at druby://130.18.14.151:38815 -2025-07-29 01:04:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_metpg2o1 is success, jobid=5952677 -2025-07-29 01:04:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_metpg2o1, jobid=5952677, in state RUNNING (RUNNING) -2025-07-29 01:10:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_metpg2o1, jobid=5952677, in state SUCCEEDED (COMPLETED), ran for 404.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:10:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_metppcp1 -2025-07-29 01:10:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_metppcp1 is pending at druby://130.18.14.151:42859 -2025-07-29 01:17:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_metppcp1 is success, jobid=5952716 -2025-07-29 01:17:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_metppcp1, jobid=5952716, in state SUCCEEDED (COMPLETED), ran for 28.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:17:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_cleanup -2025-07-29 01:17:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_cleanup is pending at druby://130.18.14.151:44137 -2025-07-29 01:23:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_cleanup is success, jobid=5952724 -2025-07-29 01:23:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_cleanup, jobid=5952724, in state SUCCEEDED (COMPLETED), ran for 20.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:23:04 -0500 :: hercules-login-1.hpc.msstate.edu :: This cycle is complete: Success diff --git a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96_atm3DVar_logs_2021122106.log b/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96_atm3DVar_logs_2021122106.log deleted file mode 100644 index 35ed5190..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96_atm3DVar_logs_2021122106.log +++ /dev/null @@ -1,75 +0,0 @@ -2025-07-29 00:34:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_prep -2025-07-29 00:34:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_prep succeeded, jobid=5952537 -2025-07-29 00:40:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5952537, in state QUEUED (PENDING) -2025-07-29 00:46:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_prep, jobid=5952537, in state SUCCEEDED (COMPLETED), ran for 116.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:46:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_anal -2025-07-29 00:46:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_anal is pending at druby://130.18.14.151:43769 -2025-07-29 00:52:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_anal is success, jobid=5952627 -2025-07-29 00:52:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5952627, in state QUEUED (PENDING) -2025-07-29 00:58:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5952627, in state QUEUED (PENDING) -2025-07-29 01:04:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5952627, in state RUNNING (RUNNING) -2025-07-29 01:10:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5952627, in state RUNNING (RUNNING) -2025-07-29 01:17:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5952627, in state RUNNING (RUNNING) -2025-07-29 01:23:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_anal, jobid=5952627, in state SUCCEEDED (COMPLETED), ran for 1171.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:23:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_sfcanl -2025-07-29 01:23:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_analdiag -2025-07-29 01:23:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_vminmon -2025-07-29 01:23:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_sfcanl is pending at druby://130.18.14.151:41693 -2025-07-29 01:23:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_analdiag is pending at druby://130.18.14.151:41693 -2025-07-29 01:23:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_vminmon is pending at druby://130.18.14.151:41693 -2025-07-29 01:29:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_sfcanl is success, jobid=5952732 -2025-07-29 01:29:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_analdiag is success, jobid=5952733 -2025-07-29 01:29:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_vminmon is success, jobid=5952734 -2025-07-29 01:29:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_sfcanl, jobid=5952732, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:29:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analdiag, jobid=5952733, in state SUCCEEDED (COMPLETED), ran for 122.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:29:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_vminmon, jobid=5952734, in state SUCCEEDED (COMPLETED), ran for 21.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:29:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_analcalc -2025-07-29 01:29:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_fcst_seg0 -2025-07-29 01:29:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_verfozn -2025-07-29 01:29:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_verfrad -2025-07-29 01:29:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_analcalc succeeded, jobid=5952752 -2025-07-29 01:29:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_fcst_seg0 succeeded, jobid=5952753 -2025-07-29 01:29:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_verfozn succeeded, jobid=5952754 -2025-07-29 01:29:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_verfrad succeeded, jobid=5952755 -2025-07-29 01:35:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_analcalc, jobid=5952752, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:35:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952753, in state RUNNING (RUNNING) -2025-07-29 01:35:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfozn, jobid=5952754, in state SUCCEEDED (COMPLETED), ran for 61.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:35:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5952755, in state RUNNING (RUNNING) -2025-07-29 01:35:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmanlupp -2025-07-29 01:35:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_atmanlupp is pending at druby://130.18.14.151:43293 -2025-07-29 01:41:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_atmanlupp is success, jobid=5952795 -2025-07-29 01:41:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlupp, jobid=5952795, in state SUCCEEDED (COMPLETED), ran for 43.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:41:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952753, in state RUNNING (RUNNING) -2025-07-29 01:41:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_verfrad, jobid=5952755, in state SUCCEEDED (COMPLETED), ran for 678.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:41:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f000 -2025-07-29 01:41:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f003 -2025-07-29 01:41:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_atmos_prod_f000 is pending at druby://130.18.14.151:46007 -2025-07-29 01:41:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_atmos_prod_f003 is pending at druby://130.18.14.151:46007 -2025-07-29 01:47:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_atmos_prod_f000 is success, jobid=5952814 -2025-07-29 01:47:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_atmos_prod_f003 is success, jobid=5952815 -2025-07-29 01:47:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fcst_seg0, jobid=5952753, in state SUCCEEDED (COMPLETED), ran for 443.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:47:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f000, jobid=5952814, in state SUCCEEDED (COMPLETED), ran for 66.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:47:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f003, jobid=5952815, in state SUCCEEDED (COMPLETED), ran for 65.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:47:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmanlprod -2025-07-29 01:47:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f006 -2025-07-29 01:47:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_atmos_prod_f009 -2025-07-29 01:47:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmanlprod succeeded, jobid=5952849 -2025-07-29 01:47:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gdas_atmos_prod_f006 succeeded, jobid=5952850 -2025-07-29 01:47:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_atmos_prod_f009 is pending at druby://130.18.14.151:45699 -2025-07-29 01:53:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_atmos_prod_f009 is success, jobid=5952851 -2025-07-29 01:53:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmanlprod, jobid=5952849, in state SUCCEEDED (COMPLETED), ran for 64.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:53:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f006, jobid=5952850, in state SUCCEEDED (COMPLETED), ran for 66.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:53:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_atmos_prod_f009, jobid=5952851, in state SUCCEEDED (COMPLETED), ran for 66.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:53:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_fit2obs -2025-07-29 01:53:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_fit2obs is pending at druby://130.18.14.151:44613 -2025-07-29 01:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_fit2obs is success, jobid=5952889 -2025-07-29 01:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_fit2obs, jobid=5952889, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 01:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_arch_vrfy -2025-07-29 01:59:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_arch_vrfy is pending at druby://130.18.14.151:41715 -2025-07-29 02:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_arch_vrfy is success, jobid=5952922 -2025-07-29 02:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_arch_vrfy, jobid=5952922, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gdas_cleanup -2025-07-29 02:05:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gdas_cleanup is pending at druby://130.18.14.151:37107 -2025-07-29 02:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gdas_cleanup is success, jobid=5952997 -2025-07-29 02:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gdas_cleanup, jobid=5952997, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 02:11:25 -0500 :: hercules-login-1.hpc.msstate.edu :: This cycle is complete: Success diff --git a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96mx100_S2S_logs_1994050100.log b/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96mx100_S2S_logs_1994050100.log deleted file mode 100644 index dad119b2..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/EXPDIR_C96mx100_S2S_logs_1994050100.log +++ /dev/null @@ -1,791 +0,0 @@ -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_stage_ic -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of sfs_stage_ic is pending at druby://130.18.14.151:37127 -2025-07-28 14:48:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending sfs_stage_ic is success, jobid=5950265 -2025-07-28 14:48:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_stage_ic, jobid=5950265, in state RUNNING (RUNNING) -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_stage_ic, jobid=5950265, in state SUCCEEDED (COMPLETED), ran for 39.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_fcst_mem000_seg0 -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_fcst_mem001_seg0 -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_fcst_mem002_seg0 -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_fcst_mem000_seg0 succeeded, jobid=5950286 -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_fcst_mem001_seg0 succeeded, jobid=5950287 -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_fcst_mem002_seg0 succeeded, jobid=5950288 -2025-07-28 15:00:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg0, jobid=5950286, in state QUEUED (PENDING) -2025-07-28 15:00:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 15:00:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg0, jobid=5950286, in state QUEUED (PENDING) -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 15:06:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg0, jobid=5950286, in state QUEUED (PENDING) -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 15:12:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg0, jobid=5950286, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 15:18:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg0, jobid=5950286, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg0, jobid=5950286, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 15:30:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg0, jobid=5950286, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 15:36:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg0, jobid=5950286, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 15:42:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg0, jobid=5950286, in state RUNNING (RUNNING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_prod_mem000_f000 -2025-07-28 15:49:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of sfs_atmos_prod_mem000_f000 is pending at druby://130.18.14.151:46473 -2025-07-28 15:55:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending sfs_atmos_prod_mem000_f000 is success, jobid=5950560 -2025-07-28 15:55:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg0, jobid=5950286, in state RUNNING (RUNNING) -2025-07-28 15:55:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 15:55:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 15:55:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg0, jobid=5950286, in state RUNNING (RUNNING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_prod_mem000_f024 -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ice_prod_mem000_f024 -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_atmos_prod_mem000_f024 succeeded, jobid=5950575 -2025-07-28 16:01:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of sfs_ice_prod_mem000_f024 is pending at druby://130.18.14.151:42629 -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending sfs_ice_prod_mem000_f024 is success, jobid=5950576 -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg0, jobid=5950286, in state SUCCEEDED (COMPLETED), ran for 1339.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ice_prod_mem000_f024, jobid=5950576, in state SUCCEEDED (COMPLETED), ran for 44.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_fcst_mem000_seg1 -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_prod_mem000_f048 -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ocean_prod_mem000_f024 -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ocean_prod_mem000_f048 -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ice_prod_mem000_f048 -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_fcst_mem000_seg1 succeeded, jobid=5950610 -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_atmos_prod_mem000_f048 succeeded, jobid=5950611 -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ocean_prod_mem000_f024 succeeded, jobid=5950612 -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ocean_prod_mem000_f048 succeeded, jobid=5950613 -2025-07-28 16:07:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ice_prod_mem000_f048 succeeded, jobid=5950614 -2025-07-28 16:13:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 16:13:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 16:13:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 16:13:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 16:13:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 16:13:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 16:13:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ocean_prod_mem000_f024, jobid=5950612, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:13:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ocean_prod_mem000_f048, jobid=5950613, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:13:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ice_prod_mem000_f048, jobid=5950614, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 16:19:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 16:25:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 16:25:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 16:25:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 16:25:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 16:25:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 16:25:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 16:31:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 16:31:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 16:31:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 16:31:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 16:31:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 16:31:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 16:37:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 16:37:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 16:37:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 16:37:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 16:37:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 16:37:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 16:43:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 16:43:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 16:43:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 16:43:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 16:43:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 16:43:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 16:49:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 16:49:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 16:49:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 16:49:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 16:49:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 16:49:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 16:55:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 16:55:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 16:55:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 16:55:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 16:55:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 16:55:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 17:01:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 17:01:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 17:01:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 17:01:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 17:01:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 17:01:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 17:07:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 17:07:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 17:07:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 17:07:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 17:07:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 17:07:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 17:13:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 17:13:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 17:13:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 17:13:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 17:13:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 17:13:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 17:19:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 17:19:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 17:19:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 17:19:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 17:19:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 17:19:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 17:25:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 17:25:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 17:25:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 17:25:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 17:25:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 17:25:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 17:31:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 17:31:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 17:31:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 17:31:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 17:31:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 17:31:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 17:37:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 17:37:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 17:37:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 17:37:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 17:37:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 17:37:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 17:43:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 17:43:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 17:43:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 17:43:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 17:43:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 17:43:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 17:49:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 17:49:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 17:49:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 17:49:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 17:49:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 17:49:52 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 17:55:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 17:55:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 17:55:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 17:55:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 17:55:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 17:55:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 18:01:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 18:01:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 18:01:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 18:01:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 18:01:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 18:01:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 18:08:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 18:08:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 18:08:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 18:08:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 18:08:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 18:08:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 18:14:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 18:14:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 18:14:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 18:14:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 18:14:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 18:14:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 18:20:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 18:20:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 18:20:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 18:20:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 18:20:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 18:20:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 18:26:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 18:26:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 18:26:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 18:26:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 18:26:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 18:26:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 18:32:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 18:32:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 18:32:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 18:32:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 18:32:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 18:32:10 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 18:38:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 18:38:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 18:38:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 18:38:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 18:38:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 18:38:13 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 18:44:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 18:44:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 18:44:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 18:44:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 18:44:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 18:44:15 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 18:50:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 18:50:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state QUEUED (PENDING) -2025-07-28 18:50:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state QUEUED (PENDING) -2025-07-28 18:50:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 18:50:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 18:50:18 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 18:56:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 18:56:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state RUNNING (RUNNING) -2025-07-28 18:56:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state RUNNING (RUNNING) -2025-07-28 18:56:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 18:56:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 18:56:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 19:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 19:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state RUNNING (RUNNING) -2025-07-28 19:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state RUNNING (RUNNING) -2025-07-28 19:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 19:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 19:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 19:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_prod_mem001_f000 -2025-07-28 19:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_prod_mem002_f000 -2025-07-28 19:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_atmos_prod_mem001_f000 succeeded, jobid=5950993 -2025-07-28 19:02:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of sfs_atmos_prod_mem002_f000 is pending at druby://130.18.14.151:34625 -2025-07-28 19:08:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending sfs_atmos_prod_mem002_f000 is success, jobid=5950994 -2025-07-28 19:08:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 19:08:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state RUNNING (RUNNING) -2025-07-28 19:08:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state RUNNING (RUNNING) -2025-07-28 19:08:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 19:08:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 19:08:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 19:08:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f000, jobid=5950993, in state QUEUED (PENDING) -2025-07-28 19:08:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f000, jobid=5950994, in state QUEUED (PENDING) -2025-07-28 19:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_prod_mem001_f024 -2025-07-28 19:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_prod_mem002_f024 -2025-07-28 19:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ice_prod_mem001_f024 -2025-07-28 19:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ice_prod_mem002_f024 -2025-07-28 19:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_atmos_prod_mem001_f024 succeeded, jobid=5951056 -2025-07-28 19:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_atmos_prod_mem002_f024 succeeded, jobid=5951057 -2025-07-28 19:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ice_prod_mem001_f024 succeeded, jobid=5951058 -2025-07-28 19:08:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ice_prod_mem002_f024 succeeded, jobid=5951059 -2025-07-28 19:14:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 19:14:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state RUNNING (RUNNING) -2025-07-28 19:14:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state RUNNING (RUNNING) -2025-07-28 19:14:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 19:14:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 19:14:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 19:14:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f000, jobid=5950993, in state QUEUED (PENDING) -2025-07-28 19:14:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f024, jobid=5951056, in state QUEUED (PENDING) -2025-07-28 19:14:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f000, jobid=5950994, in state QUEUED (PENDING) -2025-07-28 19:14:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f024, jobid=5951057, in state QUEUED (PENDING) -2025-07-28 19:14:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ice_prod_mem001_f024, jobid=5951058, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:14:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ice_prod_mem002_f024, jobid=5951059, in state SUCCEEDED (COMPLETED), ran for 31.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg0, jobid=5950287, in state SUCCEEDED (COMPLETED), ran for 1361.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg0, jobid=5950288, in state SUCCEEDED (COMPLETED), ran for 1313.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f000, jobid=5950993, in state QUEUED (PENDING) -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f024, jobid=5951056, in state QUEUED (PENDING) -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f000, jobid=5950994, in state QUEUED (PENDING) -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f024, jobid=5951057, in state QUEUED (PENDING) -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_fcst_mem001_seg1 -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_fcst_mem002_seg1 -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_prod_mem001_f048 -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_prod_mem002_f048 -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ocean_prod_mem001_f024 -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ocean_prod_mem001_f048 -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ocean_prod_mem002_f024 -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ocean_prod_mem002_f048 -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ice_prod_mem001_f048 -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ice_prod_mem002_f048 -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_fcst_mem001_seg1 succeeded, jobid=5951105 -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_fcst_mem002_seg1 succeeded, jobid=5951106 -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_atmos_prod_mem001_f048 succeeded, jobid=5951107 -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_atmos_prod_mem002_f048 succeeded, jobid=5951108 -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ocean_prod_mem001_f024 succeeded, jobid=5951109 -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ocean_prod_mem001_f048 succeeded, jobid=5951110 -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ocean_prod_mem002_f024 succeeded, jobid=5951111 -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ocean_prod_mem002_f048 succeeded, jobid=5951112 -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ice_prod_mem001_f048 succeeded, jobid=5951113 -2025-07-28 19:20:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ice_prod_mem002_f048 succeeded, jobid=5951114 -2025-07-28 19:26:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 19:26:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state QUEUED (PENDING) -2025-07-28 19:26:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state QUEUED (PENDING) -2025-07-28 19:26:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 19:26:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 19:26:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 19:26:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f000, jobid=5950993, in state QUEUED (PENDING) -2025-07-28 19:26:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f024, jobid=5951056, in state QUEUED (PENDING) -2025-07-28 19:26:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f048, jobid=5951107, in state QUEUED (PENDING) -2025-07-28 19:26:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f000, jobid=5950994, in state QUEUED (PENDING) -2025-07-28 19:26:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f024, jobid=5951057, in state QUEUED (PENDING) -2025-07-28 19:26:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f048, jobid=5951108, in state QUEUED (PENDING) -2025-07-28 19:26:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ocean_prod_mem001_f024, jobid=5951109, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:26:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ocean_prod_mem001_f048, jobid=5951110, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:26:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ocean_prod_mem002_f024, jobid=5951111, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:26:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ocean_prod_mem002_f048, jobid=5951112, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:26:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ice_prod_mem001_f048, jobid=5951113, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:26:34 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ice_prod_mem002_f048, jobid=5951114, in state SUCCEEDED (COMPLETED), ran for 41.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:32:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 19:32:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state QUEUED (PENDING) -2025-07-28 19:32:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state QUEUED (PENDING) -2025-07-28 19:32:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 19:32:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 19:32:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 19:32:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f000, jobid=5950993, in state QUEUED (PENDING) -2025-07-28 19:32:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f024, jobid=5951056, in state QUEUED (PENDING) -2025-07-28 19:32:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f048, jobid=5951107, in state QUEUED (PENDING) -2025-07-28 19:32:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f000, jobid=5950994, in state QUEUED (PENDING) -2025-07-28 19:32:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f024, jobid=5951057, in state QUEUED (PENDING) -2025-07-28 19:32:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f048, jobid=5951108, in state QUEUED (PENDING) -2025-07-28 19:38:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 19:38:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state QUEUED (PENDING) -2025-07-28 19:38:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state QUEUED (PENDING) -2025-07-28 19:38:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state QUEUED (PENDING) -2025-07-28 19:38:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state QUEUED (PENDING) -2025-07-28 19:38:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 19:38:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f000, jobid=5950993, in state QUEUED (PENDING) -2025-07-28 19:38:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f024, jobid=5951056, in state QUEUED (PENDING) -2025-07-28 19:38:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f048, jobid=5951107, in state QUEUED (PENDING) -2025-07-28 19:38:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f000, jobid=5950994, in state QUEUED (PENDING) -2025-07-28 19:38:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f024, jobid=5951057, in state QUEUED (PENDING) -2025-07-28 19:38:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f048, jobid=5951108, in state QUEUED (PENDING) -2025-07-28 19:44:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 19:44:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state QUEUED (PENDING) -2025-07-28 19:44:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state QUEUED (PENDING) -2025-07-28 19:44:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f000, jobid=5950560, in state SUCCEEDED (COMPLETED), ran for 92.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:44:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state RUNNING (RUNNING) -2025-07-28 19:44:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state QUEUED (PENDING) -2025-07-28 19:44:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f000, jobid=5950993, in state QUEUED (PENDING) -2025-07-28 19:44:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f024, jobid=5951056, in state QUEUED (PENDING) -2025-07-28 19:44:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f048, jobid=5951107, in state QUEUED (PENDING) -2025-07-28 19:44:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f000, jobid=5950994, in state QUEUED (PENDING) -2025-07-28 19:44:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f024, jobid=5951057, in state QUEUED (PENDING) -2025-07-28 19:44:41 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f048, jobid=5951108, in state QUEUED (PENDING) -2025-07-28 19:50:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 19:50:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state QUEUED (PENDING) -2025-07-28 19:50:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state QUEUED (PENDING) -2025-07-28 19:50:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f024, jobid=5950575, in state SUCCEEDED (COMPLETED), ran for 93.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:50:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f048, jobid=5950611, in state SUCCEEDED (COMPLETED), ran for 93.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:50:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f000, jobid=5950993, in state QUEUED (PENDING) -2025-07-28 19:50:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f024, jobid=5951056, in state QUEUED (PENDING) -2025-07-28 19:50:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f048, jobid=5951107, in state QUEUED (PENDING) -2025-07-28 19:50:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f000, jobid=5950994, in state QUEUED (PENDING) -2025-07-28 19:50:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f024, jobid=5951057, in state QUEUED (PENDING) -2025-07-28 19:50:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f048, jobid=5951108, in state QUEUED (PENDING) -2025-07-28 19:56:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 19:56:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state QUEUED (PENDING) -2025-07-28 19:56:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state QUEUED (PENDING) -2025-07-28 19:56:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f000, jobid=5950993, in state QUEUED (PENDING) -2025-07-28 19:56:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f024, jobid=5951056, in state QUEUED (PENDING) -2025-07-28 19:56:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f048, jobid=5951107, in state QUEUED (PENDING) -2025-07-28 19:56:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f000, jobid=5950994, in state QUEUED (PENDING) -2025-07-28 19:56:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f024, jobid=5951057, in state QUEUED (PENDING) -2025-07-28 19:56:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f048, jobid=5951108, in state QUEUED (PENDING) -2025-07-28 20:02:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 20:02:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state QUEUED (PENDING) -2025-07-28 20:02:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state QUEUED (PENDING) -2025-07-28 20:02:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f000, jobid=5950993, in state SUCCEEDED (COMPLETED), ran for 94.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 20:02:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f024, jobid=5951056, in state QUEUED (PENDING) -2025-07-28 20:02:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f048, jobid=5951107, in state QUEUED (PENDING) -2025-07-28 20:02:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f000, jobid=5950994, in state QUEUED (PENDING) -2025-07-28 20:02:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f024, jobid=5951057, in state QUEUED (PENDING) -2025-07-28 20:02:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f048, jobid=5951108, in state QUEUED (PENDING) -2025-07-28 20:08:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 20:08:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state QUEUED (PENDING) -2025-07-28 20:08:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state QUEUED (PENDING) -2025-07-28 20:08:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f024, jobid=5951056, in state QUEUED (PENDING) -2025-07-28 20:08:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f048, jobid=5951107, in state QUEUED (PENDING) -2025-07-28 20:08:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f000, jobid=5950994, in state QUEUED (PENDING) -2025-07-28 20:08:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f024, jobid=5951057, in state QUEUED (PENDING) -2025-07-28 20:08:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f048, jobid=5951108, in state QUEUED (PENDING) -2025-07-28 20:14:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 20:14:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state QUEUED (PENDING) -2025-07-28 20:14:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state QUEUED (PENDING) -2025-07-28 20:14:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f024, jobid=5951056, in state QUEUED (PENDING) -2025-07-28 20:14:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f048, jobid=5951107, in state QUEUED (PENDING) -2025-07-28 20:14:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f000, jobid=5950994, in state QUEUED (PENDING) -2025-07-28 20:14:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f024, jobid=5951057, in state QUEUED (PENDING) -2025-07-28 20:14:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f048, jobid=5951108, in state QUEUED (PENDING) -2025-07-28 20:20:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 20:20:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state QUEUED (PENDING) -2025-07-28 20:20:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state QUEUED (PENDING) -2025-07-28 20:20:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f024, jobid=5951056, in state QUEUED (PENDING) -2025-07-28 20:20:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f048, jobid=5951107, in state QUEUED (PENDING) -2025-07-28 20:20:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f000, jobid=5950994, in state QUEUED (PENDING) -2025-07-28 20:20:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f024, jobid=5951057, in state QUEUED (PENDING) -2025-07-28 20:20:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f048, jobid=5951108, in state QUEUED (PENDING) -2025-07-28 20:26:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 20:26:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state QUEUED (PENDING) -2025-07-28 20:26:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state QUEUED (PENDING) -2025-07-28 20:26:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f024, jobid=5951056, in state QUEUED (PENDING) -2025-07-28 20:26:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f048, jobid=5951107, in state QUEUED (PENDING) -2025-07-28 20:26:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f000, jobid=5950994, in state SUCCEEDED (COMPLETED), ran for 94.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 20:26:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f024, jobid=5951057, in state QUEUED (PENDING) -2025-07-28 20:26:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f048, jobid=5951108, in state QUEUED (PENDING) -2025-07-28 20:26:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_ensstat_f000 -2025-07-28 20:26:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of sfs_atmos_ensstat_f000 is pending at druby://130.18.14.151:46077 -2025-07-28 20:33:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending sfs_atmos_ensstat_f000 is success, jobid=5951343 -2025-07-28 20:33:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 20:33:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state QUEUED (PENDING) -2025-07-28 20:33:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state QUEUED (PENDING) -2025-07-28 20:33:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f024, jobid=5951056, in state QUEUED (PENDING) -2025-07-28 20:33:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f048, jobid=5951107, in state QUEUED (PENDING) -2025-07-28 20:33:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f024, jobid=5951057, in state QUEUED (PENDING) -2025-07-28 20:33:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f048, jobid=5951108, in state QUEUED (PENDING) -2025-07-28 20:33:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f000, jobid=5951343, in state QUEUED (PENDING) -2025-07-28 20:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 20:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state QUEUED (PENDING) -2025-07-28 20:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state QUEUED (PENDING) -2025-07-28 20:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f024, jobid=5951056, in state QUEUED (PENDING) -2025-07-28 20:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f048, jobid=5951107, in state QUEUED (PENDING) -2025-07-28 20:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f024, jobid=5951057, in state QUEUED (PENDING) -2025-07-28 20:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f048, jobid=5951108, in state QUEUED (PENDING) -2025-07-28 20:39:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f000, jobid=5951343, in state QUEUED (PENDING) -2025-07-28 20:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 20:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state QUEUED (PENDING) -2025-07-28 20:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state QUEUED (PENDING) -2025-07-28 20:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f024, jobid=5951056, in state QUEUED (PENDING) -2025-07-28 20:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f048, jobid=5951107, in state QUEUED (PENDING) -2025-07-28 20:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f024, jobid=5951057, in state QUEUED (PENDING) -2025-07-28 20:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f048, jobid=5951108, in state QUEUED (PENDING) -2025-07-28 20:45:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f000, jobid=5951343, in state QUEUED (PENDING) -2025-07-28 20:51:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 20:51:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state QUEUED (PENDING) -2025-07-28 20:51:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state QUEUED (PENDING) -2025-07-28 20:51:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f024, jobid=5951056, in state QUEUED (PENDING) -2025-07-28 20:51:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f048, jobid=5951107, in state QUEUED (PENDING) -2025-07-28 20:51:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f024, jobid=5951057, in state QUEUED (PENDING) -2025-07-28 20:51:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f048, jobid=5951108, in state QUEUED (PENDING) -2025-07-28 20:51:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f000, jobid=5951343, in state QUEUED (PENDING) -2025-07-28 20:57:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 20:57:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state QUEUED (PENDING) -2025-07-28 20:57:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state QUEUED (PENDING) -2025-07-28 20:57:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f024, jobid=5951056, in state QUEUED (PENDING) -2025-07-28 20:57:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f048, jobid=5951107, in state QUEUED (PENDING) -2025-07-28 20:57:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f024, jobid=5951057, in state QUEUED (PENDING) -2025-07-28 20:57:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f048, jobid=5951108, in state QUEUED (PENDING) -2025-07-28 20:57:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f000, jobid=5951343, in state QUEUED (PENDING) -2025-07-28 21:03:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state QUEUED (PENDING) -2025-07-28 21:03:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state QUEUED (PENDING) -2025-07-28 21:03:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state QUEUED (PENDING) -2025-07-28 21:03:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f024, jobid=5951056, in state QUEUED (PENDING) -2025-07-28 21:03:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f048, jobid=5951107, in state QUEUED (PENDING) -2025-07-28 21:03:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f024, jobid=5951057, in state QUEUED (PENDING) -2025-07-28 21:03:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f048, jobid=5951108, in state QUEUED (PENDING) -2025-07-28 21:03:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f000, jobid=5951343, in state QUEUED (PENDING) -2025-07-28 21:09:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state RUNNING (RUNNING) -2025-07-28 21:09:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state QUEUED (PENDING) -2025-07-28 21:09:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state QUEUED (PENDING) -2025-07-28 21:09:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f024, jobid=5951056, in state RUNNING (RUNNING) -2025-07-28 21:09:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f048, jobid=5951107, in state QUEUED (PENDING) -2025-07-28 21:09:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f024, jobid=5951057, in state RUNNING (RUNNING) -2025-07-28 21:09:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f048, jobid=5951108, in state QUEUED (PENDING) -2025-07-28 21:09:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f000, jobid=5951343, in state QUEUED (PENDING) -2025-07-28 21:15:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state RUNNING (RUNNING) -2025-07-28 21:15:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state QUEUED (PENDING) -2025-07-28 21:15:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state QUEUED (PENDING) -2025-07-28 21:15:19 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f024, jobid=5951056, in state SUCCEEDED (COMPLETED), ran for 96.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:15:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f048, jobid=5951107, in state QUEUED (PENDING) -2025-07-28 21:15:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f024, jobid=5951057, in state SUCCEEDED (COMPLETED), ran for 95.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:15:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f048, jobid=5951108, in state QUEUED (PENDING) -2025-07-28 21:15:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f000, jobid=5951343, in state QUEUED (PENDING) -2025-07-28 21:15:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_ensstat_f024 -2025-07-28 21:15:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of sfs_atmos_ensstat_f024 is pending at druby://130.18.14.151:41741 -2025-07-28 21:21:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending sfs_atmos_ensstat_f024 is success, jobid=5951564 -2025-07-28 21:21:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state RUNNING (RUNNING) -2025-07-28 21:21:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state QUEUED (PENDING) -2025-07-28 21:21:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state QUEUED (PENDING) -2025-07-28 21:21:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f048, jobid=5951107, in state QUEUED (PENDING) -2025-07-28 21:21:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f048, jobid=5951108, in state QUEUED (PENDING) -2025-07-28 21:21:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f000, jobid=5951343, in state QUEUED (PENDING) -2025-07-28 21:21:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f024, jobid=5951564, in state QUEUED (PENDING) -2025-07-28 21:21:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_prod_mem000_f072 -2025-07-28 21:21:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ice_prod_mem000_f072 -2025-07-28 21:21:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_atmos_prod_mem000_f072 succeeded, jobid=5951608 -2025-07-28 21:21:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of sfs_ice_prod_mem000_f072 is pending at druby://130.18.14.151:33203 -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending sfs_ice_prod_mem000_f072 is success, jobid=5951609 -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg1, jobid=5950610, in state SUCCEEDED (COMPLETED), ran for 1382.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state RUNNING (RUNNING) -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state RUNNING (RUNNING) -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f072, jobid=5951608, in state SUCCEEDED (COMPLETED), ran for 98.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f048, jobid=5951107, in state SUCCEEDED (COMPLETED), ran for 98.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f048, jobid=5951108, in state SUCCEEDED (COMPLETED), ran for 97.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f000, jobid=5951343, in state SUCCEEDED (COMPLETED), ran for 30.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f024, jobid=5951564, in state SUCCEEDED (COMPLETED), ran for 34.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ice_prod_mem000_f072, jobid=5951609, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_fcst_mem000_seg2 -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_prod_mem000_f096 -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_ensstat_f048 -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ocean_prod_mem000_f072 -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ocean_prod_mem000_f096 -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ice_prod_mem000_f096 -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_fcst_mem000_seg2 succeeded, jobid=5951661 -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_atmos_prod_mem000_f096 succeeded, jobid=5951662 -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_atmos_ensstat_f048 succeeded, jobid=5951663 -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ocean_prod_mem000_f072 succeeded, jobid=5951664 -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ocean_prod_mem000_f096 succeeded, jobid=5951665 -2025-07-28 21:27:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ice_prod_mem000_f096 succeeded, jobid=5951666 -2025-07-28 21:33:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg2, jobid=5951661, in state QUEUED (PENDING) -2025-07-28 21:33:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state RUNNING (RUNNING) -2025-07-28 21:33:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state RUNNING (RUNNING) -2025-07-28 21:33:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f096, jobid=5951662, in state SUCCEEDED (COMPLETED), ran for 93.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f048, jobid=5951663, in state SUCCEEDED (COMPLETED), ran for 31.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ocean_prod_mem000_f072, jobid=5951664, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ocean_prod_mem000_f096, jobid=5951665, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ice_prod_mem000_f096, jobid=5951666, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:39:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg2, jobid=5951661, in state RUNNING (RUNNING) -2025-07-28 21:39:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state RUNNING (RUNNING) -2025-07-28 21:39:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state RUNNING (RUNNING) -2025-07-28 21:39:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_prod_mem001_f072 -2025-07-28 21:39:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_prod_mem002_f072 -2025-07-28 21:39:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ice_prod_mem001_f072 -2025-07-28 21:39:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ice_prod_mem002_f072 -2025-07-28 21:39:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_atmos_prod_mem001_f072 succeeded, jobid=5951776 -2025-07-28 21:39:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_atmos_prod_mem002_f072 succeeded, jobid=5951777 -2025-07-28 21:39:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ice_prod_mem001_f072 succeeded, jobid=5951778 -2025-07-28 21:39:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ice_prod_mem002_f072 succeeded, jobid=5951779 -2025-07-28 21:45:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg2, jobid=5951661, in state RUNNING (RUNNING) -2025-07-28 21:45:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state RUNNING (RUNNING) -2025-07-28 21:45:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state RUNNING (RUNNING) -2025-07-28 21:45:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f072, jobid=5951776, in state QUEUED (PENDING) -2025-07-28 21:45:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f072, jobid=5951777, in state QUEUED (PENDING) -2025-07-28 21:45:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ice_prod_mem001_f072, jobid=5951778, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:45:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ice_prod_mem002_f072, jobid=5951779, in state SUCCEEDED (COMPLETED), ran for 21.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg2, jobid=5951661, in state RUNNING (RUNNING) -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg1, jobid=5951105, in state SUCCEEDED (COMPLETED), ran for 1361.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg1, jobid=5951106, in state SUCCEEDED (COMPLETED), ran for 1357.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f072, jobid=5951776, in state QUEUED (PENDING) -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f072, jobid=5951777, in state QUEUED (PENDING) -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_fcst_mem001_seg2 -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_fcst_mem002_seg2 -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_prod_mem001_f096 -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_prod_mem002_f096 -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ocean_prod_mem001_f072 -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ocean_prod_mem001_f096 -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ocean_prod_mem002_f072 -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ocean_prod_mem002_f096 -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ice_prod_mem001_f096 -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ice_prod_mem002_f096 -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_fcst_mem001_seg2 succeeded, jobid=5951868 -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_fcst_mem002_seg2 succeeded, jobid=5951869 -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_atmos_prod_mem001_f096 succeeded, jobid=5951870 -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_atmos_prod_mem002_f096 succeeded, jobid=5951871 -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ocean_prod_mem001_f072 succeeded, jobid=5951872 -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ocean_prod_mem001_f096 succeeded, jobid=5951873 -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ocean_prod_mem002_f072 succeeded, jobid=5951874 -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ocean_prod_mem002_f096 succeeded, jobid=5951875 -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ice_prod_mem001_f096 succeeded, jobid=5951876 -2025-07-28 21:51:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ice_prod_mem002_f096 succeeded, jobid=5951877 -2025-07-28 21:57:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg2, jobid=5951661, in state RUNNING (RUNNING) -2025-07-28 21:57:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg2, jobid=5951868, in state QUEUED (PENDING) -2025-07-28 21:57:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg2, jobid=5951869, in state QUEUED (PENDING) -2025-07-28 21:57:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f072, jobid=5951776, in state QUEUED (PENDING) -2025-07-28 21:57:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f096, jobid=5951870, in state QUEUED (PENDING) -2025-07-28 21:57:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f072, jobid=5951777, in state QUEUED (PENDING) -2025-07-28 21:57:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f096, jobid=5951871, in state QUEUED (PENDING) -2025-07-28 21:57:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ocean_prod_mem001_f072, jobid=5951872, in state SUCCEEDED (COMPLETED), ran for 16.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:57:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ocean_prod_mem001_f096, jobid=5951873, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:57:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ocean_prod_mem002_f072, jobid=5951874, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:57:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ocean_prod_mem002_f096, jobid=5951875, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:57:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ice_prod_mem001_f096, jobid=5951876, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:57:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ice_prod_mem002_f096, jobid=5951877, in state SUCCEEDED (COMPLETED), ran for 49.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:57:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_prod_mem000_f120 -2025-07-28 21:57:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ice_prod_mem000_f120 -2025-07-28 21:57:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of sfs_atmos_prod_mem000_f120 is pending at druby://130.18.14.151:46243 -2025-07-28 21:57:38 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of sfs_ice_prod_mem000_f120 is pending at druby://130.18.14.151:46243 -2025-07-28 22:03:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending sfs_atmos_prod_mem000_f120 is success, jobid=5951899 -2025-07-28 22:03:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending sfs_ice_prod_mem000_f120 is success, jobid=5951900 -2025-07-28 22:03:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem000_seg2, jobid=5951661, in state SUCCEEDED (COMPLETED), ran for 1369.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:03:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg2, jobid=5951868, in state QUEUED (PENDING) -2025-07-28 22:03:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg2, jobid=5951869, in state QUEUED (PENDING) -2025-07-28 22:03:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f120, jobid=5951899, in state QUEUED (PENDING) -2025-07-28 22:03:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f072, jobid=5951776, in state QUEUED (PENDING) -2025-07-28 22:03:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f096, jobid=5951870, in state QUEUED (PENDING) -2025-07-28 22:03:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f072, jobid=5951777, in state QUEUED (PENDING) -2025-07-28 22:03:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f096, jobid=5951871, in state QUEUED (PENDING) -2025-07-28 22:03:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ice_prod_mem000_f120, jobid=5951900, in state SUCCEEDED (COMPLETED), ran for 15.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:03:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_prod_mem000_f144 -2025-07-28 22:03:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ocean_prod_mem000_f120 -2025-07-28 22:03:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ocean_prod_mem000_f144 -2025-07-28 22:03:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ice_prod_mem000_f144 -2025-07-28 22:03:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_atmos_prod_mem000_f144 succeeded, jobid=5951912 -2025-07-28 22:03:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ocean_prod_mem000_f120 succeeded, jobid=5951913 -2025-07-28 22:03:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ocean_prod_mem000_f144 succeeded, jobid=5951914 -2025-07-28 22:03:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ice_prod_mem000_f144 succeeded, jobid=5951915 -2025-07-28 22:09:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg2, jobid=5951868, in state QUEUED (PENDING) -2025-07-28 22:09:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg2, jobid=5951869, in state QUEUED (PENDING) -2025-07-28 22:09:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f120, jobid=5951899, in state QUEUED (PENDING) -2025-07-28 22:09:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f144, jobid=5951912, in state QUEUED (PENDING) -2025-07-28 22:09:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f072, jobid=5951776, in state SUCCEEDED (COMPLETED), ran for 91.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:09:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f096, jobid=5951870, in state QUEUED (PENDING) -2025-07-28 22:09:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f072, jobid=5951777, in state SUCCEEDED (COMPLETED), ran for 88.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:09:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f096, jobid=5951871, in state QUEUED (PENDING) -2025-07-28 22:09:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ocean_prod_mem000_f120, jobid=5951913, in state SUCCEEDED (COMPLETED), ran for 15.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:09:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ocean_prod_mem000_f144, jobid=5951914, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:09:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ice_prod_mem000_f144, jobid=5951915, in state SUCCEEDED (COMPLETED), ran for 16.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:09:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_ensstat_f072 -2025-07-28 22:09:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of sfs_atmos_ensstat_f072 is pending at druby://130.18.14.151:39601 -2025-07-28 22:15:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending sfs_atmos_ensstat_f072 is success, jobid=5951954 -2025-07-28 22:15:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg2, jobid=5951868, in state QUEUED (PENDING) -2025-07-28 22:15:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg2, jobid=5951869, in state QUEUED (PENDING) -2025-07-28 22:15:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f120, jobid=5951899, in state QUEUED (PENDING) -2025-07-28 22:15:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f144, jobid=5951912, in state QUEUED (PENDING) -2025-07-28 22:15:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f096, jobid=5951870, in state QUEUED (PENDING) -2025-07-28 22:15:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f096, jobid=5951871, in state QUEUED (PENDING) -2025-07-28 22:15:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f072, jobid=5951954, in state QUEUED (PENDING) -2025-07-28 22:21:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg2, jobid=5951868, in state QUEUED (PENDING) -2025-07-28 22:21:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg2, jobid=5951869, in state QUEUED (PENDING) -2025-07-28 22:21:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f120, jobid=5951899, in state QUEUED (PENDING) -2025-07-28 22:21:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f144, jobid=5951912, in state QUEUED (PENDING) -2025-07-28 22:21:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f096, jobid=5951870, in state SUCCEEDED (COMPLETED), ran for 94.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:21:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f096, jobid=5951871, in state SUCCEEDED (COMPLETED), ran for 90.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:21:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f072, jobid=5951954, in state QUEUED (PENDING) -2025-07-28 22:21:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_ensstat_f096 -2025-07-28 22:21:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of sfs_atmos_ensstat_f096 is pending at druby://130.18.14.151:34745 -2025-07-28 22:27:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending sfs_atmos_ensstat_f096 is success, jobid=5952002 -2025-07-28 22:27:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg2, jobid=5951868, in state QUEUED (PENDING) -2025-07-28 22:27:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg2, jobid=5951869, in state QUEUED (PENDING) -2025-07-28 22:27:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f120, jobid=5951899, in state QUEUED (PENDING) -2025-07-28 22:27:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f144, jobid=5951912, in state QUEUED (PENDING) -2025-07-28 22:27:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f072, jobid=5951954, in state QUEUED (PENDING) -2025-07-28 22:27:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f096, jobid=5952002, in state QUEUED (PENDING) -2025-07-28 22:33:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg2, jobid=5951868, in state QUEUED (PENDING) -2025-07-28 22:33:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg2, jobid=5951869, in state QUEUED (PENDING) -2025-07-28 22:33:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f120, jobid=5951899, in state SUCCEEDED (COMPLETED), ran for 92.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:33:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f144, jobid=5951912, in state QUEUED (PENDING) -2025-07-28 22:33:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f072, jobid=5951954, in state QUEUED (PENDING) -2025-07-28 22:33:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f096, jobid=5952002, in state QUEUED (PENDING) -2025-07-28 22:39:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg2, jobid=5951868, in state QUEUED (PENDING) -2025-07-28 22:39:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg2, jobid=5951869, in state QUEUED (PENDING) -2025-07-28 22:39:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f144, jobid=5951912, in state QUEUED (PENDING) -2025-07-28 22:39:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f072, jobid=5951954, in state QUEUED (PENDING) -2025-07-28 22:39:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f096, jobid=5952002, in state QUEUED (PENDING) -2025-07-28 22:45:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg2, jobid=5951868, in state QUEUED (PENDING) -2025-07-28 22:45:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg2, jobid=5951869, in state QUEUED (PENDING) -2025-07-28 22:45:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f144, jobid=5951912, in state QUEUED (PENDING) -2025-07-28 22:45:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f072, jobid=5951954, in state QUEUED (PENDING) -2025-07-28 22:45:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f096, jobid=5952002, in state QUEUED (PENDING) -2025-07-28 22:52:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg2, jobid=5951868, in state QUEUED (PENDING) -2025-07-28 22:52:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg2, jobid=5951869, in state QUEUED (PENDING) -2025-07-28 22:52:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem000_f144, jobid=5951912, in state SUCCEEDED (COMPLETED), ran for 95.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:52:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f072, jobid=5951954, in state QUEUED (PENDING) -2025-07-28 22:52:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f096, jobid=5952002, in state QUEUED (PENDING) -2025-07-28 22:58:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg2, jobid=5951868, in state RUNNING (RUNNING) -2025-07-28 22:58:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg2, jobid=5951869, in state QUEUED (PENDING) -2025-07-28 22:58:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f072, jobid=5951954, in state QUEUED (PENDING) -2025-07-28 22:58:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f096, jobid=5952002, in state QUEUED (PENDING) -2025-07-28 23:04:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg2, jobid=5951868, in state RUNNING (RUNNING) -2025-07-28 23:04:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg2, jobid=5951869, in state QUEUED (PENDING) -2025-07-28 23:04:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f072, jobid=5951954, in state QUEUED (PENDING) -2025-07-28 23:04:07 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f096, jobid=5952002, in state QUEUED (PENDING) -2025-07-28 23:10:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg2, jobid=5951868, in state RUNNING (RUNNING) -2025-07-28 23:10:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg2, jobid=5951869, in state RUNNING (RUNNING) -2025-07-28 23:10:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f072, jobid=5951954, in state QUEUED (PENDING) -2025-07-28 23:10:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f096, jobid=5952002, in state QUEUED (PENDING) -2025-07-28 23:10:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_prod_mem001_f120 -2025-07-28 23:10:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ice_prod_mem001_f120 -2025-07-28 23:10:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of sfs_atmos_prod_mem001_f120 is pending at druby://130.18.14.151:44389 -2025-07-28 23:10:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of sfs_ice_prod_mem001_f120 is pending at druby://130.18.14.151:44389 -2025-07-28 23:16:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending sfs_atmos_prod_mem001_f120 is success, jobid=5952116 -2025-07-28 23:16:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending sfs_ice_prod_mem001_f120 is success, jobid=5952117 -2025-07-28 23:16:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg2, jobid=5951868, in state RUNNING (RUNNING) -2025-07-28 23:16:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg2, jobid=5951869, in state RUNNING (RUNNING) -2025-07-28 23:16:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f120, jobid=5952116, in state QUEUED (PENDING) -2025-07-28 23:16:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f072, jobid=5951954, in state QUEUED (PENDING) -2025-07-28 23:16:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f096, jobid=5952002, in state QUEUED (PENDING) -2025-07-28 23:16:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ice_prod_mem001_f120, jobid=5952117, in state SUCCEEDED (COMPLETED), ran for 14.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:22:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem001_seg2, jobid=5951868, in state SUCCEEDED (COMPLETED), ran for 1356.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:22:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg2, jobid=5951869, in state RUNNING (RUNNING) -2025-07-28 23:22:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f120, jobid=5952116, in state QUEUED (PENDING) -2025-07-28 23:22:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f072, jobid=5951954, in state SUCCEEDED (COMPLETED), ran for 30.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:22:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f096, jobid=5952002, in state QUEUED (PENDING) -2025-07-28 23:22:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_prod_mem001_f144 -2025-07-28 23:22:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_prod_mem002_f120 -2025-07-28 23:22:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ocean_prod_mem001_f120 -2025-07-28 23:22:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ocean_prod_mem001_f144 -2025-07-28 23:22:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ice_prod_mem001_f144 -2025-07-28 23:22:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ice_prod_mem002_f120 -2025-07-28 23:22:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_atmos_prod_mem001_f144 succeeded, jobid=5952146 -2025-07-28 23:22:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_atmos_prod_mem002_f120 succeeded, jobid=5952147 -2025-07-28 23:22:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ocean_prod_mem001_f120 succeeded, jobid=5952148 -2025-07-28 23:22:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ocean_prod_mem001_f144 succeeded, jobid=5952149 -2025-07-28 23:22:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ice_prod_mem001_f144 succeeded, jobid=5952150 -2025-07-28 23:22:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ice_prod_mem002_f120 succeeded, jobid=5952151 -2025-07-28 23:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_fcst_mem002_seg2, jobid=5951869, in state SUCCEEDED (COMPLETED), ran for 1344.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f120, jobid=5952116, in state QUEUED (PENDING) -2025-07-28 23:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f144, jobid=5952146, in state QUEUED (PENDING) -2025-07-28 23:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f120, jobid=5952147, in state QUEUED (PENDING) -2025-07-28 23:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f096, jobid=5952002, in state QUEUED (PENDING) -2025-07-28 23:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ocean_prod_mem001_f120, jobid=5952148, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ocean_prod_mem001_f144, jobid=5952149, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ice_prod_mem001_f144, jobid=5952150, in state SUCCEEDED (COMPLETED), ran for 15.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ice_prod_mem002_f120, jobid=5952151, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_prod_mem002_f144 -2025-07-28 23:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ocean_prod_mem002_f120 -2025-07-28 23:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ocean_prod_mem002_f144 -2025-07-28 23:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_ice_prod_mem002_f144 -2025-07-28 23:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_atmos_prod_mem002_f144 succeeded, jobid=5952158 -2025-07-28 23:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ocean_prod_mem002_f120 succeeded, jobid=5952159 -2025-07-28 23:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ocean_prod_mem002_f144 succeeded, jobid=5952160 -2025-07-28 23:28:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of sfs_ice_prod_mem002_f144 succeeded, jobid=5952161 -2025-07-28 23:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f120, jobid=5952116, in state QUEUED (PENDING) -2025-07-28 23:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f144, jobid=5952146, in state QUEUED (PENDING) -2025-07-28 23:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f120, jobid=5952147, in state QUEUED (PENDING) -2025-07-28 23:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f144, jobid=5952158, in state QUEUED (PENDING) -2025-07-28 23:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f096, jobid=5952002, in state QUEUED (PENDING) -2025-07-28 23:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ocean_prod_mem002_f120, jobid=5952159, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ocean_prod_mem002_f144, jobid=5952160, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:34:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_ice_prod_mem002_f144, jobid=5952161, in state SUCCEEDED (COMPLETED), ran for 16.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 23:40:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f120, jobid=5952116, in state QUEUED (PENDING) -2025-07-28 23:40:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f144, jobid=5952146, in state QUEUED (PENDING) -2025-07-28 23:40:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f120, jobid=5952147, in state QUEUED (PENDING) -2025-07-28 23:40:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f144, jobid=5952158, in state QUEUED (PENDING) -2025-07-28 23:40:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f096, jobid=5952002, in state QUEUED (PENDING) -2025-07-28 23:46:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f120, jobid=5952116, in state QUEUED (PENDING) -2025-07-28 23:46:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f144, jobid=5952146, in state QUEUED (PENDING) -2025-07-28 23:46:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f120, jobid=5952147, in state QUEUED (PENDING) -2025-07-28 23:46:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f144, jobid=5952158, in state QUEUED (PENDING) -2025-07-28 23:46:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f096, jobid=5952002, in state QUEUED (PENDING) -2025-07-28 23:52:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f120, jobid=5952116, in state QUEUED (PENDING) -2025-07-28 23:52:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f144, jobid=5952146, in state QUEUED (PENDING) -2025-07-28 23:52:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f120, jobid=5952147, in state QUEUED (PENDING) -2025-07-28 23:52:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f144, jobid=5952158, in state QUEUED (PENDING) -2025-07-28 23:52:27 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f096, jobid=5952002, in state QUEUED (PENDING) -2025-07-28 23:58:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f120, jobid=5952116, in state QUEUED (PENDING) -2025-07-28 23:58:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f144, jobid=5952146, in state QUEUED (PENDING) -2025-07-28 23:58:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f120, jobid=5952147, in state QUEUED (PENDING) -2025-07-28 23:58:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f144, jobid=5952158, in state QUEUED (PENDING) -2025-07-28 23:58:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f096, jobid=5952002, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:04:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f120, jobid=5952116, in state QUEUED (PENDING) -2025-07-29 00:04:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f144, jobid=5952146, in state QUEUED (PENDING) -2025-07-29 00:04:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f120, jobid=5952147, in state QUEUED (PENDING) -2025-07-29 00:04:32 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f144, jobid=5952158, in state QUEUED (PENDING) -2025-07-29 00:10:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f120, jobid=5952116, in state QUEUED (PENDING) -2025-07-29 00:10:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f144, jobid=5952146, in state QUEUED (PENDING) -2025-07-29 00:10:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f120, jobid=5952147, in state QUEUED (PENDING) -2025-07-29 00:10:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f144, jobid=5952158, in state QUEUED (PENDING) -2025-07-29 00:16:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f120, jobid=5952116, in state SUCCEEDED (COMPLETED), ran for 92.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:16:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f144, jobid=5952146, in state QUEUED (PENDING) -2025-07-29 00:16:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f120, jobid=5952147, in state QUEUED (PENDING) -2025-07-29 00:16:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f144, jobid=5952158, in state QUEUED (PENDING) -2025-07-29 00:22:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem001_f144, jobid=5952146, in state SUCCEEDED (COMPLETED), ran for 94.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:22:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f120, jobid=5952147, in state SUCCEEDED (COMPLETED), ran for 95.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:22:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_prod_mem002_f144, jobid=5952158, in state SUCCEEDED (COMPLETED), ran for 93.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:22:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_ensstat_f120 -2025-07-29 00:22:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_atmos_ensstat_f144 -2025-07-29 00:22:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of sfs_atmos_ensstat_f120 is pending at druby://130.18.14.151:35571 -2025-07-29 00:22:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of sfs_atmos_ensstat_f144 is pending at druby://130.18.14.151:35571 -2025-07-29 00:28:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending sfs_atmos_ensstat_f120 is success, jobid=5952438 -2025-07-29 00:28:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending sfs_atmos_ensstat_f144 is success, jobid=5952439 -2025-07-29 00:28:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f120, jobid=5952438, in state QUEUED (PENDING) -2025-07-29 00:28:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f144, jobid=5952439, in state QUEUED (PENDING) -2025-07-29 00:34:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f120, jobid=5952438, in state QUEUED (PENDING) -2025-07-29 00:34:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f144, jobid=5952439, in state QUEUED (PENDING) -2025-07-29 00:40:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f120, jobid=5952438, in state SUCCEEDED (COMPLETED), ran for 31.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:40:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_atmos_ensstat_f144, jobid=5952439, in state SUCCEEDED (COMPLETED), ran for 33.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:40:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting sfs_cleanup -2025-07-29 00:40:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of sfs_cleanup is pending at druby://130.18.14.151:43303 -2025-07-29 00:46:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending sfs_cleanup is success, jobid=5952570 -2025-07-29 00:46:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task sfs_cleanup, jobid=5952570, in state SUCCEEDED (COMPLETED), ran for 17.0 seconds, exit status=0, try=1 (of 2) -2025-07-29 00:46:50 -0500 :: hercules-login-1.hpc.msstate.edu :: This cycle is complete: Success diff --git a/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C48_ATM_rocotostat.log b/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C48_ATM_rocotostat.log deleted file mode 100644 index 489ff03f..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C48_ATM_rocotostat.log +++ /dev/null @@ -1,306 +0,0 @@ -2025-07-28 14:49:35,218 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 14:49:35,878 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 84/1028698 threads (0.01% utilization) -2025-07-28 14:49:35,878 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 14:49:36,552 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 80/1028698 threads (0.01% utilization) -2025-07-28 14:49:36,552 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 14:49:36,611 - INFO - root : [END] USER_THREAD_COUNT: mterry has 75/1028698 threads (0.01% utilization) -2025-07-28 14:55:37,735 - INFO - root : [START] USER_THREAD_COUNT: mterry has 67/1028698 threads (0.01% utilization) -2025-07-28 14:55:38,389 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 109/1028698 threads (0.01% utilization) -2025-07-28 14:55:38,389 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 14:55:39,036 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 109/1028698 threads (0.01% utilization) -2025-07-28 14:55:39,037 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 14:55:39,099 - INFO - root : [END] USER_THREAD_COUNT: mterry has 101/1028698 threads (0.01% utilization) -2025-07-28 15:01:40,273 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:01:40,926 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 70/1028698 threads (0.01% utilization) -2025-07-28 15:01:40,926 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 15:01:41,567 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 72/1028698 threads (0.01% utilization) -2025-07-28 15:01:41,567 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 15:01:41,626 - INFO - root : [END] USER_THREAD_COUNT: mterry has 61/1028698 threads (0.01% utilization) -2025-07-28 15:07:42,762 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 15:07:43,432 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 15:07:43,432 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 15:07:44,086 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 15:07:44,086 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 15:07:44,150 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 15:13:45,805 - INFO - root : [START] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 15:13:46,489 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 49/1028698 threads (0.0% utilization) -2025-07-28 15:13:46,490 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 15:13:47,134 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 71/1028698 threads (0.01% utilization) -2025-07-28 15:13:47,135 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 15:13:47,192 - INFO - root : [END] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 15:19:48,691 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:19:49,390 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 15:19:49,390 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 15:19:50,036 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 69/1028698 threads (0.01% utilization) -2025-07-28 15:19:50,036 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 15:19:50,092 - INFO - root : [END] USER_THREAD_COUNT: mterry has 60/1028698 threads (0.01% utilization) -2025-07-28 15:25:51,315 - INFO - root : [START] USER_THREAD_COUNT: mterry has 46/1028698 threads (0.0% utilization) -2025-07-28 15:25:51,973 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 58/1028698 threads (0.01% utilization) -2025-07-28 15:25:51,973 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 15:25:52,622 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 70/1028698 threads (0.01% utilization) -2025-07-28 15:25:52,622 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 15:25:52,681 - INFO - root : [END] USER_THREAD_COUNT: mterry has 75/1028698 threads (0.01% utilization) -2025-07-28 15:31:53,842 - INFO - root : [START] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 15:31:54,512 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 85/1028698 threads (0.01% utilization) -2025-07-28 15:31:54,513 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 15:31:55,153 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 82/1028698 threads (0.01% utilization) -2025-07-28 15:31:55,153 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 15:31:55,211 - INFO - root : [END] USER_THREAD_COUNT: mterry has 73/1028698 threads (0.01% utilization) -2025-07-28 15:37:56,536 - INFO - root : [START] USER_THREAD_COUNT: mterry has 33/1028698 threads (0.0% utilization) -2025-07-28 15:37:57,181 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 15:37:57,181 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 15:37:57,815 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 33/1028698 threads (0.0% utilization) -2025-07-28 15:37:57,815 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 15:37:57,870 - INFO - root : [END] USER_THREAD_COUNT: mterry has 34/1028698 threads (0.0% utilization) -2025-07-28 15:43:59,114 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:43:59,770 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 59/1028698 threads (0.01% utilization) -2025-07-28 15:43:59,770 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 15:44:00,421 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 59/1028698 threads (0.01% utilization) -2025-07-28 15:44:00,421 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 15:44:00,481 - INFO - root : [END] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 15:50:01,615 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:50:02,368 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 99/1028698 threads (0.01% utilization) -2025-07-28 15:50:02,368 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 15:50:03,065 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 90/1028698 threads (0.01% utilization) -2025-07-28 15:50:03,065 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 15:50:03,127 - INFO - root : [END] USER_THREAD_COUNT: mterry has 86/1028698 threads (0.01% utilization) -2025-07-28 15:56:04,254 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:56:04,902 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 62/1028698 threads (0.01% utilization) -2025-07-28 15:56:04,902 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 15:56:05,553 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 15:56:05,554 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 15:56:05,613 - INFO - root : [END] USER_THREAD_COUNT: mterry has 63/1028698 threads (0.01% utilization) -2025-07-28 16:02:06,777 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:02:07,454 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 68/1028698 threads (0.01% utilization) -2025-07-28 16:02:07,454 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 16:02:08,092 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 60/1028698 threads (0.01% utilization) -2025-07-28 16:02:08,092 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 16:02:08,152 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 16:08:09,325 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:08:10,013 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 53/1028698 threads (0.01% utilization) -2025-07-28 16:08:10,013 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 16:08:10,671 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 46/1028698 threads (0.0% utilization) -2025-07-28 16:08:10,672 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 16:08:10,733 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 16:14:11,846 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:14:12,499 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 46/1028698 threads (0.0% utilization) -2025-07-28 16:14:12,499 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 16:14:13,136 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 48/1028698 threads (0.0% utilization) -2025-07-28 16:14:13,136 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 16:14:13,192 - INFO - root : [END] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 16:20:14,303 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:20:14,957 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 16:20:14,957 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 16:20:15,591 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 16:20:15,591 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 16:20:15,647 - INFO - root : [END] USER_THREAD_COUNT: mterry has 45/1028698 threads (0.0% utilization) -2025-07-28 16:26:16,748 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:17,395 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 16:26:17,395 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 16:26:18,038 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 49/1028698 threads (0.0% utilization) -2025-07-28 16:26:18,038 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 16:26:18,094 - INFO - root : [END] USER_THREAD_COUNT: mterry has 49/1028698 threads (0.0% utilization) -2025-07-28 16:32:19,232 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:32:19,923 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 16:32:19,923 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 16:32:20,556 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 62/1028698 threads (0.01% utilization) -2025-07-28 16:32:20,556 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 16:32:20,612 - INFO - root : [END] USER_THREAD_COUNT: mterry has 67/1028698 threads (0.01% utilization) -2025-07-28 16:38:21,735 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:38:22,407 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 16:38:22,407 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 16:38:23,040 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:38:23,040 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 16:38:23,095 - INFO - root : [END] USER_THREAD_COUNT: mterry has 49/1028698 threads (0.0% utilization) -2025-07-28 16:44:24,195 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:24,854 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 16:44:24,854 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 16:44:25,488 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 16:44:25,488 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 16:44:25,543 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:50:26,790 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:50:27,437 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 16:50:27,438 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 16:50:28,073 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 36/1028698 threads (0.0% utilization) -2025-07-28 16:50:28,073 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 16:50:28,127 - INFO - root : [END] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 16:56:29,251 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:56:29,927 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 16:56:29,927 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 16:56:30,559 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 33/1028698 threads (0.0% utilization) -2025-07-28 16:56:30,559 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 16:56:30,613 - INFO - root : [END] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 17:02:31,738 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:02:32,398 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 17:02:32,398 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 17:02:33,030 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 17:02:33,030 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 17:02:33,084 - INFO - root : [END] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 17:08:34,227 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:08:34,905 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 17:08:34,905 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 17:08:35,533 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 17:08:35,533 - INFO - root : Rocoto call successful on attempt 1: call_time=0.57s, total_time=0.57s -2025-07-28 17:08:35,586 - INFO - root : [END] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 17:14:36,724 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:14:37,387 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 17:14:37,387 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 17:14:38,021 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 33/1028698 threads (0.0% utilization) -2025-07-28 17:14:38,021 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 17:14:38,074 - INFO - root : [END] USER_THREAD_COUNT: mterry has 33/1028698 threads (0.0% utilization) -2025-07-28 17:20:39,190 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:20:39,844 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 17:20:39,844 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 17:20:40,477 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 17:20:40,477 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 17:20:40,529 - INFO - root : [END] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 17:26:41,655 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:26:42,309 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 17:26:42,309 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 17:26:42,940 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 17:26:42,941 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 17:26:42,996 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:32:44,095 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:32:44,747 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:32:44,747 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 17:32:45,380 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 48/1028698 threads (0.0% utilization) -2025-07-28 17:32:45,380 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 17:32:45,434 - INFO - root : [END] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 17:38:46,590 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:38:47,242 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:38:47,242 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 17:38:47,872 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 17:38:47,872 - INFO - root : Rocoto call successful on attempt 1: call_time=0.57s, total_time=0.57s -2025-07-28 17:38:47,925 - INFO - root : [END] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 17:44:49,020 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:44:49,689 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:44:49,689 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 17:44:50,319 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 17:44:50,319 - INFO - root : Rocoto call successful on attempt 1: call_time=0.57s, total_time=0.57s -2025-07-28 17:44:50,372 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 17:50:51,499 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:50:52,177 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:50:52,177 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 17:50:52,805 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 17:50:52,805 - INFO - root : Rocoto call successful on attempt 1: call_time=0.57s, total_time=0.57s -2025-07-28 17:50:52,858 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 17:56:53,961 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 17:56:54,611 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 17:56:54,611 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 17:56:55,245 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 17:56:55,245 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 17:56:55,297 - INFO - root : [END] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 18:02:56,407 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:02:57,056 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:02:57,056 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 18:02:57,687 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 18:02:57,687 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 18:02:57,743 - INFO - root : [END] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 18:08:58,894 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:08:59,535 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 18:08:59,536 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 18:09:00,165 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 18:09:00,165 - INFO - root : Rocoto call successful on attempt 1: call_time=0.57s, total_time=0.57s -2025-07-28 18:09:00,219 - INFO - root : [END] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 18:15:01,319 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:01,962 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 18:15:01,962 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 18:15:02,640 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 18:15:02,640 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 18:15:02,702 - INFO - root : [END] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 18:21:03,867 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:04,544 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 18:21:04,544 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 18:21:05,172 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 18:21:05,172 - INFO - root : Rocoto call successful on attempt 1: call_time=0.57s, total_time=0.57s -2025-07-28 18:21:05,228 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:27:06,338 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:06,985 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 18:27:06,985 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 18:27:07,611 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 18:27:07,611 - INFO - root : Rocoto call successful on attempt 1: call_time=0.57s, total_time=0.57s -2025-07-28 18:27:07,664 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:33:08,786 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:09,429 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 18:33:09,429 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 18:33:10,061 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 45/1028698 threads (0.0% utilization) -2025-07-28 18:33:10,061 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 18:33:10,116 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:39:11,416 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 18:39:12,066 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 18:39:12,066 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 18:39:12,696 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 18:39:12,696 - INFO - root : Rocoto call successful on attempt 1: call_time=0.57s, total_time=0.57s -2025-07-28 18:39:12,748 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 18:45:13,866 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:45:14,528 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:45:14,528 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 18:45:15,169 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:45:15,169 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 18:45:15,224 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:51:16,360 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:51:17,010 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:51:17,010 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 18:51:17,645 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 18:51:17,645 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 18:51:17,696 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:57:18,799 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:57:19,450 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:57:19,450 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 18:57:20,079 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:20,079 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 18:57:20,130 - INFO - root : [END] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 19:03:21,510 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:03:22,157 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:03:22,157 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 19:03:22,794 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:03:22,794 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 19:03:22,846 - INFO - root : [END] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 19:09:24,011 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:09:24,656 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:09:24,656 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 19:09:25,284 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:09:25,284 - INFO - root : Rocoto call successful on attempt 1: call_time=0.57s, total_time=0.57s -2025-07-28 19:09:25,336 - INFO - root : [END] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 19:15:26,465 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:15:27,117 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:15:27,117 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 19:15:27,747 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:15:27,747 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 19:15:27,800 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 19:21:28,953 - INFO - root : [START] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 19:21:29,600 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 19:21:29,600 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 19:21:30,227 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 19:21:30,227 - INFO - root : Rocoto call successful on attempt 1: call_time=0.57s, total_time=0.57s -2025-07-28 19:21:30,279 - INFO - root : [END] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 19:27:31,434 - INFO - root : [START] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 19:27:32,078 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 19:27:32,078 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 19:27:32,716 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-28 19:27:32,716 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 19:27:32,768 - INFO - root : [END] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 19:33:33,907 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 19:33:34,554 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 19:33:34,554 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 19:33:35,188 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 19:33:35,188 - INFO - root : Rocoto call successful on attempt 1: call_time=0.57s, total_time=0.57s -2025-07-28 19:33:35,239 - INFO - root : [END] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 19:39:36,322 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 19:39:36,965 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:39:36,966 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 19:39:37,618 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 19:39:37,618 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 19:39:37,670 - INFO - root : [END] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 19:45:38,839 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 19:45:39,514 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 19:45:39,514 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 19:45:40,163 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 19:45:40,163 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 19:45:40,215 - INFO - root : [END] USER_THREAD_COUNT: mterry has 45/1028698 threads (0.0% utilization) -2025-07-28 19:51:41,386 - INFO - root : [START] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 19:51:42,021 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:51:42,021 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 19:51:42,677 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:51:42,677 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 19:51:42,728 - INFO - root : [END] USER_THREAD_COUNT: mterry has 45/1028698 threads (0.0% utilization) diff --git a/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C48_S2SWA_gefs_rocotostat.log b/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C48_S2SWA_gefs_rocotostat.log deleted file mode 100644 index 2dbd0fdd..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C48_S2SWA_gefs_rocotostat.log +++ /dev/null @@ -1,606 +0,0 @@ -2025-07-28 14:49:35,677 - INFO - root : [START] USER_THREAD_COUNT: mterry has 64/1028698 threads (0.01% utilization) -2025-07-28 14:49:36,551 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 80/1028698 threads (0.01% utilization) -2025-07-28 14:49:36,551 - INFO - root : Rocoto call successful on attempt 1: call_time=0.81s, total_time=0.81s -2025-07-28 14:49:37,352 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 14:49:37,353 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 14:49:37,407 - INFO - root : [END] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 14:55:38,753 - INFO - root : [START] USER_THREAD_COUNT: mterry has 95/1028698 threads (0.01% utilization) -2025-07-28 14:55:39,609 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 100/1028698 threads (0.01% utilization) -2025-07-28 14:55:39,610 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 14:55:40,406 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 69/1028698 threads (0.01% utilization) -2025-07-28 14:55:40,407 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 14:55:40,462 - INFO - root : [END] USER_THREAD_COUNT: mterry has 69/1028698 threads (0.01% utilization) -2025-07-28 15:01:41,783 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:01:42,633 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 15:01:42,633 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 15:01:43,428 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 15:01:43,428 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 15:01:43,482 - INFO - root : [END] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 15:07:44,774 - INFO - root : [START] USER_THREAD_COUNT: mterry has 33/1028698 threads (0.0% utilization) -2025-07-28 15:07:45,625 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 15:07:45,625 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 15:07:46,426 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 53/1028698 threads (0.01% utilization) -2025-07-28 15:07:46,426 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 15:07:46,483 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 15:13:47,765 - INFO - root : [START] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 15:13:48,608 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 15:13:48,608 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 15:13:49,397 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 15:13:49,397 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 15:13:49,453 - INFO - root : [END] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 15:19:50,713 - INFO - root : [START] USER_THREAD_COUNT: mterry has 53/1028698 threads (0.01% utilization) -2025-07-28 15:19:51,547 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 15:19:51,547 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 15:19:52,337 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 15:19:52,337 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 15:19:52,391 - INFO - root : [END] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 15:25:53,658 - INFO - root : [START] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 15:25:54,511 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 51/1028698 threads (0.0% utilization) -2025-07-28 15:25:54,511 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 15:25:55,312 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 15:25:55,312 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 15:25:55,370 - INFO - root : [END] USER_THREAD_COUNT: mterry has 51/1028698 threads (0.0% utilization) -2025-07-28 15:31:56,624 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:31:57,460 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 15:31:57,460 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 15:31:58,250 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 15:31:58,251 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 15:31:58,305 - INFO - root : [END] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 15:37:59,559 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:38:00,392 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 15:38:00,392 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 15:38:01,184 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 15:38:01,184 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 15:38:01,242 - INFO - root : [END] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 15:44:02,523 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:44:03,398 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 15:44:03,398 - INFO - root : Rocoto call successful on attempt 1: call_time=0.81s, total_time=0.81s -2025-07-28 15:44:04,186 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 15:44:04,186 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 15:44:04,241 - INFO - root : [END] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 15:50:05,510 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:50:06,348 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 15:50:06,348 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 15:50:07,137 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 15:50:07,137 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 15:50:07,194 - INFO - root : [END] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 15:56:08,480 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:56:09,314 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 33/1028698 threads (0.0% utilization) -2025-07-28 15:56:09,314 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 15:56:10,098 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 15:56:10,098 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 15:56:10,153 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:02:16,430 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:02:17,269 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 16:02:17,270 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 16:02:18,060 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 51/1028698 threads (0.0% utilization) -2025-07-28 16:02:18,060 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 16:02:18,116 - INFO - root : [END] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 16:08:19,446 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:08:20,288 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 16:08:20,288 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 16:08:21,080 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:08:21,080 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 16:08:21,136 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:14:22,390 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:14:23,227 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:14:23,227 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 16:14:24,020 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 16:14:24,021 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 16:14:24,076 - INFO - root : [END] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 16:20:25,362 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:20:26,193 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:20:26,194 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 16:20:26,984 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 16:20:26,984 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 16:20:27,039 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 16:26:28,309 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:29,138 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:29,138 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 16:26:29,926 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:29,926 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 16:26:29,984 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:32:31,239 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:32:32,071 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:32:32,071 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 16:32:32,858 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:32:32,859 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 16:32:32,912 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:38:34,478 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:38:35,306 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:38:35,307 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 16:38:36,091 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:38:36,091 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 16:38:36,143 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:37,410 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:38,252 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:38,252 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 16:44:39,035 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:39,035 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 16:44:39,087 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:40,360 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:41,193 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:41,193 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 16:50:41,977 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:41,978 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 16:50:42,030 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:43,297 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:44,124 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:44,124 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 16:56:44,914 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:44,914 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 16:56:44,966 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:46,224 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:47,056 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:47,056 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 17:02:47,845 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:47,845 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 17:02:47,897 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:49,176 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:50,004 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:50,004 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 17:08:50,786 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:50,786 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 17:08:50,838 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:52,095 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:52,924 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:52,924 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 17:14:53,711 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:53,711 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 17:14:53,762 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:55,028 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:55,860 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:55,860 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 17:20:56,649 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:56,649 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 17:20:56,701 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:57,979 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:58,810 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:58,810 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 17:26:59,596 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:59,596 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 17:26:59,648 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:33:00,923 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:33:01,761 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:33:01,761 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 17:33:02,555 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:33:02,555 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 17:33:02,608 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:39:03,924 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:39:04,755 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:39:04,755 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 17:39:05,542 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:39:05,542 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 17:39:05,594 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:45:06,885 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:45:07,720 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:45:07,720 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 17:45:08,509 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:45:08,509 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 17:45:08,564 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:51:10,045 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:51:10,878 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:51:10,878 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 17:51:11,683 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:51:11,683 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 17:51:11,735 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:57:13,002 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 17:57:13,840 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 17:57:13,840 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 17:57:14,625 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 17:57:14,625 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 17:57:14,677 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:03:15,948 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:03:16,777 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:03:16,777 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 18:03:17,571 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:03:17,571 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 18:03:17,624 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:18,922 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:19,780 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:19,780 - INFO - root : Rocoto call successful on attempt 1: call_time=0.80s, total_time=0.80s -2025-07-28 18:09:20,571 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:20,571 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 18:09:20,623 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:21,890 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:22,722 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:22,722 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 18:15:23,508 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:23,508 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 18:15:23,561 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:24,858 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:25,687 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:25,687 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 18:21:26,470 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:26,470 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 18:21:26,521 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:27,772 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:28,602 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:28,603 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 18:27:29,389 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:29,389 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 18:27:29,441 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:30,720 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:31,577 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:31,577 - INFO - root : Rocoto call successful on attempt 1: call_time=0.80s, total_time=0.80s -2025-07-28 18:33:32,359 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:32,359 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 18:33:32,411 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:39:33,674 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:39:34,503 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:39:34,503 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 18:39:35,284 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:39:35,284 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 18:39:35,335 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:45:36,605 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:45:37,432 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:45:37,433 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 18:45:38,217 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:45:38,217 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 18:45:38,269 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:39,661 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:40,491 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:40,491 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 18:51:41,272 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:41,273 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 18:51:41,324 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:42,651 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:43,482 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:43,483 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 18:57:44,265 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:44,265 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 18:57:44,316 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:03:46,252 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:03:47,080 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:03:47,080 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 19:03:47,861 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:03:47,861 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 19:03:47,913 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:49,511 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:50,355 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:50,355 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 19:09:51,138 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:51,139 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 19:09:51,191 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:15:52,535 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:15:53,367 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:15:53,367 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 19:15:54,149 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:15:54,150 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 19:15:54,201 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:21:55,574 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:21:56,405 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:21:56,405 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 19:21:57,191 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:21:57,191 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 19:21:57,244 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:58,570 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:59,402 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:59,402 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 19:28:00,189 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:28:00,189 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 19:28:00,241 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:34:01,570 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:34:02,413 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:34:02,413 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 19:34:03,198 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:34:03,198 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 19:34:03,249 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:40:04,602 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:40:05,445 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:40:05,445 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 19:40:06,250 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:40:06,250 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 19:40:06,308 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:46:07,677 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:46:08,507 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:46:08,508 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 19:46:09,302 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:46:09,303 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 19:46:09,355 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:52:10,699 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:52:11,531 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:52:11,531 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 19:52:12,321 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:52:12,321 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 19:52:12,374 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:58:13,710 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:58:14,541 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:58:14,541 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 19:58:15,327 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:58:15,327 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 19:58:15,379 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:04:16,747 - INFO - root : [START] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 20:04:17,577 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 20:04:17,578 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 20:04:18,364 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 20:04:18,364 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 20:04:18,416 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 20:10:19,760 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:10:20,592 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:10:20,592 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 20:10:21,395 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:10:21,395 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 20:10:21,446 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:16:22,801 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 20:16:23,642 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 20:16:23,642 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 20:16:24,431 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:16:24,431 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 20:16:24,482 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:22:25,791 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 20:22:26,625 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 20:22:26,625 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 20:22:27,420 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:22:27,420 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 20:22:27,473 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:28:28,876 - INFO - root : [START] USER_THREAD_COUNT: mterry has 75/1028698 threads (0.01% utilization) -2025-07-28 20:28:29,708 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 80/1028698 threads (0.01% utilization) -2025-07-28 20:28:29,708 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 20:28:30,495 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:28:30,496 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 20:28:30,547 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:34:31,925 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:34:32,755 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 20:34:32,755 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 20:34:33,551 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:34:33,552 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 20:34:33,604 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:35,104 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 20:40:35,956 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:35,956 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 20:40:36,751 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:36,751 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 20:40:36,803 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:46:38,173 - INFO - root : [START] USER_THREAD_COUNT: mterry has 36/1028698 threads (0.0% utilization) -2025-07-28 20:46:39,006 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:39,006 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 20:46:39,793 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:39,793 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 20:46:39,845 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:41,204 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:42,033 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:42,033 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 20:52:42,818 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:42,818 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 20:52:42,870 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:44,207 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:45,040 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:45,040 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 20:58:45,823 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:45,823 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 20:58:45,874 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:47,342 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:48,180 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:48,181 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 21:04:48,968 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:48,968 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 21:04:49,020 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:50,407 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:51,236 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:51,236 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 21:10:52,020 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:52,020 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 21:10:52,072 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:16:53,404 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:54,277 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:54,277 - INFO - root : Rocoto call successful on attempt 1: call_time=0.82s, total_time=0.82s -2025-07-28 21:16:55,063 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:55,063 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 21:16:55,115 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:22:56,654 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:57,485 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:57,486 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 21:22:58,270 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:58,270 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 21:22:58,321 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:29:00,038 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:29:00,869 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:29:00,869 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 21:29:01,658 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:29:01,658 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 21:29:01,714 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:35:03,437 - INFO - root : [START] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:35:04,280 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:35:04,280 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 21:35:05,065 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:35:05,065 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 21:35:05,118 - INFO - root : [END] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:41:06,597 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:41:07,428 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:41:07,429 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 21:41:08,229 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:41:08,229 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 21:41:08,281 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:47:09,786 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:47:10,619 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:47:10,619 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 21:47:11,416 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:47:11,416 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 21:47:11,467 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:53:12,924 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:53:13,757 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:53:13,758 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 21:53:14,542 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:53:14,542 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 21:53:14,594 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:59:16,000 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:59:16,832 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:59:16,832 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 21:59:17,616 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:59:17,617 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 21:59:17,669 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:05:19,290 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:05:20,124 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:05:20,124 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 22:05:20,907 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:05:20,907 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 22:05:20,959 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:11:22,381 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:11:23,221 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:11:23,221 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 22:11:24,004 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:11:24,004 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 22:11:24,056 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:17:25,783 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:17:26,612 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:17:26,612 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 22:17:27,395 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:17:27,395 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 22:17:27,447 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:23:29,131 - INFO - root : [START] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 22:23:29,960 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 22:23:29,960 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 22:23:30,739 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 22:23:30,739 - INFO - root : Rocoto call successful on attempt 1: call_time=0.72s, total_time=0.72s -2025-07-28 22:23:30,791 - INFO - root : [END] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 22:29:32,182 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:29:33,013 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:29:33,014 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 22:29:33,796 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:29:33,796 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 22:29:33,847 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:35:35,322 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:35:36,153 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:35:36,153 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 22:35:36,938 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:35:36,938 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 22:35:36,989 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:41:38,388 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:41:39,218 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:41:39,218 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 22:41:40,002 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:41:40,002 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 22:41:40,053 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:47:41,442 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:47:42,272 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:47:42,272 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 22:47:43,058 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:47:43,058 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 22:47:43,109 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:53:45,842 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:53:46,673 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:53:46,674 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 22:53:47,456 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:53:47,456 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 22:53:47,508 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:59:49,080 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:59:49,914 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:59:49,914 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 22:59:50,695 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:59:50,695 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 22:59:50,746 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:05:52,266 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:05:53,099 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:05:53,099 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 23:05:53,882 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:05:53,882 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 23:05:53,933 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:55,300 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:56,135 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:56,135 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 23:11:56,931 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:56,931 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 23:11:56,982 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:58,289 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:59,119 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:59,119 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 23:17:59,902 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:59,902 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 23:17:59,953 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:24:01,312 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:24:02,172 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:24:02,172 - INFO - root : Rocoto call successful on attempt 1: call_time=0.80s, total_time=0.80s -2025-07-28 23:24:02,975 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:24:02,975 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 23:24:03,032 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:30:04,393 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:30:05,228 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:30:05,228 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 23:30:06,033 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:30:06,033 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 23:30:06,091 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:36:07,424 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:36:08,267 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:36:08,267 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 23:36:09,057 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:36:09,057 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 23:36:09,118 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:42:10,432 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:42:11,269 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:42:11,269 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 23:42:12,056 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:42:12,056 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 23:42:12,111 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:48:13,431 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:48:14,272 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:48:14,272 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 23:48:15,063 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:48:15,063 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 23:48:15,114 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:54:16,428 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:54:17,262 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:54:17,262 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 23:54:18,051 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:54:18,051 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-28 23:54:18,102 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:00:19,504 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:00:20,341 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:00:20,341 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-29 00:00:21,131 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:00:21,131 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-29 00:00:21,182 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:06:22,668 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:06:23,506 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:06:23,506 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-29 00:06:24,291 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:06:24,291 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-29 00:06:24,343 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:12:26,120 - INFO - root : [START] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:12:26,955 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:12:26,956 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-29 00:12:27,743 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:12:27,744 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-29 00:12:27,796 - INFO - root : [END] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:18:29,230 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:18:30,066 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:18:30,066 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-29 00:18:30,853 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:18:30,853 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-29 00:18:30,905 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:24:32,232 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:24:33,076 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:24:33,076 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-29 00:24:33,859 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:24:33,859 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-29 00:24:33,910 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:30:35,487 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:30:36,325 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:30:36,325 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-29 00:30:37,112 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:30:37,112 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-29 00:30:37,163 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:36:38,542 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:36:39,374 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:36:39,374 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-29 00:36:40,161 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:36:40,161 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-29 00:36:40,212 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:42:41,520 - INFO - root : [START] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:42:42,356 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:42:42,356 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-29 00:42:43,142 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:42:43,143 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-29 00:42:43,195 - INFO - root : [END] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:48:44,743 - INFO - root : [START] USER_THREAD_COUNT: mterry has 74/1028698 threads (0.01% utilization) -2025-07-29 00:48:45,578 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 74/1028698 threads (0.01% utilization) -2025-07-29 00:48:45,578 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-29 00:48:46,366 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 74/1028698 threads (0.01% utilization) -2025-07-29 00:48:46,366 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-29 00:48:46,418 - INFO - root : [END] USER_THREAD_COUNT: mterry has 74/1028698 threads (0.01% utilization) -2025-07-29 00:54:53,060 - INFO - root : [START] USER_THREAD_COUNT: mterry has 48/1028698 threads (0.0% utilization) -2025-07-29 00:54:53,903 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 48/1028698 threads (0.0% utilization) -2025-07-29 00:54:53,903 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-29 00:54:54,689 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 48/1028698 threads (0.0% utilization) -2025-07-29 00:54:54,689 - INFO - root : Rocoto call successful on attempt 1: call_time=0.73s, total_time=0.73s -2025-07-29 00:54:54,741 - INFO - root : [END] USER_THREAD_COUNT: mterry has 48/1028698 threads (0.0% utilization) diff --git a/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C48_S2SW_rocotostat.log b/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C48_S2SW_rocotostat.log deleted file mode 100644 index 28289f7e..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C48_S2SW_rocotostat.log +++ /dev/null @@ -1,456 +0,0 @@ -2025-07-28 14:49:35,427 - INFO - root : [START] USER_THREAD_COUNT: mterry has 33/1028698 threads (0.0% utilization) -2025-07-28 14:49:36,200 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 14:49:36,200 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 14:49:36,941 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 34/1028698 threads (0.0% utilization) -2025-07-28 14:49:36,941 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 14:49:36,999 - INFO - root : [END] USER_THREAD_COUNT: mterry has 49/1028698 threads (0.0% utilization) -2025-07-28 14:55:38,243 - INFO - root : [START] USER_THREAD_COUNT: mterry has 102/1028698 threads (0.01% utilization) -2025-07-28 14:55:39,016 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 95/1028698 threads (0.01% utilization) -2025-07-28 14:55:39,017 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 14:55:39,810 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 79/1028698 threads (0.01% utilization) -2025-07-28 14:55:39,810 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 14:55:39,867 - INFO - root : [END] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 15:01:41,108 - INFO - root : [START] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 15:01:41,931 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 33/1028698 threads (0.0% utilization) -2025-07-28 15:01:41,931 - INFO - root : Rocoto call successful on attempt 1: call_time=0.76s, total_time=0.76s -2025-07-28 15:01:42,670 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 15:01:42,671 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 15:01:42,726 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 15:07:43,935 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 15:07:44,712 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 15:07:44,712 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 15:07:45,470 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 58/1028698 threads (0.01% utilization) -2025-07-28 15:07:45,470 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 15:07:45,529 - INFO - root : [END] USER_THREAD_COUNT: mterry has 58/1028698 threads (0.01% utilization) -2025-07-28 15:13:46,748 - INFO - root : [START] USER_THREAD_COUNT: mterry has 35/1028698 threads (0.0% utilization) -2025-07-28 15:13:47,516 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 15:13:47,516 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 15:13:48,255 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 15:13:48,255 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 15:13:48,311 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 15:19:49,495 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 15:19:50,260 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 15:19:50,261 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 15:19:50,996 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:19:50,997 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 15:19:51,052 - INFO - root : [END] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:25:52,265 - INFO - root : [START] USER_THREAD_COUNT: mterry has 49/1028698 threads (0.0% utilization) -2025-07-28 15:25:53,048 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 15:25:53,049 - INFO - root : Rocoto call successful on attempt 1: call_time=0.72s, total_time=0.72s -2025-07-28 15:25:53,803 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 15:25:53,803 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 15:25:53,861 - INFO - root : [END] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 15:31:55,047 - INFO - root : [START] USER_THREAD_COUNT: mterry has 95/1028698 threads (0.01% utilization) -2025-07-28 15:31:55,809 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 15:31:55,809 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 15:31:56,541 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:31:56,541 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 15:31:56,597 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:37:57,781 - INFO - root : [START] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 15:37:58,707 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 15:37:58,707 - INFO - root : Rocoto call successful on attempt 1: call_time=0.87s, total_time=0.87s -2025-07-28 15:37:59,435 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:37:59,435 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 15:37:59,489 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:44:00,670 - INFO - root : [START] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 15:44:01,424 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 15:44:01,424 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 15:44:02,179 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:44:02,179 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 15:44:02,237 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:50:03,439 - INFO - root : [START] USER_THREAD_COUNT: mterry has 64/1028698 threads (0.01% utilization) -2025-07-28 15:50:04,200 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 63/1028698 threads (0.01% utilization) -2025-07-28 15:50:04,200 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 15:50:04,931 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:50:04,931 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 15:50:04,986 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:56:06,196 - INFO - root : [START] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 15:56:06,956 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 15:56:06,956 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 15:56:07,710 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:56:07,710 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 15:56:07,766 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:02:08,954 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 16:02:09,714 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 16:02:09,714 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 16:02:10,451 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:02:10,451 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 16:02:10,507 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:08:11,699 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:08:12,455 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:08:12,455 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 16:08:13,188 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:08:13,188 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 16:08:13,244 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:14:14,999 - INFO - root : [START] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 16:14:15,765 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:14:15,766 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 16:14:16,511 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:14:16,512 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 16:14:16,567 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:20:17,762 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:20:18,516 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:20:18,516 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 16:20:19,247 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:20:19,247 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 16:20:19,300 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:20,491 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:21,241 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:21,241 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 16:26:21,969 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:21,969 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 16:26:22,021 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:32:23,308 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:32:24,064 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:32:24,064 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 16:32:24,791 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:32:24,791 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 16:32:24,844 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:38:26,052 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:38:26,815 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:38:26,816 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 16:38:27,540 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:38:27,540 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 16:38:27,592 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:28,786 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:29,539 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:29,539 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 16:44:30,268 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:30,268 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 16:44:30,320 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:31,510 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:32,266 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:32,266 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 16:50:32,999 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:32,999 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 16:50:33,051 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:34,255 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:35,006 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:35,006 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 16:56:35,736 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:35,736 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 16:56:35,788 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:36,981 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:37,735 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:37,735 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 17:02:38,466 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:38,467 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 17:02:38,519 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:39,721 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:40,473 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:40,473 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 17:08:41,200 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:41,200 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 17:08:41,253 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:42,501 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:43,254 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:43,254 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 17:14:43,985 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:43,985 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 17:14:44,037 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:45,231 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:45,983 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:45,983 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 17:20:46,712 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:46,712 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 17:20:46,765 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:47,954 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:48,710 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:48,710 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 17:26:49,439 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:49,440 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 17:26:49,491 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:32:50,670 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:32:51,446 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:32:51,447 - INFO - root : Rocoto call successful on attempt 1: call_time=0.72s, total_time=0.72s -2025-07-28 17:32:52,177 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:32:52,178 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 17:32:52,231 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:38:53,457 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:38:54,214 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:38:54,214 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 17:38:54,942 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:38:54,942 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 17:38:54,994 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:44:56,169 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:44:56,918 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:44:56,918 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 17:44:57,644 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:44:57,644 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 17:44:57,695 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:50:58,881 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:50:59,633 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:50:59,633 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 17:51:00,367 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:51:00,367 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 17:51:00,419 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:57:01,599 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 17:57:02,360 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 17:57:02,360 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 17:57:03,122 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 17:57:03,122 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 17:57:03,176 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:03:04,359 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:03:05,111 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:03:05,111 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 18:03:05,834 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:03:05,834 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 18:03:05,886 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:07,083 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:07,838 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:07,838 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 18:09:08,564 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:08,564 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 18:09:08,616 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:09,809 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:10,566 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:10,566 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 18:15:11,298 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:11,298 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 18:15:11,351 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:12,530 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:13,280 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:13,280 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 18:21:14,014 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:14,014 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 18:21:14,065 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:15,240 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:15,996 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:15,996 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 18:27:16,723 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:16,723 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 18:27:16,774 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:17,955 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:18,712 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:18,713 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 18:33:19,441 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:19,441 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 18:33:19,493 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:39:21,023 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:39:21,781 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:39:21,781 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 18:39:22,508 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:39:22,508 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 18:39:22,561 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:45:23,743 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:45:24,493 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:45:24,493 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 18:45:25,222 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:45:25,222 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 18:45:25,274 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:26,455 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:27,208 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:27,208 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 18:51:27,934 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:27,934 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 18:51:27,985 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:29,253 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:35,015 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:35,015 - INFO - root : Rocoto call successful on attempt 1: call_time=5.70s, total_time=5.70s -2025-07-28 18:57:35,745 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:35,745 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 18:57:35,796 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:03:37,353 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:03:38,127 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:03:38,127 - INFO - root : Rocoto call successful on attempt 1: call_time=0.72s, total_time=0.72s -2025-07-28 19:03:38,854 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:03:38,854 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 19:03:38,906 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:40,529 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:41,289 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:41,289 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 19:09:42,022 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:42,022 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 19:09:42,074 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:15:43,493 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:15:44,247 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:15:44,247 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 19:15:44,977 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:15:44,977 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 19:15:45,030 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:21:46,404 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:21:47,155 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:21:47,155 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 19:21:47,883 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:21:47,883 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 19:21:47,936 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:54,241 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:54,998 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 19:27:54,998 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 19:27:55,728 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 69/1028698 threads (0.01% utilization) -2025-07-28 19:27:55,728 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 19:27:55,781 - INFO - root : [END] USER_THREAD_COUNT: mterry has 69/1028698 threads (0.01% utilization) -2025-07-28 19:33:57,052 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:33:57,805 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:33:57,805 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 19:33:58,535 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 19:33:58,535 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 19:33:58,588 - INFO - root : [END] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 19:39:59,865 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:40:00,626 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:40:00,626 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 19:40:01,358 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 51/1028698 threads (0.0% utilization) -2025-07-28 19:40:01,358 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 19:40:01,413 - INFO - root : [END] USER_THREAD_COUNT: mterry has 51/1028698 threads (0.0% utilization) -2025-07-28 19:46:03,023 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:46:03,779 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:46:03,780 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 19:46:04,517 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 51/1028698 threads (0.0% utilization) -2025-07-28 19:46:04,517 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 19:46:04,571 - INFO - root : [END] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 19:52:06,297 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:52:07,064 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 19:52:07,064 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 19:52:07,838 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:52:07,838 - INFO - root : Rocoto call successful on attempt 1: call_time=0.72s, total_time=0.72s -2025-07-28 19:52:07,891 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:58:09,176 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:58:09,971 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:58:09,971 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 19:58:10,748 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:58:10,748 - INFO - root : Rocoto call successful on attempt 1: call_time=0.72s, total_time=0.72s -2025-07-28 19:58:10,800 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 20:04:12,098 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:04:12,854 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 25/1028698 threads (0.0% utilization) -2025-07-28 20:04:12,854 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 20:04:13,584 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 36/1028698 threads (0.0% utilization) -2025-07-28 20:04:13,584 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 20:04:13,637 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 20:10:14,932 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:10:15,690 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:10:15,690 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 20:10:16,428 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 20:10:16,428 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 20:10:16,482 - INFO - root : [END] USER_THREAD_COUNT: mterry has 36/1028698 threads (0.0% utilization) -2025-07-28 20:16:22,754 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:16:23,505 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 20:16:23,505 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 20:16:24,234 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 20:16:24,234 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 20:16:24,288 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 20:22:25,571 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:22:26,327 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 36/1028698 threads (0.0% utilization) -2025-07-28 20:22:26,327 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 20:22:27,059 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 36/1028698 threads (0.0% utilization) -2025-07-28 20:22:27,060 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 20:22:27,113 - INFO - root : [END] USER_THREAD_COUNT: mterry has 36/1028698 threads (0.0% utilization) -2025-07-28 20:28:28,376 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:28:29,131 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 66/1028698 threads (0.01% utilization) -2025-07-28 20:28:29,131 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 20:28:29,861 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 66/1028698 threads (0.01% utilization) -2025-07-28 20:28:29,861 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 20:28:29,914 - INFO - root : [END] USER_THREAD_COUNT: mterry has 66/1028698 threads (0.01% utilization) -2025-07-28 20:34:31,203 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:34:31,957 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 20:34:31,958 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 20:34:32,685 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 20:34:32,685 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 20:34:32,741 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 20:40:34,003 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:34,761 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:34,761 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 20:40:35,506 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 20:40:35,506 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 20:40:35,562 - INFO - root : [END] USER_THREAD_COUNT: mterry has 49/1028698 threads (0.0% utilization) -2025-07-28 20:46:36,911 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:37,680 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:37,680 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 20:46:38,411 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 20:46:38,412 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 20:46:38,465 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 20:52:39,757 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:40,509 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:40,509 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 20:52:41,236 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 20:52:41,236 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 20:52:41,289 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 20:58:42,570 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:43,322 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:43,322 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 20:58:44,049 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:44,049 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 20:58:44,101 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:45,470 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:46,230 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:46,230 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 21:04:46,964 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:46,964 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 21:04:47,018 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:48,452 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:49,204 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:49,204 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 21:10:49,930 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:49,930 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 21:10:49,982 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:16:51,262 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:52,015 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:52,015 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 21:16:52,741 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:52,741 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 21:16:52,794 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:22:54,274 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:55,032 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:55,032 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 21:22:55,760 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:55,760 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 21:22:55,813 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:28:57,279 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:28:58,033 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:28:58,033 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 21:28:58,759 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:28:58,759 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 21:28:58,811 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:35:00,552 - INFO - root : [START] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:35:01,306 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:35:01,306 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 21:35:02,052 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:35:02,052 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 21:35:02,117 - INFO - root : [END] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:41:03,834 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:41:04,601 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:41:04,601 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 21:41:05,365 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:41:05,365 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 21:41:05,417 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:47:06,715 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:47:07,465 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:47:07,465 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 21:47:08,189 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:47:08,189 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 21:47:08,241 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:53:09,600 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:53:10,359 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:53:10,359 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 21:53:11,087 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:53:11,087 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 21:53:11,139 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:59:12,345 - INFO - root : [START] USER_THREAD_COUNT: mterry has 53/1028698 threads (0.01% utilization) -2025-07-28 21:59:13,106 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:59:13,106 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 21:59:13,837 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:59:13,837 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 21:59:13,889 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:05:15,075 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 22:05:15,838 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:05:15,839 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 22:05:16,568 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:05:16,568 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 22:05:16,624 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:11:17,877 - INFO - root : [START] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 22:11:18,631 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:11:18,631 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 22:11:19,365 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:11:19,365 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 22:11:19,417 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:17:20,609 - INFO - root : [START] USER_THREAD_COUNT: mterry has 66/1028698 threads (0.01% utilization) -2025-07-28 22:17:21,363 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:17:21,363 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 22:17:22,088 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:17:22,088 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 22:17:22,140 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:23:23,494 - INFO - root : [START] USER_THREAD_COUNT: mterry has 53/1028698 threads (0.01% utilization) -2025-07-28 22:23:24,247 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:23:24,247 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 22:23:24,988 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:23:24,989 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 22:23:25,040 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) diff --git a/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C48mx500_3DVarAOWCDA_rocotostat.log b/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C48mx500_3DVarAOWCDA_rocotostat.log deleted file mode 100644 index e4bb5942..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C48mx500_3DVarAOWCDA_rocotostat.log +++ /dev/null @@ -1,648 +0,0 @@ -2025-07-28 14:49:35,677 - INFO - root : [START] USER_THREAD_COUNT: mterry has 64/1028698 threads (0.01% utilization) -2025-07-28 14:49:36,552 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 79/1028698 threads (0.01% utilization) -2025-07-28 14:49:36,552 - INFO - root : Rocoto call successful on attempt 1: call_time=0.81s, total_time=0.81s -2025-07-28 14:49:37,371 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 14:49:37,371 - INFO - root : Rocoto call successful on attempt 1: call_time=0.76s, total_time=0.76s -2025-07-28 14:49:37,424 - INFO - root : [END] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 14:55:38,754 - INFO - root : [START] USER_THREAD_COUNT: mterry has 95/1028698 threads (0.01% utilization) -2025-07-28 14:55:39,615 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 100/1028698 threads (0.01% utilization) -2025-07-28 14:55:39,615 - INFO - root : Rocoto call successful on attempt 1: call_time=0.80s, total_time=0.80s -2025-07-28 14:55:40,424 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 69/1028698 threads (0.01% utilization) -2025-07-28 14:55:40,424 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 14:55:40,478 - INFO - root : [END] USER_THREAD_COUNT: mterry has 69/1028698 threads (0.01% utilization) -2025-07-28 15:01:41,783 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:01:42,642 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 15:01:42,642 - INFO - root : Rocoto call successful on attempt 1: call_time=0.80s, total_time=0.80s -2025-07-28 15:01:43,444 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 15:01:43,444 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 15:01:43,498 - INFO - root : [END] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 15:07:44,845 - INFO - root : [START] USER_THREAD_COUNT: mterry has 34/1028698 threads (0.0% utilization) -2025-07-28 15:07:45,701 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:07:45,701 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 15:07:46,511 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 15:07:46,511 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 15:07:46,567 - INFO - root : [END] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 15:13:47,973 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:13:48,815 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 15:13:48,815 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 15:13:49,621 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:13:49,621 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 15:13:49,681 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 15:19:51,016 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:19:51,864 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 15:19:51,865 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 15:19:52,685 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:19:52,685 - INFO - root : Rocoto call successful on attempt 1: call_time=0.76s, total_time=0.76s -2025-07-28 15:19:52,740 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:25:54,052 - INFO - root : [START] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 15:25:54,907 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 15:25:54,907 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 15:25:55,725 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 15:25:55,725 - INFO - root : Rocoto call successful on attempt 1: call_time=0.76s, total_time=0.76s -2025-07-28 15:25:55,780 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 15:31:57,080 - INFO - root : [START] USER_THREAD_COUNT: mterry has 64/1028698 threads (0.01% utilization) -2025-07-28 15:31:57,923 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 64/1028698 threads (0.01% utilization) -2025-07-28 15:31:57,923 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 15:31:58,729 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:31:58,729 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 15:31:58,782 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:38:00,076 - INFO - root : [START] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 15:38:00,915 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 15:38:00,916 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 15:38:01,727 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:38:01,727 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 15:38:01,783 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:44:03,110 - INFO - root : [START] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 15:44:03,958 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 15:44:03,958 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 15:44:04,756 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:44:04,756 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 15:44:04,810 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:50:06,094 - INFO - root : [START] USER_THREAD_COUNT: mterry has 69/1028698 threads (0.01% utilization) -2025-07-28 15:50:06,939 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 69/1028698 threads (0.01% utilization) -2025-07-28 15:50:06,939 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 15:50:07,748 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:50:07,748 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 15:50:07,807 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:56:09,112 - INFO - root : [START] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 15:56:09,954 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 15:56:09,955 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 15:56:10,755 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:56:10,755 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 15:56:10,808 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:02:12,096 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:02:12,940 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:02:12,940 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 16:02:13,747 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:02:13,747 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 16:02:13,803 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:08:15,168 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:08:16,024 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:08:16,024 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 16:08:16,875 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:08:16,875 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 16:08:16,933 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:14:18,229 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:14:19,072 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:14:19,072 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 16:14:19,886 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:14:19,887 - INFO - root : Rocoto call successful on attempt 1: call_time=0.76s, total_time=0.76s -2025-07-28 16:14:19,942 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:20:21,223 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:20:22,062 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:20:22,062 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 16:20:22,867 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:20:22,868 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 16:20:22,921 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:24,204 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:25,039 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:25,039 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 16:26:25,846 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:25,846 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 16:26:25,899 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:32:27,178 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:32:28,018 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:32:28,018 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 16:32:28,830 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:32:28,830 - INFO - root : Rocoto call successful on attempt 1: call_time=0.76s, total_time=0.76s -2025-07-28 16:32:28,883 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:38:30,184 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:38:31,021 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:38:31,022 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 16:38:31,825 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:38:31,826 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 16:38:31,878 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:33,154 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:33,994 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:33,995 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 16:44:34,794 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:34,795 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 16:44:34,847 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:36,131 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:36,972 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:36,972 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 16:50:37,779 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:37,779 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 16:50:37,832 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:39,110 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:39,944 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:39,944 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 16:56:40,746 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:40,746 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 16:56:40,798 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:42,078 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:42,916 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:42,916 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 17:02:43,720 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:43,720 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 17:02:43,772 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:45,049 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:45,893 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:45,894 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 17:08:46,697 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:46,697 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 17:08:46,750 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:48,071 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:48,906 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:48,907 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 17:14:49,710 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:49,710 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 17:14:49,762 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:51,064 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:51,905 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:51,905 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 17:20:52,706 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:52,706 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 17:20:52,759 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:54,043 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:54,882 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:54,882 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 17:26:55,687 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:55,687 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 17:26:55,740 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:32:57,013 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:32:57,851 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:32:57,851 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 17:32:58,656 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:32:58,656 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 17:32:58,708 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:39:00,025 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:39:00,861 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:39:00,861 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 17:39:01,660 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:39:01,660 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 17:39:01,712 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:45:03,118 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:45:03,975 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:45:03,975 - INFO - root : Rocoto call successful on attempt 1: call_time=0.80s, total_time=0.80s -2025-07-28 17:45:04,779 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:45:04,779 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 17:45:04,832 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:51:06,124 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:51:06,983 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:51:06,983 - INFO - root : Rocoto call successful on attempt 1: call_time=0.80s, total_time=0.80s -2025-07-28 17:51:07,790 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:51:07,790 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 17:51:07,842 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:57:09,134 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 17:57:09,976 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 17:57:09,977 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 17:57:10,780 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 17:57:10,780 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 17:57:10,833 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:03:12,119 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:03:12,957 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:03:12,957 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 18:03:13,762 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:03:13,762 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 18:03:13,814 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:15,105 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:15,952 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:15,952 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 18:09:16,760 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:16,760 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 18:09:16,811 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:18,090 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:18,931 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:18,931 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 18:15:19,730 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:19,730 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 18:15:19,782 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:21,058 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:21,894 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:21,895 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 18:21:22,697 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:22,697 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 18:21:22,750 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:24,009 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:24,844 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:24,844 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 18:27:25,643 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:25,643 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 18:27:25,695 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:26,969 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:27,807 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:27,807 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 18:33:28,609 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:28,609 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 18:33:28,661 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:39:30,011 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:39:30,852 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:39:30,852 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 18:39:31,652 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:39:31,652 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 18:39:31,704 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:45:33,007 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:45:33,843 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:45:33,843 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 18:45:34,647 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:45:34,647 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 18:45:34,698 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:35,960 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:36,793 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:36,793 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 18:51:37,591 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:37,591 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 18:51:37,643 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:38,952 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:39,790 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:39,791 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 18:57:40,596 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:40,596 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 18:57:40,647 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:03:41,911 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:03:42,749 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:03:42,750 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 19:03:43,549 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:03:43,549 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 19:03:43,600 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:46,115 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:46,950 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:46,950 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 19:09:47,746 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:47,746 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 19:09:47,799 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:15:49,077 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:15:49,910 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:15:49,911 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 19:15:50,710 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:15:50,710 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 19:15:50,762 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:21:52,038 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:21:52,875 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:21:52,875 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 19:21:53,675 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:21:53,675 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 19:21:53,728 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:55,003 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:55,838 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 53/1028698 threads (0.01% utilization) -2025-07-28 19:27:55,838 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 19:27:56,637 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:56,637 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 19:27:56,689 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:33:57,942 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 19:33:58,782 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:33:58,782 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 19:33:59,583 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:33:59,583 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 19:33:59,635 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:40:00,899 - INFO - root : [START] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 19:40:01,768 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:40:01,768 - INFO - root : Rocoto call successful on attempt 1: call_time=0.80s, total_time=0.80s -2025-07-28 19:40:02,660 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:40:02,660 - INFO - root : Rocoto call successful on attempt 1: call_time=0.83s, total_time=0.83s -2025-07-28 19:40:02,726 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:46:04,014 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 19:46:04,855 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:46:04,855 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 19:46:05,656 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:46:05,656 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 19:46:05,708 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:52:06,985 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:52:07,872 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:52:07,872 - INFO - root : Rocoto call successful on attempt 1: call_time=0.83s, total_time=0.83s -2025-07-28 19:52:08,682 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:52:08,682 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 19:52:08,736 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:58:10,040 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 19:58:10,921 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:58:10,921 - INFO - root : Rocoto call successful on attempt 1: call_time=0.82s, total_time=0.82s -2025-07-28 19:58:11,728 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:58:11,728 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 19:58:11,779 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:04:13,052 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 20:04:13,887 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:04:13,888 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 20:04:14,697 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 20:04:14,697 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 20:04:14,749 - INFO - root : [END] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 20:10:16,031 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 20:10:16,870 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:10:16,870 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 20:10:17,675 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:10:17,676 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 20:10:17,728 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:16:19,008 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:16:19,854 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:16:19,854 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 20:16:20,655 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:16:20,655 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 20:16:20,707 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:22:21,988 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:22:22,827 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:22:22,827 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 20:22:23,635 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:22:23,635 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 20:22:23,688 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:28:24,991 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:28:25,830 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:28:25,830 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 20:28:26,633 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:28:26,634 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 20:28:26,686 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:34:27,989 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:34:28,833 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:34:28,833 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 20:34:29,634 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:34:29,634 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 20:34:29,686 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:30,983 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:31,827 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:31,827 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 20:40:32,636 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:32,636 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 20:40:32,688 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:46:34,039 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:34,886 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:34,887 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 20:46:35,691 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:35,691 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 20:46:35,743 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:37,029 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:37,870 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:37,870 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 20:52:38,670 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:38,670 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 20:52:38,722 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:40,004 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:40,844 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:40,844 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 20:58:41,643 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:41,643 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 20:58:41,695 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 21:04:42,990 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:43,823 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:43,823 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 21:04:44,626 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:44,626 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 21:04:44,678 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:45,986 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:46,848 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:46,848 - INFO - root : Rocoto call successful on attempt 1: call_time=0.80s, total_time=0.80s -2025-07-28 21:10:47,650 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:47,650 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 21:10:47,702 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:16:48,972 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:49,812 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:49,812 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 21:16:50,614 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:50,614 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 21:16:50,666 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:22:52,020 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:52,862 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:52,862 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 21:22:53,660 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:53,660 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 21:22:53,712 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:28:55,056 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:28:55,903 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:28:55,903 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 21:28:56,703 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:28:56,703 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 21:28:56,755 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:34:58,071 - INFO - root : [START] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:34:58,914 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:34:58,915 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 21:34:59,727 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:34:59,728 - INFO - root : Rocoto call successful on attempt 1: call_time=0.76s, total_time=0.76s -2025-07-28 21:34:59,780 - INFO - root : [END] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:41:01,086 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:41:01,963 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:41:01,963 - INFO - root : Rocoto call successful on attempt 1: call_time=0.82s, total_time=0.82s -2025-07-28 21:41:02,797 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:41:02,797 - INFO - root : Rocoto call successful on attempt 1: call_time=0.77s, total_time=0.77s -2025-07-28 21:41:02,853 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:47:04,158 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:47:05,020 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:47:05,020 - INFO - root : Rocoto call successful on attempt 1: call_time=0.80s, total_time=0.80s -2025-07-28 21:47:05,826 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:47:05,826 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 21:47:05,881 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:53:07,513 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:53:08,381 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:53:08,382 - INFO - root : Rocoto call successful on attempt 1: call_time=0.81s, total_time=0.81s -2025-07-28 21:53:09,184 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:53:09,185 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 21:53:09,237 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:59:10,612 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:59:11,484 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:59:11,484 - INFO - root : Rocoto call successful on attempt 1: call_time=0.81s, total_time=0.81s -2025-07-28 21:59:12,288 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:59:12,289 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 21:59:12,345 - INFO - root : [END] USER_THREAD_COUNT: mterry has 53/1028698 threads (0.01% utilization) -2025-07-28 22:05:13,648 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:05:14,486 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:05:14,486 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 22:05:15,294 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 22:05:15,294 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 22:05:15,348 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 22:11:16,630 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:11:17,476 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:11:17,476 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 22:11:18,281 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 22:11:18,281 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 22:11:18,334 - INFO - root : [END] USER_THREAD_COUNT: mterry has 62/1028698 threads (0.01% utilization) -2025-07-28 22:17:19,648 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:17:20,496 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:17:20,496 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 22:17:21,298 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 80/1028698 threads (0.01% utilization) -2025-07-28 22:17:21,298 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 22:17:21,353 - INFO - root : [END] USER_THREAD_COUNT: mterry has 66/1028698 threads (0.01% utilization) -2025-07-28 22:23:22,642 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:23:23,478 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 53/1028698 threads (0.01% utilization) -2025-07-28 22:23:23,479 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 22:23:24,273 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:23:24,274 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 22:23:24,326 - INFO - root : [END] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 22:29:25,603 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:29:26,439 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:29:26,439 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 22:29:27,237 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:29:27,237 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 22:29:27,289 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:35:29,199 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:35:30,039 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:35:30,039 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 22:35:30,841 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:35:30,841 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 22:35:30,892 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:41:32,177 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:41:33,015 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:41:33,016 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 22:41:33,814 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:41:33,814 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 22:41:33,865 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:47:35,206 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:47:36,050 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:47:36,050 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 22:47:36,853 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:47:36,854 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 22:47:36,905 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:53:38,211 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:53:39,046 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:53:39,046 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 22:53:39,847 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:53:39,847 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 22:53:39,898 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:59:41,168 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:59:42,009 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:59:42,009 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 22:59:42,807 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:59:42,807 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 22:59:42,858 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:05:44,135 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:05:44,973 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:05:44,973 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 23:05:45,778 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:05:45,778 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 23:05:45,829 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:47,098 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:47,940 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:47,940 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 23:11:48,743 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:48,743 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 23:11:48,795 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:50,067 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:50,906 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:50,906 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 23:17:51,706 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:51,707 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 23:17:51,758 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:23:53,069 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:23:53,903 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:23:53,903 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 23:23:54,709 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:23:54,709 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 23:23:54,760 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:29:56,068 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:29:56,901 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:29:56,902 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 23:29:57,701 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:29:57,701 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 23:29:57,752 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:35:59,520 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:36:00,354 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:36:00,354 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 23:36:01,151 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:36:01,151 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 23:36:01,202 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:42:02,815 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:42:03,659 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:42:03,659 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-28 23:42:04,457 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:42:04,457 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-28 23:42:04,509 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:48:06,100 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:48:06,979 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:48:06,979 - INFO - root : Rocoto call successful on attempt 1: call_time=0.82s, total_time=0.82s -2025-07-28 23:48:07,798 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:48:07,799 - INFO - root : Rocoto call successful on attempt 1: call_time=0.76s, total_time=0.76s -2025-07-28 23:48:07,852 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:54:09,438 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:54:10,280 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:54:10,280 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-28 23:54:11,083 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:54:11,083 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-28 23:54:11,135 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:00:12,682 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:00:13,529 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:00:13,529 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-29 00:00:14,340 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:00:14,340 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-29 00:00:14,393 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:06:15,942 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:06:16,781 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:06:16,781 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-29 00:06:17,584 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:06:17,584 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-29 00:06:17,635 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:12:19,184 - INFO - root : [START] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:12:20,055 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:12:20,055 - INFO - root : Rocoto call successful on attempt 1: call_time=0.81s, total_time=0.81s -2025-07-29 00:12:20,863 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:12:20,863 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-29 00:12:20,915 - INFO - root : [END] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:18:22,299 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:18:23,137 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:18:23,137 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-29 00:18:23,937 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:18:23,937 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-29 00:18:23,989 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:24:25,398 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:24:26,240 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:24:26,240 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-29 00:24:27,040 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:24:27,040 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-29 00:24:27,091 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:30:28,978 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:30:29,819 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:30:29,819 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-29 00:30:30,637 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:30:30,637 - INFO - root : Rocoto call successful on attempt 1: call_time=0.76s, total_time=0.76s -2025-07-29 00:30:30,691 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:36:32,580 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:36:33,419 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:36:33,419 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-29 00:36:34,217 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:36:34,217 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-29 00:36:34,267 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:42:36,152 - INFO - root : [START] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:42:36,992 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:42:36,992 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-29 00:42:37,793 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:42:37,793 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-29 00:42:37,846 - INFO - root : [END] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:48:39,517 - INFO - root : [START] USER_THREAD_COUNT: mterry has 74/1028698 threads (0.01% utilization) -2025-07-29 00:48:40,359 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 74/1028698 threads (0.01% utilization) -2025-07-29 00:48:40,359 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-29 00:48:41,169 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 74/1028698 threads (0.01% utilization) -2025-07-29 00:48:41,169 - INFO - root : Rocoto call successful on attempt 1: call_time=0.76s, total_time=0.76s -2025-07-29 00:48:41,221 - INFO - root : [END] USER_THREAD_COUNT: mterry has 74/1028698 threads (0.01% utilization) -2025-07-29 00:54:42,649 - INFO - root : [START] USER_THREAD_COUNT: mterry has 48/1028698 threads (0.0% utilization) -2025-07-29 00:54:43,489 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 48/1028698 threads (0.0% utilization) -2025-07-29 00:54:43,490 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-29 00:54:44,288 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 48/1028698 threads (0.0% utilization) -2025-07-29 00:54:44,288 - INFO - root : Rocoto call successful on attempt 1: call_time=0.74s, total_time=0.74s -2025-07-29 00:54:44,339 - INFO - root : [END] USER_THREAD_COUNT: mterry has 48/1028698 threads (0.0% utilization) -2025-07-29 01:00:45,683 - INFO - root : [START] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:00:46,526 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:00:46,526 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-29 01:00:47,332 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:00:47,332 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-29 01:00:47,384 - INFO - root : [END] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:06:49,025 - INFO - root : [START] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:06:49,877 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:06:49,877 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-29 01:06:50,680 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:06:50,680 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-29 01:06:50,732 - INFO - root : [END] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:12:52,298 - INFO - root : [START] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:12:53,140 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:12:53,140 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-29 01:12:53,942 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:12:53,942 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-29 01:12:53,993 - INFO - root : [END] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:18:55,603 - INFO - root : [START] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:18:56,445 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:18:56,445 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-29 01:18:57,249 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:18:57,249 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-29 01:18:57,301 - INFO - root : [END] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:24:58,584 - INFO - root : [START] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:24:59,424 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:24:59,424 - INFO - root : Rocoto call successful on attempt 1: call_time=0.78s, total_time=0.78s -2025-07-29 01:25:00,230 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:25:00,230 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-29 01:25:00,282 - INFO - root : [END] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:31:01,863 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:31:02,715 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:31:02,715 - INFO - root : Rocoto call successful on attempt 1: call_time=0.79s, total_time=0.79s -2025-07-29 01:31:03,520 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:31:03,520 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-29 01:31:03,572 - INFO - root : [END] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:37:05,201 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:37:06,069 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:37:06,069 - INFO - root : Rocoto call successful on attempt 1: call_time=0.81s, total_time=0.81s -2025-07-29 01:37:06,871 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:37:06,871 - INFO - root : Rocoto call successful on attempt 1: call_time=0.75s, total_time=0.75s -2025-07-29 01:37:06,922 - INFO - root : [END] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) diff --git a/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C48mx500_hybAOWCDA_rocotostat.log b/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C48mx500_hybAOWCDA_rocotostat.log deleted file mode 100644 index b1862a72..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C48mx500_hybAOWCDA_rocotostat.log +++ /dev/null @@ -1,612 +0,0 @@ -2025-07-28 14:49:35,229 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 14:49:35,904 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 69/1028698 threads (0.01% utilization) -2025-07-28 14:49:35,904 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 14:49:36,580 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 63/1028698 threads (0.01% utilization) -2025-07-28 14:49:36,580 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 14:49:36,642 - INFO - root : [END] USER_THREAD_COUNT: mterry has 71/1028698 threads (0.01% utilization) -2025-07-28 14:55:37,788 - INFO - root : [START] USER_THREAD_COUNT: mterry has 71/1028698 threads (0.01% utilization) -2025-07-28 14:55:38,457 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 82/1028698 threads (0.01% utilization) -2025-07-28 14:55:38,458 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 14:55:39,110 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 85/1028698 threads (0.01% utilization) -2025-07-28 14:55:39,110 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 14:55:39,170 - INFO - root : [END] USER_THREAD_COUNT: mterry has 81/1028698 threads (0.01% utilization) -2025-07-28 15:01:40,332 - INFO - root : [START] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 15:01:41,003 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 15:01:41,004 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 15:01:41,662 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 49/1028698 threads (0.0% utilization) -2025-07-28 15:01:41,662 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 15:01:41,721 - INFO - root : [END] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 15:07:43,081 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:07:43,747 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 46/1028698 threads (0.0% utilization) -2025-07-28 15:07:43,747 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 15:07:44,423 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 49/1028698 threads (0.0% utilization) -2025-07-28 15:07:44,423 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 15:07:44,488 - INFO - root : [END] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 15:13:45,805 - INFO - root : [START] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 15:13:46,510 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:13:46,510 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 15:13:47,171 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 66/1028698 threads (0.01% utilization) -2025-07-28 15:13:47,171 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 15:13:47,229 - INFO - root : [END] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 15:19:48,691 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:19:49,389 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:19:49,389 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 15:19:50,053 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 15:19:50,053 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 15:19:50,108 - INFO - root : [END] USER_THREAD_COUNT: mterry has 47/1028698 threads (0.0% utilization) -2025-07-28 15:25:51,294 - INFO - root : [START] USER_THREAD_COUNT: mterry has 46/1028698 threads (0.0% utilization) -2025-07-28 15:25:51,973 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 58/1028698 threads (0.01% utilization) -2025-07-28 15:25:51,973 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 15:25:52,647 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 69/1028698 threads (0.01% utilization) -2025-07-28 15:25:52,647 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 15:25:52,709 - INFO - root : [END] USER_THREAD_COUNT: mterry has 67/1028698 threads (0.01% utilization) -2025-07-28 15:31:53,840 - INFO - root : [START] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 15:31:54,529 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 71/1028698 threads (0.01% utilization) -2025-07-28 15:31:54,529 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 15:31:55,183 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 81/1028698 threads (0.01% utilization) -2025-07-28 15:31:55,184 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 15:31:55,241 - INFO - root : [END] USER_THREAD_COUNT: mterry has 73/1028698 threads (0.01% utilization) -2025-07-28 15:37:56,358 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 15:37:57,034 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 49/1028698 threads (0.0% utilization) -2025-07-28 15:37:57,034 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 15:37:57,685 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 49/1028698 threads (0.0% utilization) -2025-07-28 15:37:57,685 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 15:37:57,741 - INFO - root : [END] USER_THREAD_COUNT: mterry has 59/1028698 threads (0.01% utilization) -2025-07-28 15:43:59,114 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:43:59,790 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 15:43:59,790 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 15:44:00,453 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 46/1028698 threads (0.0% utilization) -2025-07-28 15:44:00,453 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 15:44:00,509 - INFO - root : [END] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:50:01,665 - INFO - root : [START] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-28 15:50:02,432 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 87/1028698 threads (0.01% utilization) -2025-07-28 15:50:02,433 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 15:50:03,128 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 86/1028698 threads (0.01% utilization) -2025-07-28 15:50:03,128 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 15:50:03,190 - INFO - root : [END] USER_THREAD_COUNT: mterry has 71/1028698 threads (0.01% utilization) -2025-07-28 15:56:04,330 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:56:05,003 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 49/1028698 threads (0.0% utilization) -2025-07-28 15:56:05,003 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 15:56:05,664 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 15:56:05,664 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 15:56:05,725 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 16:02:06,859 - INFO - root : [START] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 16:02:07,525 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 64/1028698 threads (0.01% utilization) -2025-07-28 16:02:07,525 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 16:02:08,179 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 53/1028698 threads (0.01% utilization) -2025-07-28 16:02:08,179 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 16:02:08,241 - INFO - root : [END] USER_THREAD_COUNT: mterry has 62/1028698 threads (0.01% utilization) -2025-07-28 16:08:09,383 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 16:08:10,078 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 16:08:10,078 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 16:08:10,743 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 16:08:10,743 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 16:08:10,802 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 16:14:11,893 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 16:14:12,561 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 16:14:12,561 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 16:14:13,217 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 16:14:13,218 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 16:14:13,274 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 16:20:14,399 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 16:20:15,068 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 34/1028698 threads (0.0% utilization) -2025-07-28 16:20:15,068 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 16:20:15,721 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 33/1028698 threads (0.0% utilization) -2025-07-28 16:20:15,721 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 16:20:15,777 - INFO - root : [END] USER_THREAD_COUNT: mterry has 33/1028698 threads (0.0% utilization) -2025-07-28 16:26:16,872 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:26:17,538 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 16:26:17,538 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 16:26:18,202 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 45/1028698 threads (0.0% utilization) -2025-07-28 16:26:18,202 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 16:26:18,257 - INFO - root : [END] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 16:32:19,377 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 16:32:20,040 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 16:32:20,040 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 16:32:20,696 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 58/1028698 threads (0.01% utilization) -2025-07-28 16:32:20,696 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 16:32:20,751 - INFO - root : [END] USER_THREAD_COUNT: mterry has 58/1028698 threads (0.01% utilization) -2025-07-28 16:38:21,905 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:38:22,567 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 16:38:22,567 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 16:38:23,213 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 16:38:23,213 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 16:38:23,267 - INFO - root : [END] USER_THREAD_COUNT: mterry has 45/1028698 threads (0.0% utilization) -2025-07-28 16:44:24,353 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:44:25,023 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 16:44:25,023 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 16:44:25,669 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 16:44:25,669 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 16:44:25,722 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 16:50:26,802 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:27,460 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:50:27,460 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 16:50:28,110 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 16:50:28,110 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 16:50:28,166 - INFO - root : [END] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 16:56:29,251 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:56:29,922 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 16:56:29,922 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 16:56:30,572 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 16:56:30,572 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 16:56:30,625 - INFO - root : [END] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 17:02:31,735 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:02:32,413 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 17:02:32,413 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 17:02:33,061 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 33/1028698 threads (0.0% utilization) -2025-07-28 17:02:33,061 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 17:02:33,115 - INFO - root : [END] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 17:08:34,227 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:08:34,925 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:08:34,925 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 17:08:35,565 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 33/1028698 threads (0.0% utilization) -2025-07-28 17:08:35,566 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 17:08:35,620 - INFO - root : [END] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 17:14:36,724 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:14:37,396 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 17:14:37,397 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 17:14:38,040 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 35/1028698 threads (0.0% utilization) -2025-07-28 17:14:38,040 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 17:14:38,092 - INFO - root : [END] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 17:20:39,190 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:20:39,860 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 17:20:39,860 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 17:20:40,501 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:20:40,501 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 17:20:40,554 - INFO - root : [END] USER_THREAD_COUNT: mterry has 34/1028698 threads (0.0% utilization) -2025-07-28 17:26:41,655 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:26:42,325 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 17:26:42,325 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 17:26:42,970 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:26:42,971 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 17:26:43,024 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:32:44,097 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:32:44,766 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:32:44,766 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 17:32:45,416 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 17:32:45,416 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 17:32:45,471 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:38:46,590 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:38:47,259 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:38:47,259 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 17:38:47,906 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 17:38:47,906 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 17:38:47,962 - INFO - root : [END] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 17:44:49,026 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:44:49,704 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:44:49,704 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 17:44:50,349 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 17:44:50,349 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 17:44:50,404 - INFO - root : [END] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 17:50:51,500 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:50:52,193 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:50:52,193 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 17:50:52,837 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 17:50:52,837 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 17:50:52,891 - INFO - root : [END] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 17:56:53,965 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 17:56:54,628 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 17:56:54,628 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 17:56:55,271 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 17:56:55,271 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 17:56:55,325 - INFO - root : [END] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 18:02:56,407 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:02:57,077 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:02:57,077 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 18:02:57,754 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 18:02:57,755 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 18:02:57,809 - INFO - root : [END] USER_THREAD_COUNT: mterry has 36/1028698 threads (0.0% utilization) -2025-07-28 18:08:58,936 - INFO - root : [START] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 18:08:59,591 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 18:08:59,591 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 18:09:00,232 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 18:09:00,232 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 18:09:00,286 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 18:15:01,355 - INFO - root : [START] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 18:15:02,012 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 18:15:02,012 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 18:15:02,702 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 18:15:02,702 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 18:15:02,768 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:21:03,953 - INFO - root : [START] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 18:21:04,611 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 18:21:04,611 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 18:21:05,255 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:21:05,255 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 18:21:05,310 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:27:06,383 - INFO - root : [START] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 18:27:07,041 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 18:27:07,041 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 18:27:07,680 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:27:07,680 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 18:27:07,733 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:33:08,811 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:09,468 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 18:33:09,468 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 18:33:10,113 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:33:10,113 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 18:33:10,168 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:39:11,416 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 18:39:12,085 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 18:39:12,085 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 18:39:12,726 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 18:39:12,727 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 18:39:12,781 - INFO - root : [END] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 18:45:13,866 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:45:14,545 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:45:14,545 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 18:45:15,196 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:45:15,196 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 18:45:15,256 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:51:16,362 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:51:17,026 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:51:17,026 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 18:51:17,668 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:51:17,668 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 18:51:17,721 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:57:18,799 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:57:19,465 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:57:19,465 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 18:57:20,105 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 18:57:20,105 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 18:57:20,157 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:03:21,510 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:03:22,175 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:03:22,175 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 19:03:22,823 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 19:03:22,823 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 19:03:22,877 - INFO - root : [END] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 19:09:24,011 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:09:24,678 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:24,678 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 19:09:25,319 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 19:09:25,319 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 19:09:25,372 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:15:26,465 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:15:27,133 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:15:27,133 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 19:15:27,780 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 19:15:27,780 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 19:15:27,834 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 19:21:28,953 - INFO - root : [START] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 19:21:29,620 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:21:29,620 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 19:21:30,259 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 19:21:30,259 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 19:21:30,313 - INFO - root : [END] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 19:27:31,432 - INFO - root : [START] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 19:27:32,103 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:32,103 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 19:27:32,743 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-28 19:27:32,743 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 19:27:32,795 - INFO - root : [END] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 19:33:33,907 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 19:33:34,569 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 19:33:34,570 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 19:33:35,212 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:33:35,212 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 19:33:35,264 - INFO - root : [END] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 19:39:36,337 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 19:39:36,995 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 19:39:36,995 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 19:39:37,640 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:39:37,640 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 19:39:37,693 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:45:38,839 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 19:45:39,529 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 19:45:39,530 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 19:45:40,186 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 19:45:40,186 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 19:45:40,238 - INFO - root : [END] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 19:51:41,348 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:51:42,011 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 19:51:42,011 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 19:51:42,658 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 19:51:42,658 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 19:51:42,710 - INFO - root : [END] USER_THREAD_COUNT: mterry has 45/1028698 threads (0.0% utilization) -2025-07-28 19:57:43,841 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:57:44,502 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:57:44,502 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 19:57:45,142 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:57:45,142 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 19:57:45,195 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:03:46,292 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:03:46,954 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:03:46,954 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 20:03:47,596 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:03:47,597 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 20:03:47,649 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:09:48,747 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:09:49,406 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:09:49,406 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 20:09:50,050 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:09:50,050 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 20:09:50,103 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:15:51,232 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:15:51,891 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:15:51,891 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 20:15:52,536 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:15:52,536 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 20:15:52,589 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:21:53,774 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:21:54,432 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:21:54,432 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 20:21:55,074 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:21:55,075 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 20:21:55,128 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:27:56,244 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:27:56,900 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:27:56,900 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 20:27:57,543 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:27:57,543 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 20:27:57,597 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:33:58,714 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:33:59,373 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:33:59,374 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 20:34:00,016 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:34:00,017 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 20:34:00,070 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:01,204 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:01,890 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:01,890 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 20:40:02,615 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:02,616 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 20:40:02,679 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:46:03,857 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:04,518 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:04,518 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 20:46:05,164 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:05,164 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 20:46:05,218 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:06,370 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:07,027 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:07,027 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 20:52:07,677 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:07,677 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 20:52:07,732 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:08,842 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:09,498 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:09,498 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 20:58:10,140 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:10,140 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 20:58:10,193 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:11,301 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:11,958 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:11,958 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 21:04:12,600 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:12,601 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 21:04:12,655 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:13,766 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:14,430 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:14,430 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 21:10:15,070 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:15,070 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 21:10:15,123 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:16:16,238 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:16,901 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:16,901 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 21:16:17,542 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:17,542 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 21:16:17,595 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:22:18,724 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:19,385 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:19,385 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 21:22:20,026 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:20,026 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 21:22:20,078 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:28:21,308 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:28:21,970 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:28:21,970 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 21:28:22,611 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:28:22,611 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 21:28:22,663 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:34:23,783 - INFO - root : [START] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:34:24,445 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:34:24,445 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 21:34:25,086 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:34:25,086 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 21:34:25,139 - INFO - root : [END] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:40:26,288 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:40:26,948 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:40:26,949 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 21:40:27,588 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:40:27,588 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 21:40:27,639 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:46:28,767 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:46:29,427 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:46:29,427 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 21:46:30,068 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:46:30,068 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 21:46:30,120 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:52:31,232 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:52:31,890 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:52:31,890 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 21:52:32,529 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:52:32,529 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 21:52:32,582 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:58:33,689 - INFO - root : [START] USER_THREAD_COUNT: mterry has 66/1028698 threads (0.01% utilization) -2025-07-28 21:58:34,347 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:58:34,347 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 21:58:34,989 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:58:34,989 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 21:58:35,041 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:04:36,135 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:04:36,813 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 22:04:36,813 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 22:04:37,459 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:04:37,459 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 22:04:37,512 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:10:38,661 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:10:39,322 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:10:39,322 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 22:10:39,963 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 53/1028698 threads (0.01% utilization) -2025-07-28 22:10:39,963 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 22:10:40,016 - INFO - root : [END] USER_THREAD_COUNT: mterry has 53/1028698 threads (0.01% utilization) -2025-07-28 22:16:41,144 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:16:41,816 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:16:41,816 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 22:16:42,455 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:16:42,455 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 22:16:42,506 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:22:43,771 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:22:44,437 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:22:44,437 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 22:22:45,078 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:22:45,079 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 22:22:45,132 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:28:46,240 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:28:46,899 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:28:46,899 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 22:28:47,537 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:28:47,538 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 22:28:47,589 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:34:48,695 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:34:49,363 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:34:49,364 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 22:34:50,005 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:34:50,005 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 22:34:50,057 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:40:51,152 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:40:51,817 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:40:51,817 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 22:40:52,455 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:40:52,456 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 22:40:52,507 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:46:53,652 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:46:54,311 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:46:54,311 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 22:46:54,951 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:46:54,951 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 22:46:55,002 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:52:56,093 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:52:56,752 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:52:56,753 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 22:52:57,396 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:52:57,397 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 22:52:57,448 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:58:58,536 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:58:59,198 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:58:59,198 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 22:58:59,841 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:58:59,841 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 22:58:59,892 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:05:01,071 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:05:01,775 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:05:01,775 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 23:05:02,463 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:05:02,463 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 23:05:02,519 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:03,853 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:04,522 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:04,522 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 23:11:05,166 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:05,166 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 23:11:05,217 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:06,399 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:07,088 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:07,088 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 23:17:07,733 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:07,734 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 23:17:07,785 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:23:08,954 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:23:09,636 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:23:09,636 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 23:23:10,276 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:23:10,276 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 23:23:10,328 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:29:11,448 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:29:12,138 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:29:12,138 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 23:29:12,786 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:29:12,786 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 23:29:12,839 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:35:14,019 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:35:14,717 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:35:14,717 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 23:35:15,356 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:35:15,356 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 23:35:15,408 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:41:16,532 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:41:17,191 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:41:17,192 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 23:41:17,830 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:41:17,830 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 23:41:17,881 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:47:18,999 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:47:19,661 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:47:19,661 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 23:47:20,305 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:47:20,305 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 23:47:20,357 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:53:21,476 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:53:22,136 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:53:22,136 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 23:53:22,772 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:53:22,772 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-28 23:53:22,824 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:59:23,977 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:59:24,666 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:59:24,666 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 23:59:25,308 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:59:25,308 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-28 23:59:25,360 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:05:26,489 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:05:27,187 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:05:27,188 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 00:05:27,827 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:05:27,827 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-29 00:05:27,878 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:11:29,689 - INFO - root : [START] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:11:30,351 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:11:30,351 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-29 00:11:30,989 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:11:30,989 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-29 00:11:31,041 - INFO - root : [END] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:17:32,465 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:17:33,126 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:17:33,126 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-29 00:17:33,764 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:17:33,764 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-29 00:17:33,815 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:23:34,961 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:23:35,624 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:23:35,624 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-29 00:23:36,262 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:23:36,262 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-29 00:23:36,314 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:29:37,445 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:29:38,107 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:29:38,107 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-29 00:29:38,747 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:29:38,747 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-29 00:29:38,797 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:35:40,026 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:35:40,691 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:35:40,691 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-29 00:35:41,329 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:35:41,329 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-29 00:35:41,379 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:41:43,004 - INFO - root : [START] USER_THREAD_COUNT: mterry has 77/1028698 threads (0.01% utilization) -2025-07-29 00:41:43,667 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 77/1028698 threads (0.01% utilization) -2025-07-29 00:41:43,667 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-29 00:41:44,307 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 76/1028698 threads (0.01% utilization) -2025-07-29 00:41:44,307 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-29 00:41:44,359 - INFO - root : [END] USER_THREAD_COUNT: mterry has 76/1028698 threads (0.01% utilization) -2025-07-29 00:47:45,789 - INFO - root : [START] USER_THREAD_COUNT: mterry has 90/1028698 threads (0.01% utilization) -2025-07-29 00:47:46,450 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 90/1028698 threads (0.01% utilization) -2025-07-29 00:47:46,451 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-29 00:47:47,089 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 90/1028698 threads (0.01% utilization) -2025-07-29 00:47:47,089 - INFO - root : Rocoto call successful on attempt 1: call_time=0.58s, total_time=0.58s -2025-07-29 00:47:47,141 - INFO - root : [END] USER_THREAD_COUNT: mterry has 90/1028698 threads (0.01% utilization) -2025-07-29 00:53:48,557 - INFO - root : [START] USER_THREAD_COUNT: mterry has 70/1028698 threads (0.01% utilization) -2025-07-29 00:53:49,220 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 70/1028698 threads (0.01% utilization) -2025-07-29 00:53:49,220 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-29 00:53:49,863 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 70/1028698 threads (0.01% utilization) -2025-07-29 00:53:49,864 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-29 00:53:49,916 - INFO - root : [END] USER_THREAD_COUNT: mterry has 70/1028698 threads (0.01% utilization) -2025-07-29 00:59:51,353 - INFO - root : [START] USER_THREAD_COUNT: mterry has 59/1028698 threads (0.01% utilization) -2025-07-29 00:59:52,039 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 59/1028698 threads (0.01% utilization) -2025-07-29 00:59:52,039 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-29 00:59:52,683 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 59/1028698 threads (0.01% utilization) -2025-07-29 00:59:52,683 - INFO - root : Rocoto call successful on attempt 1: call_time=0.59s, total_time=0.59s -2025-07-29 00:59:52,735 - INFO - root : [END] USER_THREAD_COUNT: mterry has 59/1028698 threads (0.01% utilization) diff --git a/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C96C48_hybatmDA_rocotostat.log b/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C96C48_hybatmDA_rocotostat.log deleted file mode 100644 index a9a70197..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C96C48_hybatmDA_rocotostat.log +++ /dev/null @@ -1,714 +0,0 @@ -2025-07-28 14:49:35,407 - INFO - root : [START] USER_THREAD_COUNT: mterry has 33/1028698 threads (0.0% utilization) -2025-07-28 14:49:36,169 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 14:49:36,169 - INFO - root : Rocoto call successful on attempt 1: call_time=0.70s, total_time=0.70s -2025-07-28 14:49:36,910 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 36/1028698 threads (0.0% utilization) -2025-07-28 14:49:36,910 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 14:49:36,965 - INFO - root : [END] USER_THREAD_COUNT: mterry has 34/1028698 threads (0.0% utilization) -2025-07-28 14:55:38,202 - INFO - root : [START] USER_THREAD_COUNT: mterry has 97/1028698 threads (0.01% utilization) -2025-07-28 14:55:38,938 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 116/1028698 threads (0.01% utilization) -2025-07-28 14:55:38,938 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 14:55:39,658 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 90/1028698 threads (0.01% utilization) -2025-07-28 14:55:39,658 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 14:55:39,715 - INFO - root : [END] USER_THREAD_COUNT: mterry has 88/1028698 threads (0.01% utilization) -2025-07-28 15:01:40,883 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 15:01:41,626 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 60/1028698 threads (0.01% utilization) -2025-07-28 15:01:41,626 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 15:01:42,339 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 15:01:42,339 - INFO - root : Rocoto call successful on attempt 1: call_time=0.65s, total_time=0.65s -2025-07-28 15:01:42,396 - INFO - root : [END] USER_THREAD_COUNT: mterry has 68/1028698 threads (0.01% utilization) -2025-07-28 15:07:43,680 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:07:44,437 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 48/1028698 threads (0.0% utilization) -2025-07-28 15:07:44,437 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 15:07:45,159 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 15:07:45,159 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 15:07:45,222 - INFO - root : [END] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 15:13:46,667 - INFO - root : [START] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 15:13:47,401 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 15:13:47,401 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 15:13:48,103 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 46/1028698 threads (0.0% utilization) -2025-07-28 15:13:48,103 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 15:13:48,160 - INFO - root : [END] USER_THREAD_COUNT: mterry has 46/1028698 threads (0.0% utilization) -2025-07-28 15:19:49,414 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:19:50,149 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 15:19:50,149 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 15:19:50,851 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 15:19:50,851 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 15:19:50,907 - INFO - root : [END] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 15:25:52,091 - INFO - root : [START] USER_THREAD_COUNT: mterry has 47/1028698 threads (0.0% utilization) -2025-07-28 15:25:52,844 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 15:25:52,844 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 15:25:53,567 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 51/1028698 threads (0.0% utilization) -2025-07-28 15:25:53,567 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 15:25:53,626 - INFO - root : [END] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 15:31:54,878 - INFO - root : [START] USER_THREAD_COUNT: mterry has 61/1028698 threads (0.01% utilization) -2025-07-28 15:31:55,607 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 63/1028698 threads (0.01% utilization) -2025-07-28 15:31:55,608 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 15:31:56,306 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 64/1028698 threads (0.01% utilization) -2025-07-28 15:31:56,306 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 15:31:56,360 - INFO - root : [END] USER_THREAD_COUNT: mterry has 63/1028698 threads (0.01% utilization) -2025-07-28 15:37:57,524 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 15:37:58,249 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 15:37:58,249 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 15:37:58,941 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 15:37:58,941 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 15:37:58,995 - INFO - root : [END] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 15:44:00,161 - INFO - root : [START] USER_THREAD_COUNT: mterry has 36/1028698 threads (0.0% utilization) -2025-07-28 15:44:00,888 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 15:44:00,889 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 15:44:01,592 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 15:44:01,592 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 15:44:01,650 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 15:50:02,859 - INFO - root : [START] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 15:50:03,646 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 15:50:03,646 - INFO - root : Rocoto call successful on attempt 1: call_time=0.72s, total_time=0.72s -2025-07-28 15:50:04,348 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 15:50:04,348 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 15:50:04,403 - INFO - root : [END] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 15:56:05,580 - INFO - root : [START] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 15:56:06,319 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 15:56:06,319 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 15:56:07,018 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:56:07,018 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 15:56:07,074 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:02:08,244 - INFO - root : [START] USER_THREAD_COUNT: mterry has 62/1028698 threads (0.01% utilization) -2025-07-28 16:02:08,971 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 16:02:08,972 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 16:02:09,671 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 16:02:09,671 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 16:02:09,726 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:08:10,945 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 16:08:11,680 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:08:11,680 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 16:08:12,393 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 16:08:12,393 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 16:08:12,452 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:14:13,624 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:14:14,349 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:14:14,350 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 16:14:15,054 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 16:14:15,054 - INFO - root : Rocoto call successful on attempt 1: call_time=0.65s, total_time=0.65s -2025-07-28 16:14:15,109 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:20:16,279 - INFO - root : [START] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 16:20:17,000 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:20:17,000 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 16:20:17,700 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:20:17,700 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 16:20:17,757 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:26:18,913 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:26:19,632 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:19,633 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 16:26:20,328 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:20,329 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 16:26:20,383 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:32:21,551 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 16:32:22,271 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:32:22,271 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 16:32:22,968 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:32:22,969 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 16:32:23,021 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:38:24,196 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:38:24,913 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:38:24,913 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 16:38:25,609 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:38:25,609 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 16:38:25,661 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:26,817 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:27,535 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:27,535 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 16:44:28,230 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:28,230 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 16:44:28,283 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:29,439 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:30,160 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:30,160 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 16:50:30,857 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:30,857 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 16:50:30,910 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:32,095 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:32,815 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:32,815 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 16:56:33,509 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:33,509 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 16:56:33,561 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:34,715 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:35,434 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:35,434 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 17:02:36,131 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:36,131 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 17:02:36,184 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:37,337 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:38,059 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:38,059 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 17:08:38,753 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:38,753 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 17:08:38,806 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:39,958 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:40,681 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:40,681 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 17:14:41,376 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:41,376 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 17:14:41,429 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:42,564 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:43,292 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:43,292 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 17:20:43,990 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:43,990 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 17:20:44,042 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:45,176 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:45,895 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:45,895 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 17:26:46,591 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:46,591 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 17:26:46,644 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:32:47,784 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:32:48,505 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:32:48,505 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 17:32:49,200 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:32:49,200 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 17:32:49,252 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:38:50,439 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:38:51,157 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:38:51,157 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 17:38:51,846 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:38:51,847 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 17:38:51,898 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:44:53,035 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:44:53,752 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:44:53,752 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 17:44:54,450 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:44:54,450 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 17:44:54,501 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:50:55,636 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:50:56,356 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:50:56,356 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 17:50:57,048 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:50:57,048 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 17:50:57,101 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:56:59,896 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 17:57:00,622 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 17:57:00,623 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 17:57:01,319 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 17:57:01,320 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 17:57:01,377 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:03:02,618 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:03:03,345 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:03:03,346 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 18:03:04,043 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:03:04,043 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 18:03:04,095 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:05,247 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:05,973 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:05,973 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 18:09:06,676 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:06,676 - INFO - root : Rocoto call successful on attempt 1: call_time=0.65s, total_time=0.65s -2025-07-28 18:09:06,732 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:07,891 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:08,613 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:08,613 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 18:15:09,315 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:09,315 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 18:15:09,369 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:10,522 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:11,241 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:11,241 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 18:21:11,937 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:11,937 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 18:21:11,990 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:13,157 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:13,875 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:13,875 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 18:27:14,569 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:14,569 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 18:27:14,621 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:15,777 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:16,498 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:16,498 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 18:33:17,197 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:17,197 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 18:33:17,249 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:39:18,392 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:39:19,114 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:39:19,114 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 18:39:19,808 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:39:19,808 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 18:39:19,860 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:45:21,011 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:45:21,732 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:45:21,733 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 18:45:22,456 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:45:22,456 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 18:45:22,509 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:23,759 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:24,477 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:24,477 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 18:51:25,175 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:25,175 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 18:51:25,227 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:26,457 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:27,174 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:27,174 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 18:57:27,867 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:27,868 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 18:57:27,919 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:03:29,070 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:03:29,791 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:03:29,791 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 19:03:30,484 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:03:30,484 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 19:03:30,536 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:31,690 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:32,411 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:32,411 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 19:09:33,104 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:33,104 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 19:09:33,157 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:15:34,311 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:15:35,033 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:15:35,033 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 19:15:35,730 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:15:35,730 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 19:15:35,784 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:21:36,928 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:21:37,647 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:21:37,647 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 19:21:38,343 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:21:38,343 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 19:21:38,396 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:39,545 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:40,262 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:40,263 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 19:27:40,956 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:40,956 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 19:27:41,008 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:33:47,175 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:33:47,896 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:33:47,896 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 19:33:48,598 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:33:48,598 - INFO - root : Rocoto call successful on attempt 1: call_time=0.65s, total_time=0.65s -2025-07-28 19:33:48,652 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:39:50,162 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:39:50,882 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:39:50,882 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 19:39:51,579 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:39:51,579 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 19:39:51,633 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:45:52,778 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:45:53,497 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:45:53,497 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 19:45:54,191 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:45:54,191 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 19:45:54,243 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:51:55,398 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:51:56,121 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:51:56,121 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 19:51:56,815 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:51:56,815 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 19:51:56,867 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:57:58,032 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:57:58,758 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:57:58,758 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 19:57:59,452 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:57:59,452 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 19:57:59,504 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:04:00,658 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:04:01,379 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:04:01,379 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 20:04:02,075 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:04:02,075 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 20:04:02,136 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:10:03,316 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:10:04,094 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:10:04,095 - INFO - root : Rocoto call successful on attempt 1: call_time=0.72s, total_time=0.72s -2025-07-28 20:10:04,801 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:10:04,801 - INFO - root : Rocoto call successful on attempt 1: call_time=0.65s, total_time=0.65s -2025-07-28 20:10:04,855 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:16:06,047 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:16:06,788 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:16:06,788 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 20:16:07,488 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:16:07,488 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 20:16:07,543 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:22:08,709 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:22:09,432 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:22:09,432 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 20:22:10,138 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:22:10,138 - INFO - root : Rocoto call successful on attempt 1: call_time=0.65s, total_time=0.65s -2025-07-28 20:22:10,190 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:28:11,340 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:28:12,067 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:28:12,067 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 20:28:12,762 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:28:12,762 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 20:28:12,817 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:34:13,982 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:34:14,706 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:34:14,706 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 20:34:20,409 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:34:20,409 - INFO - root : Rocoto call successful on attempt 1: call_time=5.65s, total_time=5.65s -2025-07-28 20:34:20,461 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:21,617 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:22,342 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:22,342 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 20:40:23,037 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:23,037 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 20:40:23,090 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:46:24,254 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:24,973 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:24,973 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 20:46:25,670 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:25,671 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 20:46:25,723 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:26,879 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:27,600 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:27,600 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 20:52:28,294 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:28,295 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 20:52:28,346 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:29,489 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:30,208 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:30,208 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 20:58:30,901 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:30,901 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 20:58:30,953 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:32,106 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:32,827 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:32,827 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 21:04:33,521 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:33,521 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 21:04:33,573 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:34,729 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:35,450 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:35,450 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 21:10:36,144 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:36,144 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 21:10:36,197 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:16:37,354 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:38,072 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:38,072 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 21:16:38,764 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:38,764 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 21:16:38,817 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:22:40,001 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:40,719 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:40,719 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 21:22:41,411 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:41,411 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 21:22:41,463 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:28:42,657 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:28:43,376 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:28:43,376 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 21:28:44,080 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:28:44,080 - INFO - root : Rocoto call successful on attempt 1: call_time=0.65s, total_time=0.65s -2025-07-28 21:28:44,132 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:34:45,306 - INFO - root : [START] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:34:46,026 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:34:46,026 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 21:34:46,718 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:34:46,718 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 21:34:46,772 - INFO - root : [END] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:40:47,956 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:40:48,676 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:40:48,676 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 21:40:49,371 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:40:49,371 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 21:40:49,423 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:46:50,584 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:46:51,306 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:46:51,306 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 21:46:52,000 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:46:52,000 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 21:46:52,052 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:52:53,256 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:52:53,975 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:52:53,975 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 21:52:54,696 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:52:54,696 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 21:52:54,748 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:58:55,951 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:58:56,670 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:58:56,670 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 21:58:57,364 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:58:57,364 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 21:58:57,416 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:04:58,593 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:04:59,317 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:04:59,317 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 22:05:00,019 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:05:00,019 - INFO - root : Rocoto call successful on attempt 1: call_time=0.65s, total_time=0.65s -2025-07-28 22:05:00,072 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:11:01,237 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:11:01,962 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:11:01,963 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 22:11:02,685 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:11:02,685 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 22:11:02,738 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:17:03,937 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:17:04,658 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:17:04,658 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 22:17:10,369 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:17:10,369 - INFO - root : Rocoto call successful on attempt 1: call_time=5.66s, total_time=5.66s -2025-07-28 22:17:10,422 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:23:11,675 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:23:12,401 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:23:12,401 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 22:23:13,093 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:23:13,093 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 22:23:13,145 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:29:14,318 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:29:15,071 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:29:15,071 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 22:29:15,767 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:29:15,767 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 22:29:15,819 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:35:16,997 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:35:17,721 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:35:17,721 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 22:35:18,425 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:35:18,425 - INFO - root : Rocoto call successful on attempt 1: call_time=0.65s, total_time=0.65s -2025-07-28 22:35:18,478 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:41:19,744 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:41:20,463 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:41:20,463 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 22:41:21,157 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:41:21,157 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 22:41:21,209 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:47:22,410 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:47:23,128 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:47:23,128 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 22:47:23,822 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:47:23,822 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 22:47:23,874 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:53:25,033 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:53:25,773 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:53:25,773 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 22:53:26,466 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:53:26,466 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 22:53:26,517 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:59:27,684 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:59:28,406 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:59:28,406 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 22:59:29,100 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:59:29,101 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 22:59:29,152 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:05:30,304 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:05:31,028 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:05:31,028 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 23:05:31,724 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:05:31,724 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 23:05:31,775 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:32,931 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:33,650 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:33,650 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 23:11:34,348 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:34,348 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 23:11:34,400 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:35,559 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:36,277 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:36,277 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 23:17:36,970 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:36,970 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 23:17:37,022 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:23:38,176 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:23:38,893 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:23:38,893 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 23:23:39,584 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:23:39,584 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 23:23:39,634 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:29:40,799 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:29:41,516 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:29:41,516 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 23:29:42,210 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:29:42,210 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 23:29:42,261 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:35:48,435 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:35:49,155 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:35:49,156 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 23:35:49,851 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:35:49,851 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 23:35:49,904 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:41:51,067 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:41:51,787 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:41:51,787 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 23:41:52,478 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:41:52,478 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 23:41:52,529 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:47:53,727 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:47:54,445 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:47:54,445 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 23:47:55,136 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:47:55,136 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 23:47:55,188 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:53:56,348 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:53:57,069 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:53:57,069 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 23:53:57,780 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:53:57,781 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 23:53:57,832 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:59:59,080 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:59:59,798 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:59:59,798 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-29 00:00:00,493 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:00:00,493 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 00:00:00,545 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:06:01,846 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:06:02,584 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:06:02,584 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-29 00:06:03,289 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:06:03,289 - INFO - root : Rocoto call successful on attempt 1: call_time=0.65s, total_time=0.65s -2025-07-29 00:06:03,341 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:12:04,627 - INFO - root : [START] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:12:05,367 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:12:05,367 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-29 00:12:06,091 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:12:06,091 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-29 00:12:06,149 - INFO - root : [END] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:18:12,452 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:18:13,172 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:18:13,173 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-29 00:18:13,866 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:18:13,866 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 00:18:13,917 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:24:15,313 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:24:16,031 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:24:16,032 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-29 00:24:16,724 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:24:16,724 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 00:24:16,775 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:30:18,166 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:30:18,912 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:30:18,912 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-29 00:30:19,605 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:30:19,605 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 00:30:19,657 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:36:21,142 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:36:21,865 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:36:21,865 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-29 00:36:22,560 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:36:22,560 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 00:36:22,611 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:42:24,415 - INFO - root : [START] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:42:25,165 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:42:25,166 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-29 00:42:25,859 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:42:25,859 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 00:42:25,912 - INFO - root : [END] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:48:27,463 - INFO - root : [START] USER_THREAD_COUNT: mterry has 74/1028698 threads (0.01% utilization) -2025-07-29 00:48:28,181 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 74/1028698 threads (0.01% utilization) -2025-07-29 00:48:28,181 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-29 00:48:28,874 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 74/1028698 threads (0.01% utilization) -2025-07-29 00:48:28,874 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 00:48:28,926 - INFO - root : [END] USER_THREAD_COUNT: mterry has 74/1028698 threads (0.01% utilization) -2025-07-29 00:54:30,499 - INFO - root : [START] USER_THREAD_COUNT: mterry has 48/1028698 threads (0.0% utilization) -2025-07-29 00:54:31,221 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 48/1028698 threads (0.0% utilization) -2025-07-29 00:54:31,221 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-29 00:54:31,918 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 48/1028698 threads (0.0% utilization) -2025-07-29 00:54:31,918 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 00:54:31,971 - INFO - root : [END] USER_THREAD_COUNT: mterry has 48/1028698 threads (0.0% utilization) -2025-07-29 01:00:33,238 - INFO - root : [START] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:00:33,965 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:00:33,965 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-29 01:00:34,661 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:00:34,661 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 01:00:34,713 - INFO - root : [END] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:06:36,216 - INFO - root : [START] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:06:36,940 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:06:36,940 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-29 01:06:37,638 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:06:37,638 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 01:06:37,689 - INFO - root : [END] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:12:38,877 - INFO - root : [START] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:12:39,600 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:12:39,600 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-29 01:12:40,292 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:12:40,292 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 01:12:40,344 - INFO - root : [END] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:18:41,865 - INFO - root : [START] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:18:42,614 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:18:42,614 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-29 01:18:43,310 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:18:43,310 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 01:18:43,361 - INFO - root : [END] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:24:44,586 - INFO - root : [START] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:24:45,311 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:24:45,311 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-29 01:24:46,003 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:24:46,003 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 01:24:46,055 - INFO - root : [END] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:30:47,611 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:30:48,332 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:30:48,332 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-29 01:30:49,025 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:30:49,025 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 01:30:49,076 - INFO - root : [END] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:36:50,324 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:36:51,047 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:36:51,047 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-29 01:36:51,739 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:36:51,739 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 01:36:51,791 - INFO - root : [END] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:42:53,290 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:42:54,012 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:42:54,012 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-29 01:42:54,704 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:42:54,704 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 01:42:54,756 - INFO - root : [END] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:48:56,310 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:48:57,031 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:48:57,031 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-29 01:48:57,723 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:48:57,723 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 01:48:57,774 - INFO - root : [END] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:54:58,974 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:54:59,692 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:54:59,692 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-29 01:55:00,386 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:55:00,387 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 01:55:00,437 - INFO - root : [END] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 02:01:01,677 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-29 02:01:02,430 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-29 02:01:02,430 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-29 02:01:03,129 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-29 02:01:03,129 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 02:01:03,181 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-29 02:07:04,405 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 02:07:05,148 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 02:07:05,148 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-29 02:07:05,841 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 02:07:05,841 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 02:07:05,893 - INFO - root : [END] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 02:13:07,166 - INFO - root : [START] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:13:07,887 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:13:07,887 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-29 02:13:08,584 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:13:08,584 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 02:13:08,635 - INFO - root : [END] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:19:10,048 - INFO - root : [START] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:19:10,775 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:19:10,775 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-29 02:19:11,470 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:19:11,470 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 02:19:11,522 - INFO - root : [END] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:25:12,758 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-29 02:25:13,509 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-29 02:25:13,509 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-29 02:25:14,208 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-29 02:25:14,208 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 02:25:14,260 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-29 02:31:15,575 - INFO - root : [START] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:31:16,307 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:31:16,307 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-29 02:31:17,002 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:31:17,002 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 02:31:17,053 - INFO - root : [END] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:37:18,286 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-29 02:37:19,010 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-29 02:37:19,010 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-29 02:37:19,702 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-29 02:37:19,702 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 02:37:19,753 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-29 02:43:20,948 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-29 02:43:21,672 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-29 02:43:21,672 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-29 02:43:22,366 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-29 02:43:22,366 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 02:43:22,418 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) diff --git a/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C96C48mx500_S2SW_cyc_gfs_rocotostat.log b/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C96C48mx500_S2SW_cyc_gfs_rocotostat.log deleted file mode 100644 index 7e018fe1..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C96C48mx500_S2SW_cyc_gfs_rocotostat.log +++ /dev/null @@ -1,708 +0,0 @@ -2025-07-28 14:49:35,835 - INFO - root : [START] USER_THREAD_COUNT: mterry has 101/1028698 threads (0.01% utilization) -2025-07-28 14:49:36,817 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 60/1028698 threads (0.01% utilization) -2025-07-28 14:49:36,817 - INFO - root : Rocoto call successful on attempt 1: call_time=0.92s, total_time=0.92s -2025-07-28 14:49:37,723 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 14:49:37,723 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 14:49:37,775 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 14:55:39,265 - INFO - root : [START] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-28 14:55:40,276 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 97/1028698 threads (0.01% utilization) -2025-07-28 14:55:40,276 - INFO - root : Rocoto call successful on attempt 1: call_time=0.95s, total_time=0.95s -2025-07-28 14:55:41,184 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 67/1028698 threads (0.01% utilization) -2025-07-28 14:55:41,184 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 14:55:41,238 - INFO - root : [END] USER_THREAD_COUNT: mterry has 67/1028698 threads (0.01% utilization) -2025-07-28 15:01:42,666 - INFO - root : [START] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 15:01:43,624 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:01:43,624 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 15:01:44,526 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:01:44,526 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 15:01:44,579 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:07:45,995 - INFO - root : [START] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 15:07:46,956 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:07:46,956 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 15:07:47,857 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:07:47,857 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 15:07:47,912 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:13:49,681 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 15:13:50,637 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:13:50,638 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 15:13:51,538 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:13:51,538 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 15:13:51,592 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:19:53,105 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:19:54,054 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:19:54,054 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 15:19:54,953 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:19:54,953 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 15:19:55,006 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:25:56,427 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 15:25:57,391 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 15:25:57,391 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 15:25:58,310 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 15:25:58,310 - INFO - root : Rocoto call successful on attempt 1: call_time=0.86s, total_time=0.86s -2025-07-28 15:25:58,367 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 15:31:59,798 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:32:00,755 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:32:00,756 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 15:32:01,669 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:32:01,669 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 15:32:01,728 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:38:03,193 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:38:04,174 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:38:04,174 - INFO - root : Rocoto call successful on attempt 1: call_time=0.92s, total_time=0.92s -2025-07-28 15:38:05,069 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 15:38:05,070 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 15:38:05,120 - INFO - root : [END] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 15:44:06,567 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:44:07,516 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:44:07,517 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 15:44:08,423 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:44:08,423 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 15:44:08,478 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:50:09,911 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:50:10,860 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:50:10,860 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 15:50:11,766 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:50:11,766 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 15:50:11,822 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:56:13,254 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:56:14,207 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:56:14,207 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 15:56:15,106 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:56:15,106 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 15:56:15,159 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:02:16,572 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 16:02:17,529 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 16:02:17,529 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 16:02:18,430 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:02:18,430 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 16:02:18,484 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:08:19,944 - INFO - root : [START] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 16:08:20,902 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 16:08:20,902 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 16:08:21,812 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:08:21,812 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 16:08:21,870 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:14:23,302 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 16:14:24,257 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:14:24,257 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 16:14:25,163 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:14:25,163 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 16:14:25,216 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:20:26,623 - INFO - root : [START] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 16:20:27,583 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:20:27,583 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 16:20:28,486 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:20:28,486 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 16:20:28,540 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:29,984 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:26:30,943 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:30,943 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 16:26:31,840 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:31,840 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 16:26:31,893 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:32:33,307 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:32:34,256 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:32:34,256 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 16:32:35,166 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:32:35,166 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 16:32:35,219 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:38:36,753 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:38:37,709 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:38:37,709 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 16:38:38,611 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:38:38,611 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 16:38:38,662 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:40,127 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:41,158 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:41,158 - INFO - root : Rocoto call successful on attempt 1: call_time=0.97s, total_time=0.97s -2025-07-28 16:44:42,058 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:42,058 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 16:44:42,109 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:43,529 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:44,484 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:44,484 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 16:50:45,390 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:45,390 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 16:50:45,443 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:46,851 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:47,798 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:47,798 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 16:56:48,694 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:48,694 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 16:56:48,746 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:50,179 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:51,128 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:51,128 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 17:02:52,033 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:52,033 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 17:02:52,086 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:53,502 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:54,448 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:54,448 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 17:08:55,345 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:55,345 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 17:08:55,397 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:56,845 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:57,804 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:57,804 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 17:14:58,704 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:58,704 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 17:14:58,756 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:21:00,311 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:21:01,265 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:21:01,265 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 17:21:02,172 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:21:02,172 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 17:21:02,228 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:27:03,654 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:27:04,632 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:27:04,632 - INFO - root : Rocoto call successful on attempt 1: call_time=0.92s, total_time=0.92s -2025-07-28 17:27:05,534 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:27:05,534 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 17:27:05,589 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:33:07,002 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:33:07,954 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:33:07,954 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 17:33:08,857 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:33:08,857 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 17:33:08,912 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:39:10,336 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:39:11,305 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:39:11,305 - INFO - root : Rocoto call successful on attempt 1: call_time=0.91s, total_time=0.91s -2025-07-28 17:39:12,211 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:39:12,211 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 17:39:12,263 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:45:13,664 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:45:14,625 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:45:14,625 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 17:45:15,524 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:45:15,525 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 17:45:15,575 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:51:16,968 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:51:17,916 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:51:17,916 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 17:51:18,818 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:51:18,818 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 17:51:18,870 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:57:20,271 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 17:57:21,220 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 17:57:21,220 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 17:57:22,116 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 17:57:22,116 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 17:57:22,168 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:03:23,565 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:03:24,539 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:03:24,539 - INFO - root : Rocoto call successful on attempt 1: call_time=0.91s, total_time=0.91s -2025-07-28 18:03:25,455 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:03:25,455 - INFO - root : Rocoto call successful on attempt 1: call_time=0.86s, total_time=0.86s -2025-07-28 18:03:25,507 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:26,911 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:27,860 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:27,860 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 18:09:28,760 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:28,761 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 18:09:28,812 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:30,223 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:31,172 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:31,173 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 18:15:32,065 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:32,065 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 18:15:32,118 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:33,524 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:34,479 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:34,479 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 18:21:35,371 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:35,371 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 18:21:35,422 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:36,840 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:37,785 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:37,785 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 18:27:38,686 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:38,686 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 18:27:38,738 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:40,141 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:41,093 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:41,093 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 18:33:41,988 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:41,988 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 18:33:42,039 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:39:43,514 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:39:44,490 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:39:44,490 - INFO - root : Rocoto call successful on attempt 1: call_time=0.92s, total_time=0.92s -2025-07-28 18:39:45,387 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:39:45,387 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 18:39:45,439 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:45:46,849 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:45:47,799 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:45:47,800 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 18:45:48,695 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:45:48,695 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 18:45:48,747 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:50,181 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:51,127 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:51,127 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 18:51:52,023 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:52,023 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 18:51:52,074 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:53,585 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:54,535 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:54,535 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 18:57:55,432 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:55,432 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 18:57:55,483 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:03:56,918 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:03:57,881 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:03:57,881 - INFO - root : Rocoto call successful on attempt 1: call_time=0.91s, total_time=0.91s -2025-07-28 19:03:58,784 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:03:58,785 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 19:03:58,836 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:10:00,241 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:10:01,189 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:10:01,189 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 19:10:02,108 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:10:02,108 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 19:10:02,173 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:16:03,696 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:16:04,651 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:16:04,652 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 19:16:05,559 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:16:05,559 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 19:16:05,618 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:22:07,009 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:22:07,966 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:22:07,966 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 19:22:08,875 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:22:08,875 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 19:22:08,929 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:28:10,321 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:28:11,278 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:28:11,278 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 19:28:12,186 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:28:12,186 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 19:28:12,239 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:34:13,634 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:34:14,587 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:34:14,587 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 19:34:15,492 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:34:15,492 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 19:34:15,545 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:40:16,948 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:40:17,922 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:40:17,922 - INFO - root : Rocoto call successful on attempt 1: call_time=0.91s, total_time=0.91s -2025-07-28 19:40:18,820 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:40:18,820 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 19:40:18,872 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:46:20,276 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:46:21,232 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:46:21,233 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 19:46:22,136 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:46:22,136 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 19:46:22,189 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:52:23,570 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:52:24,525 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:52:24,525 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 19:52:25,428 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:52:25,428 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 19:52:25,480 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:58:26,889 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:58:27,843 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:58:27,843 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 19:58:28,743 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:58:28,743 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 19:58:28,795 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:04:30,185 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:04:31,141 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:04:31,141 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 20:04:32,044 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:04:32,045 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 20:04:32,096 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:10:33,504 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:10:34,460 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:10:34,460 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 20:10:35,367 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:10:35,367 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 20:10:35,419 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:16:36,812 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:16:37,767 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:16:37,767 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 20:16:38,668 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:16:38,668 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 20:16:38,720 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:22:40,140 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:22:41,089 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:22:41,089 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 20:22:41,986 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:22:41,986 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 20:22:42,038 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:28:43,478 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:28:44,527 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:28:44,528 - INFO - root : Rocoto call successful on attempt 1: call_time=0.99s, total_time=0.99s -2025-07-28 20:28:45,432 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:28:45,432 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 20:28:45,484 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:34:46,876 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:34:47,832 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:34:47,832 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 20:34:48,731 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:34:48,731 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 20:34:48,784 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:50,211 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:51,164 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:51,164 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 20:40:52,066 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:52,066 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 20:40:52,117 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:46:53,515 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:54,473 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:54,473 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 20:46:55,377 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:55,377 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 20:46:55,428 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:56,834 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:57,794 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:57,794 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 20:52:58,691 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:58,691 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 20:52:58,743 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:59:00,148 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:59:01,106 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:59:01,106 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 20:59:02,019 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:59:02,019 - INFO - root : Rocoto call successful on attempt 1: call_time=0.86s, total_time=0.86s -2025-07-28 20:59:02,073 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:05:03,502 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:05:04,472 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:05:04,472 - INFO - root : Rocoto call successful on attempt 1: call_time=0.91s, total_time=0.91s -2025-07-28 21:05:05,378 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:05:05,378 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 21:05:05,431 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:11:06,829 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:11:07,784 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:11:07,784 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 21:11:08,685 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:11:08,685 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 21:11:08,737 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:17:10,144 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:17:11,101 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:17:11,101 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 21:17:12,011 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:17:12,011 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 21:17:12,065 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:23:14,059 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:23:15,018 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:23:15,018 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 21:23:15,914 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:23:15,914 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 21:23:15,966 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:29:17,482 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:29:18,462 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:29:18,462 - INFO - root : Rocoto call successful on attempt 1: call_time=0.92s, total_time=0.92s -2025-07-28 21:29:19,358 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:29:19,359 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 21:29:19,412 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:35:20,886 - INFO - root : [START] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:35:21,840 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:35:21,840 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 21:35:22,740 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:35:22,740 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 21:35:22,793 - INFO - root : [END] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:41:24,246 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:41:25,193 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:41:25,194 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 21:41:26,101 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:41:26,101 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 21:41:26,153 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:47:27,584 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:47:28,540 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:47:28,541 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 21:47:29,439 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:47:29,439 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 21:47:29,491 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:53:30,948 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:53:31,896 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:53:31,896 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 21:53:32,802 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:53:32,802 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 21:53:32,854 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:59:34,272 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:59:35,224 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:59:35,225 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 21:59:36,121 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:59:36,121 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 21:59:36,174 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:05:37,669 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:05:38,625 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:05:38,625 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 22:05:39,527 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:05:39,527 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 22:05:39,579 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:11:41,121 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:11:42,081 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:11:42,081 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 22:11:42,982 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:11:42,982 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 22:11:43,034 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:17:44,478 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:17:45,430 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:17:45,430 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 22:17:46,333 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:17:46,333 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 22:17:46,387 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:23:47,817 - INFO - root : [START] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 22:23:48,763 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 22:23:48,763 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 22:23:49,666 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 22:23:49,666 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 22:23:49,718 - INFO - root : [END] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 22:29:51,128 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:29:52,081 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:29:52,081 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 22:29:52,983 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:29:52,983 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 22:29:53,036 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:35:54,432 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:35:55,381 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:35:55,381 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 22:35:56,277 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:35:56,277 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 22:35:56,328 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:41:57,740 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:41:58,697 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:41:58,697 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 22:41:59,597 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:41:59,597 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 22:41:59,648 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:48:01,097 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:48:02,049 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:48:02,049 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 22:48:02,977 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:48:02,977 - INFO - root : Rocoto call successful on attempt 1: call_time=0.87s, total_time=0.87s -2025-07-28 22:48:03,032 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:54:04,436 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:54:05,395 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:54:05,396 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 22:54:06,305 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:54:06,305 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 22:54:06,358 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:00:07,801 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:00:08,822 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:00:08,823 - INFO - root : Rocoto call successful on attempt 1: call_time=0.96s, total_time=0.96s -2025-07-28 23:00:09,739 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:00:09,739 - INFO - root : Rocoto call successful on attempt 1: call_time=0.86s, total_time=0.86s -2025-07-28 23:00:09,794 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:06:11,205 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:06:12,157 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:06:12,157 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 23:06:13,055 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:06:13,055 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 23:06:13,107 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:12:14,561 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:12:15,517 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:12:15,517 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 23:12:16,421 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:12:16,421 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-28 23:12:16,473 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:18:17,963 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:18:18,916 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:18:18,917 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 23:18:19,812 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:18:19,813 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 23:18:19,864 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:24:21,300 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:24:22,251 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:24:22,252 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 23:24:23,147 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:24:23,147 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 23:24:23,198 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:30:24,610 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:30:25,606 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:30:25,606 - INFO - root : Rocoto call successful on attempt 1: call_time=0.94s, total_time=0.94s -2025-07-28 23:30:26,554 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:30:26,554 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 23:30:26,607 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:36:28,184 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:36:29,163 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:36:29,164 - INFO - root : Rocoto call successful on attempt 1: call_time=0.92s, total_time=0.92s -2025-07-28 23:36:30,063 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:36:30,063 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 23:36:30,114 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:42:31,762 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:42:32,718 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:42:32,718 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-28 23:42:33,614 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:42:33,614 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 23:42:33,665 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:48:35,226 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:48:36,177 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:48:36,178 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 23:48:37,074 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:48:37,074 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 23:48:37,125 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:54:38,676 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:54:39,626 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:54:39,626 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-28 23:54:40,523 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:54:40,523 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-28 23:54:40,573 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:00:42,120 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:00:43,106 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:00:43,106 - INFO - root : Rocoto call successful on attempt 1: call_time=0.93s, total_time=0.93s -2025-07-29 00:00:44,003 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:00:44,003 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-29 00:00:44,053 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:06:45,889 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:06:46,835 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:06:46,835 - INFO - root : Rocoto call successful on attempt 1: call_time=0.89s, total_time=0.89s -2025-07-29 00:06:47,730 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:06:47,730 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-29 00:06:47,781 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:12:49,416 - INFO - root : [START] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:12:50,374 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:12:50,374 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-29 00:12:51,275 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:12:51,275 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-29 00:12:51,326 - INFO - root : [END] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:18:53,075 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-29 00:18:54,034 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:18:54,035 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-29 00:18:54,940 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:18:54,941 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-29 00:18:54,991 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:24:56,648 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:24:57,608 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:24:57,608 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-29 00:24:58,504 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:24:58,504 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-29 00:24:58,555 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:31:00,321 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:31:01,276 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:31:01,276 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-29 00:31:02,187 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:31:02,187 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-29 00:31:02,241 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:37:03,919 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:37:04,879 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:37:04,879 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-29 00:37:05,778 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:37:05,778 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-29 00:37:05,829 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:43:07,367 - INFO - root : [START] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:43:08,324 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:43:08,324 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-29 00:43:09,223 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:43:09,223 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-29 00:43:09,275 - INFO - root : [END] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:49:10,924 - INFO - root : [START] USER_THREAD_COUNT: mterry has 74/1028698 threads (0.01% utilization) -2025-07-29 00:49:11,899 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 74/1028698 threads (0.01% utilization) -2025-07-29 00:49:11,899 - INFO - root : Rocoto call successful on attempt 1: call_time=0.92s, total_time=0.92s -2025-07-29 00:49:12,802 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 74/1028698 threads (0.01% utilization) -2025-07-29 00:49:12,802 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-29 00:49:12,853 - INFO - root : [END] USER_THREAD_COUNT: mterry has 74/1028698 threads (0.01% utilization) -2025-07-29 00:55:14,775 - INFO - root : [START] USER_THREAD_COUNT: mterry has 46/1028698 threads (0.0% utilization) -2025-07-29 00:55:15,739 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 46/1028698 threads (0.0% utilization) -2025-07-29 00:55:15,739 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-29 00:55:16,641 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 46/1028698 threads (0.0% utilization) -2025-07-29 00:55:16,641 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-29 00:55:16,693 - INFO - root : [END] USER_THREAD_COUNT: mterry has 46/1028698 threads (0.0% utilization) -2025-07-29 01:01:18,173 - INFO - root : [START] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:01:19,126 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:01:19,126 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-29 01:01:20,023 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:01:20,023 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-29 01:01:20,075 - INFO - root : [END] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:07:21,949 - INFO - root : [START] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:07:22,914 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:07:22,915 - INFO - root : Rocoto call successful on attempt 1: call_time=0.91s, total_time=0.91s -2025-07-29 01:07:23,809 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:07:23,809 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-29 01:07:23,859 - INFO - root : [END] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:13:25,742 - INFO - root : [START] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:13:26,712 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:13:26,712 - INFO - root : Rocoto call successful on attempt 1: call_time=0.91s, total_time=0.91s -2025-07-29 01:13:27,614 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:13:27,614 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-29 01:13:27,666 - INFO - root : [END] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:19:29,160 - INFO - root : [START] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:19:30,121 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:19:30,121 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-29 01:19:31,019 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:19:31,019 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-29 01:19:31,070 - INFO - root : [END] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:25:32,591 - INFO - root : [START] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:25:33,543 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:25:33,543 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-29 01:25:34,448 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:25:34,448 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-29 01:25:34,500 - INFO - root : [END] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:31:36,595 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:31:37,585 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:31:37,585 - INFO - root : Rocoto call successful on attempt 1: call_time=0.93s, total_time=0.93s -2025-07-29 01:31:38,485 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:31:38,485 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-29 01:31:38,536 - INFO - root : [END] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:37:40,101 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-29 01:37:41,058 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-29 01:37:41,058 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-29 01:37:41,953 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-29 01:37:41,953 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-29 01:37:42,004 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-29 01:43:43,971 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:43:44,926 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:43:44,926 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-29 01:43:45,826 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:43:45,826 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-29 01:43:45,877 - INFO - root : [END] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:49:47,797 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:49:48,751 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:49:48,751 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-29 01:49:49,645 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:49:49,645 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-29 01:49:49,696 - INFO - root : [END] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:55:51,519 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:55:52,475 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:55:52,475 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-29 01:55:53,390 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:55:53,390 - INFO - root : Rocoto call successful on attempt 1: call_time=0.86s, total_time=0.86s -2025-07-29 01:55:53,441 - INFO - root : [END] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 02:01:55,273 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-29 02:01:56,234 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-29 02:01:56,234 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-29 02:01:57,135 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-29 02:01:57,135 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-29 02:01:57,186 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-29 02:07:58,676 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 02:07:59,631 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 02:07:59,631 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-29 02:08:00,525 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 02:08:00,525 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-29 02:08:00,577 - INFO - root : [END] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 02:14:02,118 - INFO - root : [START] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:14:03,098 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:14:03,098 - INFO - root : Rocoto call successful on attempt 1: call_time=0.92s, total_time=0.92s -2025-07-29 02:14:04,007 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:14:04,007 - INFO - root : Rocoto call successful on attempt 1: call_time=0.85s, total_time=0.85s -2025-07-29 02:14:04,058 - INFO - root : [END] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:20:05,594 - INFO - root : [START] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:20:06,592 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:20:06,592 - INFO - root : Rocoto call successful on attempt 1: call_time=0.94s, total_time=0.94s -2025-07-29 02:20:07,510 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:20:07,510 - INFO - root : Rocoto call successful on attempt 1: call_time=0.86s, total_time=0.86s -2025-07-29 02:20:07,565 - INFO - root : [END] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:26:09,388 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-29 02:26:10,343 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-29 02:26:10,343 - INFO - root : Rocoto call successful on attempt 1: call_time=0.90s, total_time=0.90s -2025-07-29 02:26:11,239 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-29 02:26:11,239 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-29 02:26:11,291 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-29 02:32:12,900 - INFO - root : [START] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:32:13,868 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:32:13,868 - INFO - root : Rocoto call successful on attempt 1: call_time=0.91s, total_time=0.91s -2025-07-29 02:32:14,766 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:32:14,766 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-29 02:32:14,817 - INFO - root : [END] USER_THREAD_COUNT: mterry has 14/1028698 threads (0.0% utilization) -2025-07-29 02:38:16,673 - INFO - root : [START] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-29 02:38:17,655 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-29 02:38:17,655 - INFO - root : Rocoto call successful on attempt 1: call_time=0.92s, total_time=0.92s -2025-07-29 02:38:18,550 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-29 02:38:18,550 - INFO - root : Rocoto call successful on attempt 1: call_time=0.84s, total_time=0.84s -2025-07-29 02:38:18,601 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) diff --git a/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C96_atm3DVar_rocotostat.log b/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C96_atm3DVar_rocotostat.log deleted file mode 100644 index 19c968ac..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C96_atm3DVar_rocotostat.log +++ /dev/null @@ -1,684 +0,0 @@ -2025-07-28 14:49:35,245 - INFO - root : [START] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 14:49:35,933 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 61/1028698 threads (0.01% utilization) -2025-07-28 14:49:35,933 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 14:49:36,601 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 70/1028698 threads (0.01% utilization) -2025-07-28 14:49:36,601 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 14:49:36,660 - INFO - root : [END] USER_THREAD_COUNT: mterry has 66/1028698 threads (0.01% utilization) -2025-07-28 14:55:37,785 - INFO - root : [START] USER_THREAD_COUNT: mterry has 71/1028698 threads (0.01% utilization) -2025-07-28 14:55:38,470 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 87/1028698 threads (0.01% utilization) -2025-07-28 14:55:38,470 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 14:55:39,146 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 83/1028698 threads (0.01% utilization) -2025-07-28 14:55:39,146 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 14:55:39,208 - INFO - root : [END] USER_THREAD_COUNT: mterry has 104/1028698 threads (0.01% utilization) -2025-07-28 15:01:40,331 - INFO - root : [START] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 15:01:41,020 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 15:01:41,020 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 15:01:41,693 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 15:01:41,693 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 15:01:41,752 - INFO - root : [END] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 15:07:42,974 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 15:07:43,656 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 15:07:43,656 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 15:07:44,343 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 67/1028698 threads (0.01% utilization) -2025-07-28 15:07:44,343 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 15:07:44,408 - INFO - root : [END] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 15:13:45,802 - INFO - root : [START] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 15:13:46,528 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 33/1028698 threads (0.0% utilization) -2025-07-28 15:13:46,529 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 15:13:47,205 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 15:13:47,206 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 15:13:47,266 - INFO - root : [END] USER_THREAD_COUNT: mterry has 53/1028698 threads (0.01% utilization) -2025-07-28 15:19:48,690 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:19:49,389 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:19:49,389 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 15:19:50,064 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 53/1028698 threads (0.01% utilization) -2025-07-28 15:19:50,064 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 15:19:50,120 - INFO - root : [END] USER_THREAD_COUNT: mterry has 45/1028698 threads (0.0% utilization) -2025-07-28 15:25:51,247 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 15:25:51,950 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 58/1028698 threads (0.01% utilization) -2025-07-28 15:25:51,950 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 15:25:52,647 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 69/1028698 threads (0.01% utilization) -2025-07-28 15:25:52,647 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 15:25:52,709 - INFO - root : [END] USER_THREAD_COUNT: mterry has 67/1028698 threads (0.01% utilization) -2025-07-28 15:31:53,841 - INFO - root : [START] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 15:31:54,557 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 72/1028698 threads (0.01% utilization) -2025-07-28 15:31:54,557 - INFO - root : Rocoto call successful on attempt 1: call_time=0.65s, total_time=0.65s -2025-07-28 15:31:55,224 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 73/1028698 threads (0.01% utilization) -2025-07-28 15:31:55,224 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 15:31:55,284 - INFO - root : [END] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 15:37:56,381 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:37:57,056 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 53/1028698 threads (0.01% utilization) -2025-07-28 15:37:57,057 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 15:37:57,717 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 15:37:57,717 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 15:37:57,777 - INFO - root : [END] USER_THREAD_COUNT: mterry has 60/1028698 threads (0.01% utilization) -2025-07-28 15:43:59,114 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:43:59,811 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 33/1028698 threads (0.0% utilization) -2025-07-28 15:43:59,811 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 15:44:00,481 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 15:44:00,481 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 15:44:00,539 - INFO - root : [END] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 15:50:01,732 - INFO - root : [START] USER_THREAD_COUNT: mterry has 59/1028698 threads (0.01% utilization) -2025-07-28 15:50:02,489 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 59/1028698 threads (0.01% utilization) -2025-07-28 15:50:02,489 - INFO - root : Rocoto call successful on attempt 1: call_time=0.68s, total_time=0.68s -2025-07-28 15:50:03,226 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-28 15:50:03,226 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 15:50:03,293 - INFO - root : [END] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-28 15:56:04,448 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:56:05,138 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 45/1028698 threads (0.0% utilization) -2025-07-28 15:56:05,138 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 15:56:05,817 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 15:56:05,817 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 15:56:05,874 - INFO - root : [END] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 16:02:07,024 - INFO - root : [START] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 16:02:07,709 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 58/1028698 threads (0.01% utilization) -2025-07-28 16:02:07,709 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 16:02:08,389 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-28 16:02:08,389 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 16:02:08,446 - INFO - root : [END] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 16:08:09,589 - INFO - root : [START] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 16:08:10,282 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 16:08:10,282 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 16:08:10,961 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 16:08:10,962 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 16:08:11,019 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:14:12,126 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 16:14:12,814 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:14:12,814 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 16:14:13,520 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 16:14:13,520 - INFO - root : Rocoto call successful on attempt 1: call_time=0.65s, total_time=0.65s -2025-07-28 16:14:13,577 - INFO - root : [END] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 16:20:14,696 - INFO - root : [START] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 16:20:15,380 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 16:20:15,380 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 16:20:16,045 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 16:20:16,045 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 16:20:16,102 - INFO - root : [END] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 16:26:17,222 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 16:26:17,908 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 16:26:17,908 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 16:26:18,575 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 16:26:18,575 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 16:26:18,629 - INFO - root : [END] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 16:32:19,736 - INFO - root : [START] USER_THREAD_COUNT: mterry has 53/1028698 threads (0.01% utilization) -2025-07-28 16:32:20,419 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 55/1028698 threads (0.01% utilization) -2025-07-28 16:32:20,419 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 16:32:21,083 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 16:32:21,083 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 16:32:21,136 - INFO - root : [END] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 16:38:22,239 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 16:38:22,918 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:38:22,918 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 16:38:23,579 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:38:23,579 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 16:38:23,632 - INFO - root : [END] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 16:44:24,732 - INFO - root : [START] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 16:44:25,410 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 16:44:25,410 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 16:44:26,097 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:44:26,097 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 16:44:26,149 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:50:27,280 - INFO - root : [START] USER_THREAD_COUNT: mterry has 48/1028698 threads (0.0% utilization) -2025-07-28 16:50:27,967 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 49/1028698 threads (0.0% utilization) -2025-07-28 16:50:27,967 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 16:50:28,625 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:50:28,626 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 16:50:28,679 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:56:29,776 - INFO - root : [START] USER_THREAD_COUNT: mterry has 48/1028698 threads (0.0% utilization) -2025-07-28 16:56:30,483 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 53/1028698 threads (0.01% utilization) -2025-07-28 16:56:30,483 - INFO - root : Rocoto call successful on attempt 1: call_time=0.65s, total_time=0.65s -2025-07-28 16:56:31,146 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:56:31,146 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 16:56:31,199 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:02:32,308 - INFO - root : [START] USER_THREAD_COUNT: mterry has 46/1028698 threads (0.0% utilization) -2025-07-28 17:02:32,991 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 59/1028698 threads (0.01% utilization) -2025-07-28 17:02:32,991 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 17:02:33,649 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 17:02:33,649 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 17:02:33,703 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:08:34,812 - INFO - root : [START] USER_THREAD_COUNT: mterry has 48/1028698 threads (0.0% utilization) -2025-07-28 17:08:35,491 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 61/1028698 threads (0.01% utilization) -2025-07-28 17:08:35,491 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 17:08:36,148 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 17:08:36,148 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 17:08:36,201 - INFO - root : [END] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 17:14:37,312 - INFO - root : [START] USER_THREAD_COUNT: mterry has 58/1028698 threads (0.01% utilization) -2025-07-28 17:14:37,989 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 17:14:37,989 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 17:14:38,645 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:14:38,645 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 17:14:38,697 - INFO - root : [END] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 17:20:39,798 - INFO - root : [START] USER_THREAD_COUNT: mterry has 58/1028698 threads (0.01% utilization) -2025-07-28 17:20:40,480 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 17:20:40,480 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 17:20:41,177 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 17:20:41,177 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 17:20:41,230 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:42,327 - INFO - root : [START] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 17:26:43,002 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:26:43,002 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 17:26:43,660 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 17:26:43,660 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 17:26:43,714 - INFO - root : [END] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 17:32:44,811 - INFO - root : [START] USER_THREAD_COUNT: mterry has 34/1028698 threads (0.0% utilization) -2025-07-28 17:32:45,487 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:32:45,487 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 17:32:46,148 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 17:32:46,148 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 17:32:46,202 - INFO - root : [END] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 17:38:47,337 - INFO - root : [START] USER_THREAD_COUNT: mterry has 34/1028698 threads (0.0% utilization) -2025-07-28 17:38:48,018 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 17:38:48,018 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 17:38:48,671 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 17:38:48,671 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 17:38:48,724 - INFO - root : [END] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 17:44:49,822 - INFO - root : [START] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 17:44:50,496 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 17:44:50,496 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 17:44:51,152 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 17:44:51,152 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 17:44:51,205 - INFO - root : [END] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 17:50:52,304 - INFO - root : [START] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 17:50:52,979 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:50:52,980 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 17:50:53,636 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 17:50:53,637 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 17:50:53,689 - INFO - root : [END] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 17:56:54,796 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 17:56:55,498 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 17:56:55,498 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 17:56:56,156 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 17:56:56,156 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 17:56:56,210 - INFO - root : [END] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 18:02:57,313 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:02:58,026 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:02:58,026 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 18:02:58,689 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 18:02:58,689 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 18:02:58,742 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 18:08:59,833 - INFO - root : [START] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 18:09:00,505 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:00,506 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 18:09:01,167 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:09:01,167 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 18:09:01,221 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 18:15:02,339 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 18:15:03,095 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:03,095 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 18:15:03,789 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:15:03,790 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 18:15:03,850 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 18:21:04,979 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 18:21:05,667 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:05,667 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 18:21:06,334 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:21:06,334 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 18:21:06,389 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 18:27:07,489 - INFO - root : [START] USER_THREAD_COUNT: mterry has 47/1028698 threads (0.0% utilization) -2025-07-28 18:27:08,183 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:08,183 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 18:27:08,842 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:27:08,842 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 18:27:08,895 - INFO - root : [END] USER_THREAD_COUNT: mterry has 36/1028698 threads (0.0% utilization) -2025-07-28 18:33:09,996 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 18:33:10,675 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:10,675 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 18:33:11,333 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:33:11,333 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 18:33:11,386 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:39:12,488 - INFO - root : [START] USER_THREAD_COUNT: mterry has 59/1028698 threads (0.01% utilization) -2025-07-28 18:39:13,162 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:39:13,162 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 18:39:13,816 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 18:39:13,816 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 18:39:13,869 - INFO - root : [END] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 18:45:15,066 - INFO - root : [START] USER_THREAD_COUNT: mterry has 47/1028698 threads (0.0% utilization) -2025-07-28 18:45:15,742 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:45:15,743 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 18:45:16,401 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:45:16,401 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 18:45:16,455 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:51:17,556 - INFO - root : [START] USER_THREAD_COUNT: mterry has 47/1028698 threads (0.0% utilization) -2025-07-28 18:51:18,238 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:18,238 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 18:51:18,896 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:51:18,896 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 18:51:18,949 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:57:20,049 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 18:57:20,747 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:20,747 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 18:57:21,404 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:57:21,404 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 18:57:21,456 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 19:03:22,551 - INFO - root : [START] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 19:03:23,226 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:03:23,226 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 19:03:23,880 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:03:23,881 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 19:03:23,932 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:09:25,027 - INFO - root : [START] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 19:09:25,701 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:25,701 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 19:09:26,358 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:26,358 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 19:09:26,410 - INFO - root : [END] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 19:15:27,516 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 19:15:28,190 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:15:28,190 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 19:15:28,843 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:15:28,844 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 19:15:28,897 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:21:30,070 - INFO - root : [START] USER_THREAD_COUNT: mterry has 74/1028698 threads (0.01% utilization) -2025-07-28 19:21:30,743 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:21:30,743 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 19:21:31,426 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:21:31,426 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 19:21:31,478 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:32,667 - INFO - root : [START] USER_THREAD_COUNT: mterry has 84/1028698 threads (0.01% utilization) -2025-07-28 19:27:33,345 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:33,345 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 19:27:34,008 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:34,008 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 19:27:34,060 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:33:35,188 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 19:33:35,864 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:33:35,864 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 19:33:36,517 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:33:36,517 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 19:33:36,569 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:39:37,685 - INFO - root : [START] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 19:39:38,360 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:39:38,360 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 19:39:39,015 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:39:39,015 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 19:39:39,067 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:45:40,168 - INFO - root : [START] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 19:45:40,845 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:45:40,846 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 19:45:41,498 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:45:41,498 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 19:45:41,552 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:51:42,658 - INFO - root : [START] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 19:51:43,336 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:51:43,336 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 19:51:43,994 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:51:43,994 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 19:51:44,049 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:57:45,227 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:57:45,907 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:57:45,907 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 19:57:46,561 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:57:46,561 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 19:57:46,613 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:03:47,734 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:03:48,407 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:03:48,408 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 20:03:49,062 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:03:49,062 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 20:03:49,115 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:09:50,222 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:09:50,896 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:09:50,896 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 20:09:51,550 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:09:51,550 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 20:09:51,601 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:15:52,715 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:15:53,386 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:15:53,386 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 20:15:54,041 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:15:54,041 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 20:15:54,092 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:21:55,197 - INFO - root : [START] USER_THREAD_COUNT: mterry has 25/1028698 threads (0.0% utilization) -2025-07-28 20:21:55,867 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:21:55,867 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 20:21:56,520 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:21:56,520 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 20:21:56,572 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:27:57,681 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:27:58,353 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:27:58,353 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 20:27:59,006 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:27:59,006 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 20:27:59,058 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:34:00,172 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:34:00,849 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:34:00,849 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 20:34:01,512 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:34:01,512 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 20:34:01,567 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:02,730 - INFO - root : [START] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 20:40:03,428 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:03,428 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 20:40:04,150 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:04,150 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 20:40:04,203 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:46:05,357 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:06,032 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:06,032 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 20:46:06,687 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:06,687 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 20:46:06,739 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:07,864 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:08,541 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:08,541 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 20:52:09,247 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:09,247 - INFO - root : Rocoto call successful on attempt 1: call_time=0.65s, total_time=0.65s -2025-07-28 20:52:09,301 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:10,411 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:11,082 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:11,082 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 20:58:11,739 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:11,740 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 20:58:11,791 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:12,910 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:13,595 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:13,595 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 21:04:14,257 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:14,257 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 21:04:14,309 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:15,414 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:16,088 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:16,088 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 21:10:16,744 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:16,744 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 21:10:16,796 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:16:17,900 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:18,575 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:18,575 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 21:16:19,233 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:19,233 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 21:16:19,286 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:22:20,406 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:21,078 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:21,078 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 21:22:21,735 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:21,735 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 21:22:21,787 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:28:22,888 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:28:23,562 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:28:23,563 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 21:28:24,217 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:28:24,218 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 21:28:24,270 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:34:25,399 - INFO - root : [START] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:34:26,072 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:34:26,072 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 21:34:26,728 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:34:26,728 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 21:34:26,781 - INFO - root : [END] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:40:27,955 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:40:28,629 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:40:28,629 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 21:40:29,282 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:40:29,283 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 21:40:29,335 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:46:30,460 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:46:31,131 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:46:31,131 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 21:46:31,785 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:46:31,785 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 21:46:31,837 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:52:32,953 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:52:33,624 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:52:33,624 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 21:52:34,277 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:52:34,277 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 21:52:34,329 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:58:35,464 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:58:36,135 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:58:36,135 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 21:58:36,790 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:58:36,790 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 21:58:36,843 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:04:37,955 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:04:38,629 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:04:38,629 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 22:04:39,283 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:04:39,283 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 22:04:39,335 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:10:40,504 - INFO - root : [START] USER_THREAD_COUNT: mterry has 66/1028698 threads (0.01% utilization) -2025-07-28 22:10:41,181 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:10:41,181 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 22:10:41,837 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:10:41,837 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 22:10:41,890 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:16:43,242 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:16:43,919 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 79/1028698 threads (0.01% utilization) -2025-07-28 22:16:43,919 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 22:16:44,574 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:16:44,574 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 22:16:44,626 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:22:45,753 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:22:46,426 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:22:46,426 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 22:22:47,083 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 66/1028698 threads (0.01% utilization) -2025-07-28 22:22:47,083 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 22:22:47,136 - INFO - root : [END] USER_THREAD_COUNT: mterry has 66/1028698 threads (0.01% utilization) -2025-07-28 22:28:48,261 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:28:48,934 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:28:48,934 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 22:28:49,594 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:28:49,594 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 22:28:49,646 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:34:50,764 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:34:51,437 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:34:51,437 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 22:34:52,090 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:34:52,090 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 22:34:52,141 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:40:53,302 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:40:53,975 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:40:53,975 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 22:40:54,630 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:40:54,631 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 22:40:54,682 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:46:55,879 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:46:56,550 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:46:56,550 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 22:46:57,204 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:46:57,204 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 22:46:57,255 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:52:58,502 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:52:59,177 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:52:59,177 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 22:52:59,830 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:52:59,831 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 22:52:59,883 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:59:01,088 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:59:01,767 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:59:01,768 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 22:59:02,427 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:59:02,427 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 22:59:02,479 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:05:03,658 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:05:04,334 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:05:04,334 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 23:05:04,990 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:05:04,990 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 23:05:05,042 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:06,267 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:06,939 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:06,939 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 23:11:07,596 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:07,596 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 23:11:07,648 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:08,806 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:09,478 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:09,479 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 23:17:10,142 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:10,142 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 23:17:10,193 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:23:11,404 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:23:12,081 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:23:12,082 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 23:23:12,737 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:23:12,737 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 23:23:12,789 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:29:13,974 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:29:14,648 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:29:14,648 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 23:29:15,299 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:29:15,300 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 23:29:15,351 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:35:16,514 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:35:17,188 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:35:17,188 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 23:35:17,845 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:35:17,845 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 23:35:17,896 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:41:19,147 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:41:19,822 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:41:19,822 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 23:41:20,475 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:41:20,475 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 23:41:20,526 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:47:21,670 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:47:22,344 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:47:22,344 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 23:47:22,996 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:47:22,996 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 23:47:23,048 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:53:24,191 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:53:24,862 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:53:24,862 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 23:53:25,516 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:53:25,516 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 23:53:25,567 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:59:26,720 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:59:27,395 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:59:27,395 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 23:59:28,048 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:59:28,048 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-28 23:59:28,099 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:05:30,556 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:05:31,229 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:05:31,229 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-29 00:05:31,881 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:05:31,881 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-29 00:05:31,932 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:11:33,415 - INFO - root : [START] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:11:34,089 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:11:34,090 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-29 00:11:34,748 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:11:34,749 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-29 00:11:34,800 - INFO - root : [END] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:17:35,949 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:17:36,624 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:17:36,624 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-29 00:17:37,280 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:17:37,280 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-29 00:17:37,330 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:23:38,483 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:23:39,155 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:23:39,155 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-29 00:23:39,808 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:23:39,808 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-29 00:23:39,859 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:29:40,999 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:29:41,671 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:29:41,671 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-29 00:29:42,322 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:29:42,322 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-29 00:29:42,372 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:35:43,593 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:35:44,267 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:35:44,267 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-29 00:35:44,920 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:35:44,920 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-29 00:35:44,971 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:41:46,433 - INFO - root : [START] USER_THREAD_COUNT: mterry has 76/1028698 threads (0.01% utilization) -2025-07-29 00:41:47,109 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 76/1028698 threads (0.01% utilization) -2025-07-29 00:41:47,109 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-29 00:41:47,767 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 76/1028698 threads (0.01% utilization) -2025-07-29 00:41:47,768 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-29 00:41:47,820 - INFO - root : [END] USER_THREAD_COUNT: mterry has 76/1028698 threads (0.01% utilization) -2025-07-29 00:47:48,987 - INFO - root : [START] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:47:49,662 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:47:49,662 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-29 00:47:50,317 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:47:50,317 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-29 00:47:50,369 - INFO - root : [END] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:53:51,779 - INFO - root : [START] USER_THREAD_COUNT: mterry has 70/1028698 threads (0.01% utilization) -2025-07-29 00:53:52,482 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 81/1028698 threads (0.01% utilization) -2025-07-29 00:53:52,482 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 00:53:53,144 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 48/1028698 threads (0.0% utilization) -2025-07-29 00:53:53,144 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-29 00:53:53,196 - INFO - root : [END] USER_THREAD_COUNT: mterry has 48/1028698 threads (0.0% utilization) -2025-07-29 00:59:54,335 - INFO - root : [START] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 00:59:55,009 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 00:59:55,009 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-29 00:59:55,664 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 00:59:55,664 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-29 00:59:55,715 - INFO - root : [END] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:05:56,852 - INFO - root : [START] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:05:57,532 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:05:57,532 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-29 01:05:58,186 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:05:58,187 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-29 01:05:58,238 - INFO - root : [END] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:11:59,383 - INFO - root : [START] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:12:00,065 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:12:00,065 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-29 01:12:00,722 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:12:00,722 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-29 01:12:00,773 - INFO - root : [END] USER_THREAD_COUNT: mterry has 57/1028698 threads (0.01% utilization) -2025-07-29 01:18:01,933 - INFO - root : [START] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:18:02,627 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:18:02,627 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-29 01:18:03,298 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:18:03,298 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-29 01:18:03,349 - INFO - root : [END] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:24:04,524 - INFO - root : [START] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:24:05,203 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:24:05,203 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-29 01:24:05,870 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:24:05,870 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-29 01:24:05,941 - INFO - root : [END] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-29 01:30:07,189 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:30:07,879 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:30:07,879 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-29 01:30:08,545 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:30:08,546 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-29 01:30:08,601 - INFO - root : [END] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:36:09,755 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:36:10,435 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:36:10,435 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-29 01:36:11,091 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:36:11,091 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-29 01:36:11,142 - INFO - root : [END] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-29 01:42:12,318 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:42:13,003 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:42:13,003 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-29 01:42:13,660 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:42:13,660 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-29 01:42:13,712 - INFO - root : [END] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:48:14,910 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:48:15,591 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:48:15,592 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-29 01:48:16,247 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:48:16,247 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-29 01:48:16,298 - INFO - root : [END] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:54:17,459 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:54:18,138 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:54:18,139 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-29 01:54:18,792 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 01:54:18,792 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-29 01:54:18,843 - INFO - root : [END] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 02:00:19,990 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-29 02:00:20,673 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-29 02:00:20,674 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-29 02:00:21,332 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-29 02:00:21,332 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-29 02:00:21,385 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-29 02:06:22,847 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 02:06:23,527 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 02:06:23,528 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-29 02:06:24,189 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 02:06:24,189 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-29 02:06:24,241 - INFO - root : [END] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-29 02:12:25,697 - INFO - root : [START] USER_THREAD_COUNT: mterry has 16/1028698 threads (0.0% utilization) -2025-07-29 02:12:26,373 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 16/1028698 threads (0.0% utilization) -2025-07-29 02:12:26,373 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-29 02:12:27,030 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 16/1028698 threads (0.0% utilization) -2025-07-29 02:12:27,030 - INFO - root : Rocoto call successful on attempt 1: call_time=0.60s, total_time=0.60s -2025-07-29 02:12:27,081 - INFO - root : [END] USER_THREAD_COUNT: mterry has 16/1028698 threads (0.0% utilization) diff --git a/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C96mx100_S2S_rocotostat.log b/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C96mx100_S2S_rocotostat.log deleted file mode 100644 index d23ea72f..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/RUNTESTS_EXPDIR_C96mx100_S2S_rocotostat.log +++ /dev/null @@ -1,600 +0,0 @@ -2025-07-28 14:49:35,260 - INFO - root : [START] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 14:49:35,969 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 60/1028698 threads (0.01% utilization) -2025-07-28 14:49:35,969 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 14:49:36,656 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 69/1028698 threads (0.01% utilization) -2025-07-28 14:49:36,656 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 14:49:36,718 - INFO - root : [END] USER_THREAD_COUNT: mterry has 64/1028698 threads (0.01% utilization) -2025-07-28 14:55:37,902 - INFO - root : [START] USER_THREAD_COUNT: mterry has 70/1028698 threads (0.01% utilization) -2025-07-28 14:55:38,598 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 74/1028698 threads (0.01% utilization) -2025-07-28 14:55:38,598 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 14:55:39,293 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 91/1028698 threads (0.01% utilization) -2025-07-28 14:55:39,293 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 14:55:39,353 - INFO - root : [END] USER_THREAD_COUNT: mterry has 104/1028698 threads (0.01% utilization) -2025-07-28 15:01:40,504 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:01:41,208 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 15:01:41,209 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 15:01:41,952 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 33/1028698 threads (0.0% utilization) -2025-07-28 15:01:41,952 - INFO - root : Rocoto call successful on attempt 1: call_time=0.69s, total_time=0.69s -2025-07-28 15:01:42,009 - INFO - root : [END] USER_THREAD_COUNT: mterry has 35/1028698 threads (0.0% utilization) -2025-07-28 15:07:43,189 - INFO - root : [START] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 15:07:43,919 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 15:07:43,919 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 15:07:44,620 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 44/1028698 threads (0.0% utilization) -2025-07-28 15:07:44,620 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 15:07:44,681 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 15:13:45,850 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 15:13:46,542 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 33/1028698 threads (0.0% utilization) -2025-07-28 15:13:46,542 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 15:13:47,223 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 15:13:47,224 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 15:13:47,281 - INFO - root : [END] USER_THREAD_COUNT: mterry has 53/1028698 threads (0.01% utilization) -2025-07-28 15:19:48,691 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:19:49,405 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 15:19:49,405 - INFO - root : Rocoto call successful on attempt 1: call_time=0.65s, total_time=0.65s -2025-07-28 15:19:50,088 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 60/1028698 threads (0.01% utilization) -2025-07-28 15:19:50,088 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 15:19:50,149 - INFO - root : [END] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 15:25:51,315 - INFO - root : [START] USER_THREAD_COUNT: mterry has 46/1028698 threads (0.0% utilization) -2025-07-28 15:25:52,025 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 15:25:52,025 - INFO - root : Rocoto call successful on attempt 1: call_time=0.65s, total_time=0.65s -2025-07-28 15:25:52,745 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 15:25:52,746 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 15:25:52,804 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 15:31:53,928 - INFO - root : [START] USER_THREAD_COUNT: mterry has 63/1028698 threads (0.01% utilization) -2025-07-28 15:31:54,625 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 61/1028698 threads (0.01% utilization) -2025-07-28 15:31:54,625 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 15:31:55,308 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 64/1028698 threads (0.01% utilization) -2025-07-28 15:31:55,309 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 15:31:55,364 - INFO - root : [END] USER_THREAD_COUNT: mterry has 66/1028698 threads (0.01% utilization) -2025-07-28 15:37:56,478 - INFO - root : [START] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 15:37:57,175 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:37:57,175 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 15:37:57,850 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 34/1028698 threads (0.0% utilization) -2025-07-28 15:37:57,850 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 15:37:57,905 - INFO - root : [END] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:43:59,114 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:43:59,816 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:43:59,816 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 15:44:00,494 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 33/1028698 threads (0.0% utilization) -2025-07-28 15:44:00,494 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 15:44:00,552 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 15:50:01,727 - INFO - root : [START] USER_THREAD_COUNT: mterry has 60/1028698 threads (0.01% utilization) -2025-07-28 15:50:02,508 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 58/1028698 threads (0.01% utilization) -2025-07-28 15:50:02,508 - INFO - root : Rocoto call successful on attempt 1: call_time=0.71s, total_time=0.71s -2025-07-28 15:50:03,241 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 15:50:03,241 - INFO - root : Rocoto call successful on attempt 1: call_time=0.67s, total_time=0.67s -2025-07-28 15:50:03,305 - INFO - root : [END] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 15:56:04,469 - INFO - root : [START] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:56:05,185 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 34/1028698 threads (0.0% utilization) -2025-07-28 15:56:05,185 - INFO - root : Rocoto call successful on attempt 1: call_time=0.65s, total_time=0.65s -2025-07-28 15:56:05,874 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 15:56:05,874 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 15:56:05,933 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:02:07,087 - INFO - root : [START] USER_THREAD_COUNT: mterry has 47/1028698 threads (0.0% utilization) -2025-07-28 16:02:07,786 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 47/1028698 threads (0.0% utilization) -2025-07-28 16:02:07,787 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 16:02:08,474 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 16:02:08,475 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 16:02:08,532 - INFO - root : [END] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 16:08:09,784 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:08:10,486 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:08:10,487 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 16:08:11,170 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:08:11,170 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 16:08:11,227 - INFO - root : [END] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 16:14:12,379 - INFO - root : [START] USER_THREAD_COUNT: mterry has 49/1028698 threads (0.0% utilization) -2025-07-28 16:14:13,082 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 16:14:13,082 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 16:14:13,799 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:14:13,799 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 16:14:13,854 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:20:14,982 - INFO - root : [START] USER_THREAD_COUNT: mterry has 45/1028698 threads (0.0% utilization) -2025-07-28 16:20:15,679 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 45/1028698 threads (0.0% utilization) -2025-07-28 16:20:15,680 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 16:20:16,356 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 16:20:16,356 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 16:20:16,410 - INFO - root : [END] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 16:26:17,540 - INFO - root : [START] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 16:26:18,253 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 16:26:18,254 - INFO - root : Rocoto call successful on attempt 1: call_time=0.65s, total_time=0.65s -2025-07-28 16:26:18,924 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:26:18,925 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 16:26:18,976 - INFO - root : [END] USER_THREAD_COUNT: mterry has 31/1028698 threads (0.0% utilization) -2025-07-28 16:32:20,109 - INFO - root : [START] USER_THREAD_COUNT: mterry has 47/1028698 threads (0.0% utilization) -2025-07-28 16:32:20,806 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 16:32:20,807 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 16:32:21,479 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:32:21,479 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 16:32:21,533 - INFO - root : [END] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 16:38:22,665 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 16:38:23,358 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 16:38:23,358 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 16:38:24,026 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:38:24,026 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 16:38:24,079 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:25,193 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 16:44:25,900 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 16:44:25,900 - INFO - root : Rocoto call successful on attempt 1: call_time=0.65s, total_time=0.65s -2025-07-28 16:44:26,592 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:44:26,592 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 16:44:26,646 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:27,789 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 16:50:28,480 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 16:50:28,480 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 16:50:29,151 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:50:29,152 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 16:50:29,205 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:30,334 - INFO - root : [START] USER_THREAD_COUNT: mterry has 56/1028698 threads (0.01% utilization) -2025-07-28 16:56:31,026 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 16:56:31,026 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 16:56:31,694 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 16:56:31,694 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 16:56:31,746 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:32,878 - INFO - root : [START] USER_THREAD_COUNT: mterry has 63/1028698 threads (0.01% utilization) -2025-07-28 17:02:33,570 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 17:02:33,570 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 17:02:34,241 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:02:34,241 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 17:02:34,293 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:35,420 - INFO - root : [START] USER_THREAD_COUNT: mterry has 63/1028698 threads (0.01% utilization) -2025-07-28 17:08:36,111 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 43/1028698 threads (0.0% utilization) -2025-07-28 17:08:36,111 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 17:08:36,780 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:08:36,780 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 17:08:36,832 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:37,968 - INFO - root : [START] USER_THREAD_COUNT: mterry has 59/1028698 threads (0.01% utilization) -2025-07-28 17:14:38,659 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:14:38,659 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 17:14:39,329 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:14:39,329 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 17:14:39,382 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:40,500 - INFO - root : [START] USER_THREAD_COUNT: mterry has 30/1028698 threads (0.0% utilization) -2025-07-28 17:20:41,223 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:20:41,223 - INFO - root : Rocoto call successful on attempt 1: call_time=0.66s, total_time=0.66s -2025-07-28 17:20:41,891 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:20:41,891 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 17:20:41,944 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:43,066 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:26:43,756 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 17:26:43,756 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 17:26:44,432 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:26:44,432 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 17:26:44,483 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:32:45,598 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:32:46,286 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:32:46,286 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 17:32:46,951 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:32:46,952 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 17:32:47,003 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:38:48,179 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:38:48,863 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:38:48,863 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 17:38:49,529 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:38:49,529 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 17:38:49,581 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:44:50,695 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:44:51,383 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:44:51,383 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 17:44:52,047 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:44:52,048 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 17:44:52,100 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:50:53,213 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 17:50:53,898 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:50:53,899 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 17:50:54,567 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:50:54,567 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 17:50:54,620 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 17:56:55,738 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 17:56:56,423 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 17:56:56,423 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 17:56:57,088 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 17:56:57,088 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 17:56:57,142 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:02:58,283 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 18:02:58,972 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:02:58,972 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 18:02:59,635 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:02:59,635 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 18:02:59,687 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:00,802 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:09:01,491 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:01,491 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 18:09:02,171 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:09:02,172 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 18:09:02,228 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:03,393 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:15:04,103 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:04,103 - INFO - root : Rocoto call successful on attempt 1: call_time=0.65s, total_time=0.65s -2025-07-28 18:15:04,785 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:15:04,785 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 18:15:04,839 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:05,962 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:21:06,651 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:06,651 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 18:21:07,317 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:21:07,317 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 18:21:07,369 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:08,482 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:27:09,171 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:09,171 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 18:27:09,844 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:27:09,844 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 18:27:09,898 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:11,018 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:33:11,704 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:11,704 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 18:33:12,376 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:33:12,376 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 18:33:12,430 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:39:13,541 - INFO - root : [START] USER_THREAD_COUNT: mterry has 42/1028698 threads (0.0% utilization) -2025-07-28 18:39:14,228 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:39:14,228 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 18:39:14,893 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:39:14,893 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 18:39:14,945 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 18:45:16,061 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:45:16,748 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:45:16,748 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 18:45:17,421 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:45:17,421 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 18:45:17,473 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:18,592 - INFO - root : [START] USER_THREAD_COUNT: mterry has 29/1028698 threads (0.0% utilization) -2025-07-28 18:51:19,280 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:19,281 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 18:51:19,946 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:51:19,946 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 18:51:19,997 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:21,152 - INFO - root : [START] USER_THREAD_COUNT: mterry has 36/1028698 threads (0.0% utilization) -2025-07-28 18:57:21,838 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:21,838 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 18:57:22,503 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 18:57:22,503 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 18:57:22,555 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:03:23,709 - INFO - root : [START] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-28 19:03:24,396 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:03:24,396 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 19:03:25,063 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:03:25,063 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 19:03:25,115 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:26,333 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:09:27,019 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:27,019 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 19:09:27,684 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:09:27,684 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 19:09:27,735 - INFO - root : [END] USER_THREAD_COUNT: mterry has 40/1028698 threads (0.0% utilization) -2025-07-28 19:15:28,954 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 19:15:29,641 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:15:29,641 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 19:15:30,309 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:15:30,309 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 19:15:30,360 - INFO - root : [END] USER_THREAD_COUNT: mterry has 27/1028698 threads (0.0% utilization) -2025-07-28 19:21:31,782 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:21:32,474 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:21:32,475 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 19:21:33,144 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:21:33,144 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 19:21:33,197 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:34,387 - INFO - root : [START] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:35,075 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:35,075 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 19:27:35,741 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:27:35,741 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 19:27:35,794 - INFO - root : [END] USER_THREAD_COUNT: mterry has 54/1028698 threads (0.01% utilization) -2025-07-28 19:33:36,920 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:33:37,610 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 32/1028698 threads (0.0% utilization) -2025-07-28 19:33:37,610 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 19:33:38,276 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:33:38,276 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 19:33:38,327 - INFO - root : [END] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 19:39:39,458 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:39:40,144 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:39:40,144 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 19:39:40,812 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:39:40,813 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 19:39:40,865 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:45:42,005 - INFO - root : [START] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:45:42,693 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:45:42,693 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 19:45:43,360 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:45:43,360 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 19:45:43,413 - INFO - root : [END] USER_THREAD_COUNT: mterry has 41/1028698 threads (0.0% utilization) -2025-07-28 19:51:44,545 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:51:45,235 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:51:45,235 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 19:51:45,903 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:51:45,903 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 19:51:45,955 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:57:47,077 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:57:47,764 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:57:47,764 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 19:57:48,437 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 19:57:48,437 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 19:57:48,489 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:03:49,625 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:03:50,312 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:03:50,313 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 20:03:50,980 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:03:50,980 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 20:03:51,032 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:09:52,165 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:09:52,852 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:09:52,852 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 20:09:53,519 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:09:53,519 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 20:09:53,570 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:15:54,699 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:15:55,387 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:15:55,387 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 20:15:56,056 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:15:56,056 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 20:15:56,107 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:21:57,236 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:21:57,930 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:21:57,930 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 20:21:58,602 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:21:58,602 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 20:21:58,653 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:27:59,796 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:28:00,482 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:28:00,483 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 20:28:01,148 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:28:01,149 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 20:28:01,201 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 20:34:02,353 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:34:03,044 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:34:03,044 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 20:34:03,715 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:34:03,715 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 20:34:03,767 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:04,903 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:05,593 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:05,593 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 20:40:06,281 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:40:06,281 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 20:40:06,339 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 20:46:07,486 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:08,175 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:08,175 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 20:46:08,844 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:46:08,844 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 20:46:08,896 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:10,027 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:10,721 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:10,721 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 20:52:11,389 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:52:11,390 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 20:52:11,442 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:12,582 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:13,271 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:13,271 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 20:58:13,939 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 20:58:13,939 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 20:58:13,991 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:15,152 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:15,840 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:15,840 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 21:04:16,505 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:04:16,505 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 21:04:16,557 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:17,723 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:18,412 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:18,412 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 21:10:19,081 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:10:19,081 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 21:10:19,133 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:16:20,273 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:20,961 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:20,961 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 21:16:21,626 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:16:21,626 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 21:16:21,679 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:22:22,843 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:23,529 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:23,529 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 21:22:24,194 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:22:24,195 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 21:22:24,246 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:28:25,507 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:28:26,198 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:28:26,198 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 21:28:26,866 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:28:26,866 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 21:28:26,918 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 21:34:28,071 - INFO - root : [START] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:34:28,760 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:34:28,760 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 21:34:29,427 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:34:29,427 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 21:34:29,479 - INFO - root : [END] USER_THREAD_COUNT: mterry has 78/1028698 threads (0.01% utilization) -2025-07-28 21:40:30,670 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:40:31,359 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:40:31,359 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 21:40:32,025 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:40:32,025 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 21:40:32,078 - INFO - root : [END] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 21:46:33,257 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:46:33,945 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:46:33,945 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 21:46:34,611 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:46:34,612 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 21:46:34,663 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 21:52:35,950 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:52:36,640 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:52:36,641 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 21:52:37,305 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:52:37,305 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 21:52:37,357 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:58:38,555 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:58:39,254 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:58:39,254 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 21:58:39,920 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 21:58:39,920 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 21:58:39,972 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:04:41,203 - INFO - root : [START] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:04:41,894 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:04:41,894 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 22:04:42,560 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:04:42,560 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 22:04:42,612 - INFO - root : [END] USER_THREAD_COUNT: mterry has 26/1028698 threads (0.0% utilization) -2025-07-28 22:10:44,096 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:10:44,784 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:10:44,784 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 22:10:45,454 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:10:45,454 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 22:10:45,506 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:16:46,654 - INFO - root : [START] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:16:47,345 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:16:47,345 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 22:16:48,018 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:16:48,018 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 22:16:48,072 - INFO - root : [END] USER_THREAD_COUNT: mterry has 65/1028698 threads (0.01% utilization) -2025-07-28 22:22:49,208 - INFO - root : [START] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:22:49,897 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:22:49,897 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 22:22:50,566 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:22:50,566 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 22:22:50,619 - INFO - root : [END] USER_THREAD_COUNT: mterry has 52/1028698 threads (0.01% utilization) -2025-07-28 22:28:52,196 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:28:52,884 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:28:52,885 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 22:28:53,552 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:28:53,552 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 22:28:53,604 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:34:54,721 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:34:55,410 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:34:55,410 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 22:34:56,103 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:34:56,103 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 22:34:56,155 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:40:57,274 - INFO - root : [START] USER_THREAD_COUNT: mterry has 39/1028698 threads (0.0% utilization) -2025-07-28 22:40:57,965 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:40:57,965 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 22:40:58,632 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:40:58,633 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 22:40:58,684 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:46:59,812 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:47:00,499 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 22:47:00,499 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 22:47:01,168 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:47:01,168 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 22:47:01,220 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:53:02,360 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:53:03,071 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:53:03,071 - INFO - root : Rocoto call successful on attempt 1: call_time=0.65s, total_time=0.65s -2025-07-28 22:53:03,744 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 22:53:03,744 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 22:53:03,796 - INFO - root : [END] USER_THREAD_COUNT: mterry has 38/1028698 threads (0.0% utilization) -2025-07-28 22:59:04,925 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:59:05,611 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:59:05,612 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 22:59:06,278 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 22:59:06,278 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 22:59:06,329 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:05:07,460 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:05:08,157 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:05:08,157 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 23:05:08,824 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:05:08,824 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 23:05:08,877 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:10,010 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:10,697 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:10,697 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 23:11:11,362 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:11:11,362 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 23:11:11,413 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:12,565 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:13,254 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:13,255 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 23:17:13,920 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:17:13,920 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 23:17:13,970 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:23:15,191 - INFO - root : [START] USER_THREAD_COUNT: mterry has 28/1028698 threads (0.0% utilization) -2025-07-28 23:23:15,885 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:23:15,885 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-28 23:23:16,552 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:23:16,552 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 23:23:16,603 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:29:17,847 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:29:18,533 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:29:18,534 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 23:29:19,199 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:29:19,199 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 23:29:19,250 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:35:20,407 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:35:21,098 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:35:21,098 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 23:35:21,762 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:35:21,762 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 23:35:21,813 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-28 23:41:22,930 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:41:23,616 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:41:23,616 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 23:41:24,282 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:41:24,282 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 23:41:24,334 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:47:25,445 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:47:26,131 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:47:26,131 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 23:47:26,804 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:47:26,804 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 23:47:26,854 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:53:27,971 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:53:28,658 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:53:28,658 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 23:53:29,322 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:53:29,322 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-28 23:53:29,373 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:59:30,496 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:59:31,183 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:59:31,183 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-28 23:59:31,855 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-28 23:59:31,855 - INFO - root : Rocoto call successful on attempt 1: call_time=0.62s, total_time=0.62s -2025-07-28 23:59:31,906 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:05:33,021 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:05:33,713 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:05:33,713 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 00:05:34,380 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:05:34,380 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-29 00:05:34,431 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:11:35,550 - INFO - root : [START] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:11:36,238 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:11:36,238 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-29 00:11:36,905 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:11:36,905 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-29 00:11:36,957 - INFO - root : [END] USER_THREAD_COUNT: mterry has 50/1028698 threads (0.0% utilization) -2025-07-29 00:17:38,070 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:17:38,756 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:17:38,756 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-29 00:17:39,424 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:17:39,425 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-29 00:17:39,475 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:23:40,610 - INFO - root : [START] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:23:41,302 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:23:41,302 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 00:23:41,966 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:23:41,966 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-29 00:23:42,016 - INFO - root : [END] USER_THREAD_COUNT: mterry has 37/1028698 threads (0.0% utilization) -2025-07-29 00:29:43,207 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:29:43,893 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:29:43,893 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-29 00:29:44,555 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:29:44,556 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-29 00:29:44,606 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:35:45,716 - INFO - root : [START] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:35:46,407 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:35:46,407 - INFO - root : Rocoto call successful on attempt 1: call_time=0.64s, total_time=0.64s -2025-07-29 00:35:47,073 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:35:47,073 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-29 00:35:47,125 - INFO - root : [END] USER_THREAD_COUNT: mterry has 24/1028698 threads (0.0% utilization) -2025-07-29 00:41:48,257 - INFO - root : [START] USER_THREAD_COUNT: mterry has 76/1028698 threads (0.01% utilization) -2025-07-29 00:41:48,946 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 76/1028698 threads (0.01% utilization) -2025-07-29 00:41:48,946 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-29 00:41:49,613 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 76/1028698 threads (0.01% utilization) -2025-07-29 00:41:49,613 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-29 00:41:49,665 - INFO - root : [END] USER_THREAD_COUNT: mterry has 76/1028698 threads (0.01% utilization) -2025-07-29 00:47:51,079 - INFO - root : [START] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:47:51,766 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:47:51,766 - INFO - root : Rocoto call successful on attempt 1: call_time=0.63s, total_time=0.63s -2025-07-29 00:47:52,433 - INFO - root : [ROCOTO_SUCCESS_ATTEMPT_1] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) -2025-07-29 00:47:52,433 - INFO - root : Rocoto call successful on attempt 1: call_time=0.61s, total_time=0.61s -2025-07-29 00:47:52,485 - INFO - root : [END] USER_THREAD_COUNT: mterry has 89/1028698 threads (0.01% utilization) diff --git a/ci/error_logs/HECULES_PR_LOGS/global_mterry_RUNTESTS_ci-run_check.log b/ci/error_logs/HECULES_PR_LOGS/global_mterry_RUNTESTS_ci-run_check.log deleted file mode 100644 index 2a1c50ee..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/global_mterry_RUNTESTS_ci-run_check.log +++ /dev/null @@ -1,7 +0,0 @@ -Experiment C48_ATM Completed 1 Cycles: *SUCCESS* at Mon Jul 28 19:51:42 CDT 2025 -Experiment C48_S2SW Completed 1 Cycles: *SUCCESS* at Mon Jul 28 22:23:25 CDT 2025 -Experiment C96mx100_S2S Completed 1 Cycles: *SUCCESS* at Tue Jul 29 00:47:52 CDT 2025 -Experiment C48_S2SWA_gefs Completed 1 Cycles: *SUCCESS* at Tue Jul 29 00:54:54 CDT 2025 -Experiment C48mx500_hybAOWCDA Completed 2 Cycles: *SUCCESS* at Tue Jul 29 00:59:52 CDT 2025 -Experiment C48mx500_3DVarAOWCDA Completed 2 Cycles: *SUCCESS* at Tue Jul 29 01:37:06 CDT 2025 -Experiment C96_atm3DVar Completed 3 Cycles: *SUCCESS* at Tue Jul 29 02:12:27 CDT 2025 diff --git a/ci/error_logs/HECULES_PR_LOGS/logs_2021032312.log b/ci/error_logs/HECULES_PR_LOGS/logs_2021032312.log deleted file mode 100644 index 2b7201ec..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/logs_2021032312.log +++ /dev/null @@ -1,2417 +0,0 @@ -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_stage_ic -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_waveinit -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_stage_ic succeeded, jobid=5950267 -2025-07-28 14:48:24 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_waveinit succeeded, jobid=5950268 -2025-07-28 14:48:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_stage_ic, jobid=5950267, in state RUNNING (RUNNING) -2025-07-28 14:48:35 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_waveinit, jobid=5950268, in state QUEUED (PENDING) -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_stage_ic, jobid=5950267, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_waveinit, jobid=5950268, in state SUCCEEDED (COMPLETED), ran for 35.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_fcst_seg0 -2025-07-28 14:54:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_fcst_seg0 succeeded, jobid=5950292 -2025-07-28 15:00:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 15:06:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 15:12:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 15:18:49 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 15:24:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 15:30:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 15:36:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 15:43:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 15:49:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 15:55:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 16:01:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 16:07:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 16:13:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 16:19:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 16:25:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 16:31:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 16:37:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 16:43:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 16:49:31 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 16:55:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 17:01:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 17:07:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 17:13:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 17:19:44 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 17:25:47 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 17:31:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 17:37:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 17:43:55 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 17:49:58 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 17:56:01 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 18:02:04 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 18:08:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 18:14:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 18:20:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 18:26:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 18:32:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 18:38:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 18:44:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 18:50:26 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state QUEUED (PENDING) -2025-07-28 18:56:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state RUNNING (RUNNING) -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state RUNNING (RUNNING) -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f000-f002 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f003-f005 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f006-f008 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f009-f011 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f012-f014 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f015-f017 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f018-f020 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f006 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f006 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f012 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f018 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f000-f002 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f003-f005 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f006-f008 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f009-f011 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f012-f014 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f015-f017 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f018-f020 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f021-f023 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f024-f026 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f027-f029 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f030-f032 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f000-f002 succeeded, jobid=5950995 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f003-f005 succeeded, jobid=5950996 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f006-f008 succeeded, jobid=5950997 -2025-07-28 19:02:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f009-f011 succeeded, jobid=5950998 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f012-f014 succeeded, jobid=5950999 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f015-f017 succeeded, jobid=5951000 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f018-f020 succeeded, jobid=5951001 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f006 succeeded, jobid=5951002 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f006 succeeded, jobid=5951003 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f012 succeeded, jobid=5951004 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f018 succeeded, jobid=5951005 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f000-f002 succeeded, jobid=5951006 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f003-f005 succeeded, jobid=5951007 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f006-f008 succeeded, jobid=5951008 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f009-f011 succeeded, jobid=5951009 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f012-f014 succeeded, jobid=5951010 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f015-f017 succeeded, jobid=5951011 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f018-f020 succeeded, jobid=5951012 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f021-f023 succeeded, jobid=5951013 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f024-f026 succeeded, jobid=5951014 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f027-f029 succeeded, jobid=5951015 -2025-07-28 19:02:37 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f030-f032 succeeded, jobid=5951016 -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state RUNNING (RUNNING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f006, jobid=5951002, in state SUCCEEDED (COMPLETED), ran for 46.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f006, jobid=5951003, in state SUCCEEDED (COMPLETED), ran for 46.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f012, jobid=5951004, in state SUCCEEDED (COMPLETED), ran for 46.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f018, jobid=5951005, in state SUCCEEDED (COMPLETED), ran for 46.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f021-f023 -2025-07-28 19:08:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f024-f026 -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f027-f029 -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f030-f032 -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f033-f035 -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f036-f038 -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f021-f023 succeeded, jobid=5951060 -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f024-f026 succeeded, jobid=5951061 -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f027-f029 succeeded, jobid=5951062 -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f030-f032 succeeded, jobid=5951063 -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f033-f035 succeeded, jobid=5951064 -2025-07-28 19:08:40 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f036-f038 succeeded, jobid=5951065 -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state RUNNING (RUNNING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:14:43 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_fcst_seg0, jobid=5950292, in state SUCCEEDED (COMPLETED), ran for 1488.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 19:20:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f039-f041 -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:20:46 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f039-f041 succeeded, jobid=5951115 -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 19:26:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 19:32:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 19:38:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 19:45:02 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 19:51:05 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 19:57:08 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 20:03:11 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 20:09:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 20:15:22 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 20:21:25 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 20:27:28 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 20:33:30 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 20:39:33 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state RUNNING (RUNNING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 20:45:36 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f000-f002, jobid=5950995, in state SUCCEEDED (COMPLETED), ran for 267.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f042-f044 -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:51:39 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f042-f044 succeeded, jobid=5951482 -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5951482, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 20:57:42 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5951482, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 21:03:45 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f003-f005, jobid=5950996, in state SUCCEEDED (COMPLETED), ran for 284.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f006-f008, jobid=5950997, in state SUCCEEDED (COMPLETED), ran for 280.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5951482, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f045-f047 -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f048-f054 -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f045-f047 succeeded, jobid=5951544 -2025-07-28 21:09:48 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f048-f054 succeeded, jobid=5951545 -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state RUNNING (RUNNING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5951482, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5951544, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5951545, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state QUEUED (PENDING) -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_atmos_prod_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f012, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f018, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:50 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 21:15:51 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f009-f011, jobid=5950998, in state SUCCEEDED (COMPLETED), ran for 274.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state RUNNING (RUNNING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state RUNNING (RUNNING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state RUNNING (RUNNING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state QUEUED (PENDING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state QUEUED (PENDING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state QUEUED (PENDING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state QUEUED (PENDING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state QUEUED (PENDING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state QUEUED (PENDING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state QUEUED (PENDING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5951482, in state QUEUED (PENDING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5951544, in state QUEUED (PENDING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5951545, in state QUEUED (PENDING) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f000-f002, jobid=5951006, in state SUCCEEDED (COMPLETED), ran for 70.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f003-f005, jobid=5951007, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f006-f008, jobid=5951008, in state SUCCEEDED (COMPLETED), ran for 76.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f009-f011, jobid=5951009, in state SUCCEEDED (COMPLETED), ran for 75.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f012-f014, jobid=5951010, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f015-f017, jobid=5951011, in state SUCCEEDED (COMPLETED), ran for 74.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f018-f020, jobid=5951012, in state SUCCEEDED (COMPLETED), ran for 74.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f021-f023, jobid=5951013, in state SUCCEEDED (COMPLETED), ran for 75.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f024-f026, jobid=5951014, in state SUCCEEDED (COMPLETED), ran for 75.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f027-f029, jobid=5951015, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f030-f032, jobid=5951016, in state SUCCEEDED (COMPLETED), ran for 74.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f057-f063 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f066-f072 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f075-f081 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f084-f090 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f093-f099 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f102-f108 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f111-f114 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_atmos_prod_f117-f120 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f012 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f018 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f024 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f030 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f057-f063 succeeded, jobid=5951616 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f066-f072 succeeded, jobid=5951617 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f075-f081 succeeded, jobid=5951618 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f084-f090 succeeded, jobid=5951619 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f093-f099 succeeded, jobid=5951620 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f102-f108 succeeded, jobid=5951621 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f111-f114 succeeded, jobid=5951622 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_atmos_prod_f117-f120 succeeded, jobid=5951623 -2025-07-28 21:21:53 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f012 succeeded, jobid=5951624 -2025-07-28 21:21:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f018 succeeded, jobid=5951625 -2025-07-28 21:21:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f024 succeeded, jobid=5951626 -2025-07-28 21:21:54 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f030 succeeded, jobid=5951627 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f012-f014, jobid=5950999, in state SUCCEEDED (COMPLETED), ran for 273.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f015-f017, jobid=5951000, in state SUCCEEDED (COMPLETED), ran for 272.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f018-f020, jobid=5951001, in state SUCCEEDED (COMPLETED), ran for 268.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5951482, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5951544, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5951545, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5951616, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5951617, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5951618, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5951619, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5951620, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5951621, in state RUNNING (RUNNING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f111-f114, jobid=5951622, in state SUCCEEDED (COMPLETED), ran for 216.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5951623, in state RUNNING (COMPLETING) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f012, jobid=5951624, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f018, jobid=5951625, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f024, jobid=5951626, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f030, jobid=5951627, in state SUCCEEDED (COMPLETED), ran for 40.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f036 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f042 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f048 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f054 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f060 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f066 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f072 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f078 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ocean_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f024, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f030, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f036, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f042, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f048, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f060, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f066, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f078, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f084, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f096, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f102, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f036 succeeded, jobid=5951670 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f042 succeeded, jobid=5951671 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f048 succeeded, jobid=5951672 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f054 succeeded, jobid=5951673 -2025-07-28 21:27:56 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f060 succeeded, jobid=5951674 -2025-07-28 21:27:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f066 succeeded, jobid=5951675 -2025-07-28 21:27:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f072 succeeded, jobid=5951676 -2025-07-28 21:27:57 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f078 succeeded, jobid=5951677 -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f021-f023, jobid=5951060, in state SUCCEEDED (COMPLETED), ran for 276.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f024-f026, jobid=5951061, in state SUCCEEDED (COMPLETED), ran for 276.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f027-f029, jobid=5951062, in state SUCCEEDED (COMPLETED), ran for 273.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f030-f032, jobid=5951063, in state SUCCEEDED (COMPLETED), ran for 280.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f033-f035, jobid=5951064, in state SUCCEEDED (COMPLETED), ran for 281.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f036-f038, jobid=5951065, in state SUCCEEDED (COMPLETED), ran for 274.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f039-f041, jobid=5951115, in state SUCCEEDED (COMPLETED), ran for 279.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f042-f044, jobid=5951482, in state SUCCEEDED (COMPLETED), ran for 275.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f045-f047, jobid=5951544, in state SUCCEEDED (COMPLETED), ran for 275.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f048-f054, jobid=5951545, in state SUCCEEDED (COMPLETED), ran for 321.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f057-f063, jobid=5951616, in state SUCCEEDED (COMPLETED), ran for 286.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f066-f072, jobid=5951617, in state SUCCEEDED (COMPLETED), ran for 320.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f075-f081, jobid=5951618, in state SUCCEEDED (COMPLETED), ran for 318.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f084-f090, jobid=5951619, in state SUCCEEDED (COMPLETED), ran for 324.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f093-f099, jobid=5951620, in state SUCCEEDED (COMPLETED), ran for 325.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f102-f108, jobid=5951621, in state SUCCEEDED (COMPLETED), ran for 330.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_atmos_prod_f117-f120, jobid=5951623, in state SUCCEEDED (COMPLETED), ran for 220.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f036, jobid=5951670, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f042, jobid=5951671, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f048, jobid=5951672, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f054, jobid=5951673, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f060, jobid=5951674, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f066, jobid=5951675, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f072, jobid=5951676, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f078, jobid=5951677, in state SUCCEEDED (COMPLETED), ran for 36.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_tracker -2025-07-28 21:33:59 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_genesis -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f084 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f090 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f096 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f102 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f108 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f114 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ocean_prod_f120 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f024 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f030 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f036 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f042 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f048 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f054 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f060 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f066 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f072 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f078 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f084 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f090 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f096 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f102 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f108 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f114 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_ice_prod_f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f033-f035, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f036-f038, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f039-f041, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f042-f044, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f045-f047, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f048-f054, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f057-f063, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f066-f072, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f075-f081, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f084-f090, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f093-f099, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f102-f108, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f111-f114, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostsbs_f117-f120, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostbndpntbll, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Cannot submit gfs_wavepostpnt, because maximum global task throttle of 25 will be violated. -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_tracker succeeded, jobid=5951717 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_genesis succeeded, jobid=5951718 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f084 succeeded, jobid=5951719 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f090 succeeded, jobid=5951720 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f096 succeeded, jobid=5951721 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f102 succeeded, jobid=5951722 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f108 succeeded, jobid=5951723 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f114 succeeded, jobid=5951724 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ocean_prod_f120 succeeded, jobid=5951725 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f024 succeeded, jobid=5951726 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f030 succeeded, jobid=5951727 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f036 succeeded, jobid=5951728 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f042 succeeded, jobid=5951729 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f048 succeeded, jobid=5951730 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f054 succeeded, jobid=5951731 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f060 succeeded, jobid=5951732 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f066 succeeded, jobid=5951733 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f072 succeeded, jobid=5951734 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f078 succeeded, jobid=5951735 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f084 succeeded, jobid=5951736 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f090 succeeded, jobid=5951737 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f096 succeeded, jobid=5951738 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f102 succeeded, jobid=5951739 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f108 succeeded, jobid=5951740 -2025-07-28 21:34:00 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f114 succeeded, jobid=5951741 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_tracker, jobid=5951717, in state SUCCEEDED (COMPLETED), ran for 64.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_genesis, jobid=5951718, in state RUNNING (RUNNING) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f084, jobid=5951719, in state SUCCEEDED (COMPLETED), ran for 19.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f090, jobid=5951720, in state SUCCEEDED (COMPLETED), ran for 20.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f096, jobid=5951721, in state SUCCEEDED (COMPLETED), ran for 20.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f102, jobid=5951722, in state SUCCEEDED (COMPLETED), ran for 20.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f108, jobid=5951723, in state SUCCEEDED (COMPLETED), ran for 20.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f114, jobid=5951724, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ocean_prod_f120, jobid=5951725, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f024, jobid=5951726, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f030, jobid=5951727, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f036, jobid=5951728, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f042, jobid=5951729, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f048, jobid=5951730, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f054, jobid=5951731, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f060, jobid=5951732, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f066, jobid=5951733, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f072, jobid=5951734, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f078, jobid=5951735, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f084, jobid=5951736, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f090, jobid=5951737, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f096, jobid=5951738, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f102, jobid=5951739, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f108, jobid=5951740, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f114, jobid=5951741, in state SUCCEEDED (COMPLETED), ran for 38.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_ice_prod_f120 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f033-f035 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f036-f038 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f039-f041 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f042-f044 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f045-f047 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f048-f054 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f057-f063 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f066-f072 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f075-f081 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f084-f090 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f093-f099 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f102-f108 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f111-f114 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostsbs_f117-f120 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostbndpnt -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostbndpntbll -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_wavepostpnt -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_ice_prod_f120 succeeded, jobid=5951787 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f033-f035 succeeded, jobid=5951788 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f036-f038 succeeded, jobid=5951789 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f039-f041 succeeded, jobid=5951790 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f042-f044 succeeded, jobid=5951791 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f045-f047 succeeded, jobid=5951792 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f048-f054 succeeded, jobid=5951793 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f057-f063 succeeded, jobid=5951794 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f066-f072 succeeded, jobid=5951795 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f075-f081 succeeded, jobid=5951796 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f084-f090 succeeded, jobid=5951797 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f093-f099 succeeded, jobid=5951798 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f102-f108 succeeded, jobid=5951799 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f111-f114 succeeded, jobid=5951800 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostsbs_f117-f120 succeeded, jobid=5951801 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostbndpnt succeeded, jobid=5951802 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostbndpntbll succeeded, jobid=5951803 -2025-07-28 21:40:03 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission of gfs_wavepostpnt succeeded, jobid=5951804 -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_genesis, jobid=5951718, in state SUCCEEDED (COMPLETED), ran for 465.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_ice_prod_f120, jobid=5951787, in state SUCCEEDED (COMPLETED), ran for 18.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f033-f035, jobid=5951788, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f036-f038, jobid=5951789, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f039-f041, jobid=5951790, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f042-f044, jobid=5951791, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f045-f047, jobid=5951792, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f048-f054, jobid=5951793, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f057-f063, jobid=5951794, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f066-f072, jobid=5951795, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f075-f081, jobid=5951796, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f084-f090, jobid=5951797, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f093-f099, jobid=5951798, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f102-f108, jobid=5951799, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f111-f114, jobid=5951800, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f117-f120, jobid=5951801, in state QUEUED (PENDING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostbndpnt, jobid=5951802, in state RUNNING (RUNNING) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostbndpntbll, jobid=5951803, in state SUCCEEDED (COMPLETED), ran for 78.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:46:06 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostpnt, jobid=5951804, in state RUNNING (RUNNING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f033-f035, jobid=5951788, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f036-f038, jobid=5951789, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f039-f041, jobid=5951790, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f042-f044, jobid=5951791, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f045-f047, jobid=5951792, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f048-f054, jobid=5951793, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f057-f063, jobid=5951794, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f066-f072, jobid=5951795, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f075-f081, jobid=5951796, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f084-f090, jobid=5951797, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f093-f099, jobid=5951798, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f102-f108, jobid=5951799, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f111-f114, jobid=5951800, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f117-f120, jobid=5951801, in state QUEUED (PENDING) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostbndpnt, jobid=5951802, in state SUCCEEDED (COMPLETED), ran for 175.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:52:09 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostpnt, jobid=5951804, in state SUCCEEDED (COMPLETED), ran for 111.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f033-f035, jobid=5951788, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f036-f038, jobid=5951789, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f039-f041, jobid=5951790, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f042-f044, jobid=5951791, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f045-f047, jobid=5951792, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f048-f054, jobid=5951793, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f057-f063, jobid=5951794, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f066-f072, jobid=5951795, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f075-f081, jobid=5951796, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f084-f090, jobid=5951797, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f093-f099, jobid=5951798, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f102-f108, jobid=5951799, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f111-f114, jobid=5951800, in state QUEUED (PENDING) -2025-07-28 21:58:12 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f117-f120, jobid=5951801, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f033-f035, jobid=5951788, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f036-f038, jobid=5951789, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f039-f041, jobid=5951790, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f042-f044, jobid=5951791, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f045-f047, jobid=5951792, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f048-f054, jobid=5951793, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f057-f063, jobid=5951794, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f066-f072, jobid=5951795, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f075-f081, jobid=5951796, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f084-f090, jobid=5951797, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f093-f099, jobid=5951798, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f102-f108, jobid=5951799, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f111-f114, jobid=5951800, in state QUEUED (PENDING) -2025-07-28 22:04:14 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f117-f120, jobid=5951801, in state QUEUED (PENDING) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f033-f035, jobid=5951788, in state SUCCEEDED (COMPLETED), ran for 70.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f036-f038, jobid=5951789, in state SUCCEEDED (COMPLETED), ran for 70.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f039-f041, jobid=5951790, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f042-f044, jobid=5951791, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f045-f047, jobid=5951792, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f048-f054, jobid=5951793, in state SUCCEEDED (COMPLETED), ran for 74.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f057-f063, jobid=5951794, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f066-f072, jobid=5951795, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f075-f081, jobid=5951796, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f084-f090, jobid=5951797, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f093-f099, jobid=5951798, in state SUCCEEDED (COMPLETED), ran for 72.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f102-f108, jobid=5951799, in state SUCCEEDED (COMPLETED), ran for 73.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f111-f114, jobid=5951800, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_wavepostsbs_f117-f120, jobid=5951801, in state SUCCEEDED (COMPLETED), ran for 50.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_arch_vrfy -2025-07-28 22:10:17 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_arch_vrfy is pending at druby://130.18.14.151:37481 -2025-07-28 22:16:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_arch_vrfy is success, jobid=5951962 -2025-07-28 22:16:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_arch_vrfy, jobid=5951962, in state SUCCEEDED (COMPLETED), ran for 74.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:16:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submitting gfs_cleanup -2025-07-28 22:16:20 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of gfs_cleanup is pending at druby://130.18.14.151:33107 -2025-07-28 22:22:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Submission status of previously pending gfs_cleanup is success, jobid=5951987 -2025-07-28 22:22:23 -0500 :: hercules-login-1.hpc.msstate.edu :: Task gfs_cleanup, jobid=5951987, in state SUCCEEDED (COMPLETED), ran for 22.0 seconds, exit status=0, try=1 (of 2) -2025-07-28 22:22:23 -0500 :: hercules-login-1.hpc.msstate.edu :: This cycle is complete: Success diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f000.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f000.log deleted file mode 100644 index fb8658c0..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f000.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.120000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f001.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f001.log deleted file mode 100644 index 3d0346eb..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f001.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.130000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f002.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f002.log deleted file mode 100644 index 9399a939..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f002.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.140000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f003.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f003.log deleted file mode 100644 index 73b913fa..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f003.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.150000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f004.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f004.log deleted file mode 100644 index 79b4b47a..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f004.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.160000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f005.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f005.log deleted file mode 100644 index 1c036364..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f005.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.170000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f006.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f006.log deleted file mode 100644 index 52aa6abe..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f006.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.180000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f007.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f007.log deleted file mode 100644 index 0b7a2e11..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f007.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.190000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f008.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f008.log deleted file mode 100644 index 95333b2a..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f008.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.200000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f009.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f009.log deleted file mode 100644 index ab8eebec..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f009.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.210000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f010.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f010.log deleted file mode 100644 index e0476174..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f010.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.220000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f011.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f011.log deleted file mode 100644 index 13877d08..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f011.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.230000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f012.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f012.log deleted file mode 100644 index e5c11e5c..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f012.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.000000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f013.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f013.log deleted file mode 100644 index 9b596319..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f013.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.010000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f014.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f014.log deleted file mode 100644 index eeb0f6a8..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f014.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.020000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f015.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f015.log deleted file mode 100644 index ae84a4d2..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f015.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.030000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f016.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f016.log deleted file mode 100644 index b6c0169b..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f016.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.040000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f017.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f017.log deleted file mode 100644 index ed8b41e1..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f017.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.050000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f018.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f018.log deleted file mode 100644 index 20b105db..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f018.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.060000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f019.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f019.log deleted file mode 100644 index 7121c171..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f019.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.070000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f020.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f020.log deleted file mode 100644 index b6dd703a..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f020.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.080000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f021.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f021.log deleted file mode 100644 index ecbcfeed..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f021.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.090000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f022.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f022.log deleted file mode 100644 index b992cc83..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f022.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.100000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f023.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f023.log deleted file mode 100644 index af41774b..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f023.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.110000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f024.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f024.log deleted file mode 100644 index 1cd07a33..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f024.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.120000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f025.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f025.log deleted file mode 100644 index 7b078307..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f025.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.130000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f026.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f026.log deleted file mode 100644 index fb60f743..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f026.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.140000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f027.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f027.log deleted file mode 100644 index 88fe03d6..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f027.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.150000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f028.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f028.log deleted file mode 100644 index 878d9fb6..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f028.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.160000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f029.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f029.log deleted file mode 100644 index 155a7055..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f029.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.170000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f030.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f030.log deleted file mode 100644 index fac379ad..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f030.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.180000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f031.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f031.log deleted file mode 100644 index 2beff41c..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f031.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.190000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f032.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f032.log deleted file mode 100644 index 19d3d75f..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f032.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.200000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f033.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f033.log deleted file mode 100644 index 50ae1568..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f033.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.210000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f034.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f034.log deleted file mode 100644 index 1d22a130..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f034.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.220000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f035.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f035.log deleted file mode 100644 index f884d7c7..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f035.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.230000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f036.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f036.log deleted file mode 100644 index 2c29e7af..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f036.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.000000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f037.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f037.log deleted file mode 100644 index 8f762b1e..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f037.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.010000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f038.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f038.log deleted file mode 100644 index 4e7f5b75..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f038.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.020000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f039.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f039.log deleted file mode 100644 index 5b7c2a6a..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f039.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.030000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f040.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f040.log deleted file mode 100644 index 80055e31..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f040.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.040000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f041.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f041.log deleted file mode 100644 index 391d7ecc..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f041.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.050000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f042.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f042.log deleted file mode 100644 index 962eb002..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f042.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.060000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f043.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f043.log deleted file mode 100644 index e7173bf6..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f043.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.070000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f044.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f044.log deleted file mode 100644 index 3f5741da..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f044.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.080000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f045.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f045.log deleted file mode 100644 index 722e8cc8..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f045.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.090000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f046.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f046.log deleted file mode 100644 index 321a7e00..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f046.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.100000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f047.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f047.log deleted file mode 100644 index d2622d7b..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f047.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.110000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f048.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f048.log deleted file mode 100644 index 897dff1a..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f048.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.120000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f049.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f049.log deleted file mode 100644 index e3431e95..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f049.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.130000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f050.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f050.log deleted file mode 100644 index 51c43652..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f050.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.140000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f051.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f051.log deleted file mode 100644 index 2401fa48..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f051.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.150000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f052.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f052.log deleted file mode 100644 index 7e76caf6..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f052.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.160000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f053.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f053.log deleted file mode 100644 index c46fe94d..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f053.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.170000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f054.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f054.log deleted file mode 100644 index 4ab62d03..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f054.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.180000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f055.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f055.log deleted file mode 100644 index 82e6b2be..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f055.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.190000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f056.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f056.log deleted file mode 100644 index ad2f3b70..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f056.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.200000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f057.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f057.log deleted file mode 100644 index fc278758..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f057.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.210000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f058.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f058.log deleted file mode 100644 index cc7e1fc7..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f058.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.220000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f059.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f059.log deleted file mode 100644 index 187963db..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f059.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.230000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f060.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f060.log deleted file mode 100644 index beb901e4..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f060.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.000000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f061.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f061.log deleted file mode 100644 index 200e01a2..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f061.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.010000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f062.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f062.log deleted file mode 100644 index 68b5d061..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f062.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.020000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f063.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f063.log deleted file mode 100644 index b1e8cf1e..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f063.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.030000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f064.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f064.log deleted file mode 100644 index 0cfcabcf..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f064.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.040000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f065.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f065.log deleted file mode 100644 index 21c09a7a..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f065.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.050000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f066.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f066.log deleted file mode 100644 index 8fd9b4fe..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f066.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.060000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f067.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f067.log deleted file mode 100644 index 59364442..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f067.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.070000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f068.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f068.log deleted file mode 100644 index 0de525e3..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f068.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.080000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f069.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f069.log deleted file mode 100644 index 7109ee04..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f069.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.090000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f070.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f070.log deleted file mode 100644 index 2be83ac6..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f070.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.100000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f071.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f071.log deleted file mode 100644 index 8f9d265d..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f071.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.110000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f072.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f072.log deleted file mode 100644 index 9c117a57..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f072.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.120000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f073.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f073.log deleted file mode 100644 index e1362376..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f073.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.130000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f074.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f074.log deleted file mode 100644 index d7e8a1b0..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f074.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.140000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f075.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f075.log deleted file mode 100644 index 07809db1..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f075.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.150000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f076.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f076.log deleted file mode 100644 index 444f61cb..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f076.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.160000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f077.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f077.log deleted file mode 100644 index 69a8050c..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f077.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.170000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f078.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f078.log deleted file mode 100644 index 9f90161c..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f078.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.180000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f079.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f079.log deleted file mode 100644 index 77ee38e6..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f079.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.190000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f080.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f080.log deleted file mode 100644 index 2134be74..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f080.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.200000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f081.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f081.log deleted file mode 100644 index b182f2af..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f081.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.210000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f082.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f082.log deleted file mode 100644 index 5a02ee3b..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f082.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.220000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f083.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f083.log deleted file mode 100644 index afbe99f3..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f083.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.230000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f084.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f084.log deleted file mode 100644 index d616bd0e..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f084.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.000000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f085.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f085.log deleted file mode 100644 index 619cacd8..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f085.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.010000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f086.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f086.log deleted file mode 100644 index e5660123..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f086.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.020000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f087.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f087.log deleted file mode 100644 index 29ee5e97..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f087.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.030000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f088.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f088.log deleted file mode 100644 index 01318764..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f088.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.040000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f089.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f089.log deleted file mode 100644 index 9e6d833e..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f089.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.050000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f090.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f090.log deleted file mode 100644 index f15d8e68..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f090.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.060000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f091.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f091.log deleted file mode 100644 index 80f1e8ac..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f091.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.070000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f092.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f092.log deleted file mode 100644 index ac3451da..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f092.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.080000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f093.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f093.log deleted file mode 100644 index b8ede776..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f093.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.090000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f094.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f094.log deleted file mode 100644 index bae30e6d..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f094.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.100000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f095.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f095.log deleted file mode 100644 index 9fa951f0..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f095.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.110000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f096.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f096.log deleted file mode 100644 index 05a0e560..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f096.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.120000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f097.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f097.log deleted file mode 100644 index 5d1f5982..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f097.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.130000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f098.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f098.log deleted file mode 100644 index a48fb48a..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f098.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.140000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f099.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f099.log deleted file mode 100644 index 1440cd13..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f099.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.150000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f100.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f100.log deleted file mode 100644 index 5fcb2b15..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f100.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.160000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f101.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f101.log deleted file mode 100644 index fb684e1e..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f101.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.170000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f102.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f102.log deleted file mode 100644 index 67467d9b..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f102.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.180000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f103.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f103.log deleted file mode 100644 index b42239a7..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f103.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.190000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f104.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f104.log deleted file mode 100644 index 410c3086..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f104.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.200000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f105.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f105.log deleted file mode 100644 index b92117b9..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f105.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.210000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f106.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f106.log deleted file mode 100644 index bacdf57d..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f106.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.220000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f107.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f107.log deleted file mode 100644 index 65e3467c..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f107.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.230000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f108.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f108.log deleted file mode 100644 index 6b6e7f9c..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f108.log +++ /dev/null @@ -1 +0,0 @@ - The 20210328.000000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f109.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f109.log deleted file mode 100644 index 6bfcd4fa..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f109.log +++ /dev/null @@ -1 +0,0 @@ - The 20210328.010000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f110.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f110.log deleted file mode 100644 index 57aaf9dc..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f110.log +++ /dev/null @@ -1 +0,0 @@ - The 20210328.020000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f111.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f111.log deleted file mode 100644 index 4842972b..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f111.log +++ /dev/null @@ -1 +0,0 @@ - The 20210328.030000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f112.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f112.log deleted file mode 100644 index 03a02d27..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f112.log +++ /dev/null @@ -1 +0,0 @@ - The 20210328.040000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f113.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f113.log deleted file mode 100644 index f898565a..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f113.log +++ /dev/null @@ -1 +0,0 @@ - The 20210328.050000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f114.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f114.log deleted file mode 100644 index 4fac666d..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f114.log +++ /dev/null @@ -1 +0,0 @@ - The 20210328.060000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f115.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f115.log deleted file mode 100644 index 9ce6bb2a..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f115.log +++ /dev/null @@ -1 +0,0 @@ - The 20210328.070000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f116.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f116.log deleted file mode 100644 index 884b9bcb..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f116.log +++ /dev/null @@ -1 +0,0 @@ - The 20210328.080000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f117.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f117.log deleted file mode 100644 index ddfecb8e..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f117.log +++ /dev/null @@ -1 +0,0 @@ - The 20210328.090000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f118.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f118.log deleted file mode 100644 index aa1611e3..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f118.log +++ /dev/null @@ -1 +0,0 @@ - The 20210328.100000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f119.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f119.log deleted file mode 100644 index 7140307c..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f119.log +++ /dev/null @@ -1 +0,0 @@ - The 20210328.110000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f120.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f120.log deleted file mode 100644 index f6976412..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.points.f120.log +++ /dev/null @@ -1 +0,0 @@ - The 20210328.120000.out_pnt.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.2021032312.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.2021032312.log deleted file mode 100644 index 4d720677..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.2021032312.log +++ /dev/null @@ -1,9480 +0,0 @@ - *** WAVEWATCH III Program shell *** -=============================================== - - - Wave wav_coupling_to_cice setting is F - WW3 runtype is initial - Time interval : --------------------------------------------------- - Starting time : 2021/03/23 12:00:00 UTC - start_ymd, stop_ymd = 20210323 20210328 -(wav_comp_nuopc):(InitializeAdvertise): Wave use_restartnc setting is T -(wav_comp_nuopc):(InitializeAdvertise): Custom restart prefix is ufs.cpld.ww3.r. -(wav_comp_nuopc):(InitializeAdvertise): Wave use_historync setting is F -(wav_comp_nuopc):(InitializeAdvertise): Wave restart_from_binary setting is T - wav_pio_init : pio_netcdf_format = 64BIT_OFFSET 512 - wav_pio_init : pio_typename = PNETCDF 1 - wav_pio_init : pio_root = 1 - wav_pio_init : pio_stride = 4 - wav_pio_init : pio_numiotasks = -99 - wav_pio_init : update pio_numiotasks = 10 - wav_pio_init : pio_rearranger = BOX 1 - wav_pio_init calling pio init - wav_pio_init : pio_root = 1 - wav_pio_init : pio_stride = 4 - wav_pio_init : pio_numiotasks = 10 - wav_pio_init : pio_debug_level = 0 -(wav_comp_nuopc:wavinit_ufs) call read_shel_config - - Input fields : - -------------------------------------------------- - water levels ---/NO - currents YES/-- (coupling field) - winds YES/-- (coupling field) - ice fields YES/-- (coupling field) - momentum ---/NO - air density ---/NO - mean param. ---/NO - 1D spectra ---/NO - 2D spectra ---/NO - - Fields : Current vel. - Wind speed - Ice concentration - Wave height - Mean wave period(+2) - Mean wave period(+1) - Peak frequency - Mean wave dir. a1b1 - Peak direction - Part. wave height - Part. peak period - Part. mean direction - Charnock parameter - Point 1 : -148.02 56.31 46001 - 2 : -154.98 52.70 46066 - 3 : -146.83 60.22 46061 - 4 : -160.81 53.93 46075 - 5 : -148.00 59.50 46076 - 6 : -152.45 56.05 46078 - 7 : -152.09 59.76 46106 - 8 : -150.00 58.00 46080 - 9 : -151.83 59.60 46108 - 10 : -160.00 57.70 46021 - 11 : -146.80 60.58 46060 - 12 : -154.18 57.91 46077 - 13 : -152.23 59.05 46079 - 14 : -152.23 59.05 46105 - 15 : -147.99 59.92 46107 - 16 : -165.48 64.46 46265 - 17 : -136.10 50.93 46004 - 18 : -138.85 53.91 46184 - 19 : -143.42 59.69 46082 - 20 : -138.00 58.25 46083 - 21 : -136.16 56.59 46084 - 22 : -142.56 56.85 46085 - 23 : -134.28 54.16 46205 - 24 : -132.45 54.38 46145 - 25 : -131.22 51.83 46147 - 26 : -131.10 53.62 46183 - 27 : -129.81 52.42 46185 - 28 : -128.75 51.37 46204 - 29 : -129.92 50.87 46207 - 30 : -132.68 52.52 46208 - 31 : -129.79 52.44 46138 - 32 : -130.27 42.60 46002 - 33 : -137.48 40.80 46006 - 34 : -130.00 37.98 46059 - 35 : -120.87 34.88 46011 - 36 : -122.88 37.36 46012 - 37 : -123.32 38.23 46013 - 38 : -123.97 39.22 46014 - 39 : -124.54 40.78 46022 - 40 : -120.97 34.71 46023 - 41 : -122.82 37.75 46026 - 42 : -124.38 41.85 46027 - 43 : -124.85 42.75 46015 - 44 : -119.08 33.75 46025 - 45 : -121.89 35.74 46028 - 46 : -124.53 40.42 46030 - 47 : -122.42 36.75 46042 - 48 : -119.53 32.43 46047 - 49 : -124.53 44.62 46050 - 50 : -119.85 34.24 46053 - 51 : -120.45 34.27 46054 - 52 : -121.01 35.10 46062 - 53 : -120.70 34.27 46063 - 54 : -120.20 33.65 46069 - 55 : -118.00 32.50 46086 - 56 : -125.77 45.88 46089 - 57 : -124.74 40.29 46213 - 58 : -123.46 37.94 46214 - 59 : -119.80 34.33 46216 - 60 : -119.43 34.17 46217 - 61 : -120.78 34.45 46218 - 62 : -119.88 33.22 46219 - 63 : -118.64 33.86 46221 - 64 : -118.32 33.62 46222 - 65 : -117.77 33.46 46223 - 66 : -117.47 33.18 46224 - 67 : -117.39 32.93 46225 - 68 : -117.44 32.63 46227 - 69 : -124.55 43.77 46229 - 70 : -117.37 32.75 46231 - 71 : -117.43 32.52 46232 - 72 : -117.75 32.64 SGX01 - 73 : -118.00 30.00 TPC50 - 74 : -135.00 20.00 TPC51 - 75 : -117.00 20.00 TPC52 - 76 : -120.00 6.00 TPC53 - 77 : -95.00 15.00 TPC54 - 78 : -88.00 9.00 TPC55 - 79 : -80.00 6.00 TPC56 - 80 : -130.50 48.10 OPCP01 - 81 : -126.60 48.10 OPCP02 - 82 : -129.70 45.30 OPCP03 - 83 : -125.60 45.30 OPCP04 - 84 : -129.90 41.75 OPCP05 - 85 : -125.80 41.90 OPCP06 - 86 : -129.00 38.50 OPCP07 - 87 : -125.50 39.20 OPCP08 - 88 : -125.40 36.40 OPCP09 - 89 : -125.00 33.30 OPCP10 - 90 : -122.30 34.60 OPCP11 - 91 : -121.50 30.90 OPCP12 - 92 : -117.00 29.60 OPCP13 - 93 : -120.86 35.20 46215 - 94 : -121.95 36.76 46236 - 95 : -122.63 37.79 46237 - 96 : -119.47 33.40 46238 - 97 : -122.10 36.34 46239 - 98 : -121.91 36.62 46240 - 99 : -124.13 46.22 46243 - 100 : -124.36 40.89 46244 - 101 : -145.20 50.03 46246 - 102 : -124.64 46.13 46248 - 103 : -119.20 33.00 46024 - 104 : -121.90 36.83 46091 - 105 : -122.03 36.75 46092 - 106 : -122.41 36.69 46093 - 107 : -124.30 44.64 46094 - 108 : -124.30 44.64 46097 - 109 : -124.96 44.38 46098 - 110 : -122.33 36.69 46114 - 111 : -124.31 40.75 46212 - 112 : -117.35 32.85 46226 - 113 : -117.32 32.94 46233 - 114 : -117.17 32.57 46235 - 115 : -117.44 33.22 46242 - 116 : -122.83 37.75 46247 - 117 : -119.71 33.82 46249 - 118 : -119.09 34.03 46250 - 119 : -119.56 33.77 46251 - 120 : -119.26 33.95 46252 - 121 : -118.18 33.58 46253 - 122 : -117.27 32.87 46254 - 123 : -119.65 33.40 46255 - 124 : -118.20 33.70 46256 - 125 : -120.77 34.44 46257 - 126 : -117.50 32.75 46258 - 127 : -121.50 34.77 46259 - 128 : -119.00 33.70 46262 - 129 : -131.02 46.05 46005 - 130 : -133.94 48.35 46036 - 131 : -127.93 49.74 46132 - 132 : -126.00 48.84 46206 - 133 : -124.51 46.12 46029 - 134 : -124.75 47.34 46041 - 135 : -124.73 48.49 46087 - 136 : -124.24 46.86 46211 - 137 : -123.17 48.33 46088 - 138 : -124.13 46.17 46096 - 139 : -124.57 46.99 46099 - 140 : -124.97 46.85 46100 - 141 : -124.95 47.97 46119 - 142 : -124.06 46.22 46127 - 143 : -126.01 48.84 46139 - 144 : -151.70 57.48 46264 - 145 : -177.58 57.05 46035 - 146 : 175.28 55.00 46070 - 147 : -172.03 54.94 46073 - 148 : 179.05 51.16 46071 - 149 : -171.73 52.25 46072 - 150 : -168.00 55.88 46020 - 151 : -162.21 23.43 51001 - 152 : -157.78 17.19 51002 - 153 : -160.82 19.22 51003 - 154 : -152.48 17.52 51004 - 155 : -158.00 24.00 HNL01 - 156 : -153.00 22.50 HNL02 - 157 : -157.75 22.00 HNL10 - 158 : -158.25 21.00 HNL11 - 159 : -156.50 19.75 HNL12 - 160 : -158.12 21.67 51201 - 161 : -157.68 21.42 51202 - 162 : -177.40 28.20 MIDWAY - 163 : -169.50 16.70 JOHNSTON - 164 : -166.30 23.90 FF_SHOALS - 165 : -154.06 23.55 51000 - 166 : -153.90 23.56 51100 - 167 : -162.06 24.32 51101 - 168 : -157.00 20.79 51203 - 169 : -158.12 21.28 51204 - 170 : -156.42 21.02 51205 - 171 : -154.97 19.78 51206 - 172 : -153.87 0.02 51028 - 173 : -158.30 21.10 51200 - 174 : -159.57 22.28 51208 - 175 : -170.50 -14.27 51209 - 176 : 134.67 7.69 52212 - 177 : -157.96 21.30 51211 - 178 : -158.15 21.32 51212 - 179 : -157.00 20.75 51213 - 180 : 147.50 16.00 SAIPAN - 181 : 143.75 12.00 GUAM - 182 : 166.50 19.50 WAKE - 183 : 136.25 9.00 PALAU - 184 : 138.00 9.60 YAP - 185 : 152.50 8.00 CHUUK - 186 : 157.50 7.00 POHNPEI - 187 : 163.00 5.10 KOSRAE - 188 : 171.25 8.00 MAJURO - 189 : 163.75 13.00 ENEWETAK - 190 : 174.00 1.00 TARAWA - 191 : 144.79 13.35 52200 - 192 : 126.02 37.23 22101 - 193 : 125.77 34.80 22102 - 194 : 127.50 34.00 22103 - 195 : 128.90 34.77 22104 - 196 : 130.00 37.53 22105 - 197 : 171.39 7.04 52201 - 198 : 144.80 13.68 52202 - 199 : 145.66 15.27 52211 - 200 : 145.00 14.70 SAIPAN_W - 201 : 145.60 14.30 SAIPAN_E - 202 : 133.90 7.50 PALAU_W - 203 : 134.90 7.70 PALAU_E - 204 : 138.40 9.60 YAP_E - 205 : 151.20 7.40 CHUUK_W - 206 : 152.10 7.60 CHUUK_E - 207 : 158.40 7.10 POHNPEI_E - 208 : 162.80 5.90 KOSRAE_W - 209 : 167.80 9.50 KWAJALEIN - 210 : 167.50 8.67 KWAJ_W1 - 211 : 167.00 8.67 KWAJ_W2 - 212 : 168.17 8.67 KWAJ_E - 213 : 166.33 9.17 WOTHO - 214 : 168.00 9.17 ROI_NAMUR - 215 : 171.50 9.17 WOTJE_E - 216 : 171.40 7.60 MAJURO_02 - 217 : 171.50 6.60 MAJURO_03 - 218 : 133.62 33.19 21178 - 219 : 131.11 37.46 21229 - 220 : 125.75 36.25 22108 - 221 : 126.14 33.79 22184 - 222 : 125.43 37.09 22185 - 223 : 125.81 35.66 22186 - 224 : 127.02 33.13 22187 - 225 : 128.23 34.39 22188 - 226 : 129.84 35.35 22189 - 227 : 129.87 36.91 22190 - 228 : 146.25 -12.00 NEWGUINE_S - 229 : -168.75 -15.00 PAGO_PAGO - 230 : 176.25 -18.00 NADI - 231 : 179.20 -8.50 FUNAFUTI - 232 : -175.00 -22.00 TONGATAPU - 233 : -159.80 -21.20 RAROTONGA - 234 : 167.50 -24.00 NOUMEA - 235 : 167.50 -18.00 PORT_VILA - 236 : -149.60 -19.00 PAPEETE - 237 : -169.90 -19.10 NIUE - 238 : -171.90 -9.20 NUKUNONO - 239 : 167.00 -0.50 NAURU - 240 : 160.00 -12.00 SOLOMON_SW - 241 : 165.00 -12.00 SOLOMON_SE - 242 : 160.00 -5.00 SOLOMON_N - 243 : -85.00 -15.00 TPC01 - 244 : -110.00 -15.00 TPC02 - 245 : -135.00 -15.00 TPC03 - 246 : 150.18 -37.29 55020 - 247 : 151.07 -23.31 55033 - 248 : 153.63 -27.49 55035 - 249 : 148.19 -38.60 55039 - 250 : 118.00 -1.00 SULAWESI - 251 : 120.00 -7.50 P_BONARAT - 252 : 125.00 -5.00 P_RUNDUMA - 253 : 133.00 -8.00 P_JAMDENA - 254 : 122.00 -11.00 P_SAWA - 255 : -89.67 25.90 42001 - 256 : -94.42 25.17 42002 - 257 : -85.94 26.07 42003 - 258 : -88.77 30.09 42007 - 259 : -95.36 27.91 42019 - 260 : -96.70 26.94 42020 - 261 : -94.40 29.22 42035 - 262 : -84.52 28.50 42036 - 263 : -86.02 28.79 42039 - 264 : -88.21 29.18 42040 - 265 : -90.46 27.50 42041 - 266 : -92.55 27.42 42038 - 267 : -94.05 22.01 42055 - 268 : -84.28 27.35 42099 - 269 : -86.00 23.00 TPC26 - 270 : -87.55 30.06 42012 - 271 : -88.49 28.19 42887 - 272 : -87.73 26.00 42054 - 273 : -82.92 27.17 42013 - 274 : -82.22 25.25 42014 - 275 : -83.31 28.31 42021 - 276 : -83.74 27.50 42022 - 277 : -83.09 26.01 42023 - 278 : -94.90 28.98 42043 - 279 : -97.05 26.19 42044 - 280 : -96.50 26.22 42045 - 281 : -94.04 27.89 42046 - 282 : -93.60 27.90 42047 - 283 : -88.65 30.04 42067 - 284 : -83.65 25.70 42097 - 285 : -82.93 27.59 42098 - 286 : -90.47 26.67 42360 - 287 : -92.49 27.55 42361 - 288 : -90.65 27.80 42362 - 289 : -89.22 28.16 42363 - 290 : -88.09 29.06 42364 - 291 : -89.12 28.20 42365 - 292 : -90.28 27.21 42369 - 293 : -90.54 27.32 42370 - 294 : -88.06 28.87 42374 - 295 : -88.29 28.52 42375 - 296 : -87.94 29.11 42376 - 297 : -94.90 26.13 42390 - 298 : -90.03 27.20 42392 - 299 : -89.24 28.16 42394 - 300 : -90.79 26.40 42395 - 301 : -85.06 19.87 42056 - 302 : -81.50 16.83 42057 - 303 : -75.06 15.09 42058 - 304 : -81.95 24.39 42080 - 305 : -81.75 24.00 EYW01 - 306 : -82.25 25.00 EYW02 - 307 : -66.50 19.00 PUERTO_R_N - 308 : -66.50 17.50 PUERTO_R_S - 309 : -67.50 15.01 42059 - 310 : -85.38 -19.62 32012 - 311 : -63.00 15.00 TPC21 - 312 : -77.00 12.00 TPC22 - 313 : -80.00 15.00 TPC23 - 314 : -76.00 22.00 TPC24 - 315 : -80.00 24.00 TPC25 - 316 : -63.50 16.50 42060 - 317 : -67.50 19.00 CARCOOS01 - 318 : -65.50 19.00 CARCOOS02 - 319 : -64.00 19.00 CARCOOS03 - 320 : -64.40 17.30 CARCOOS04 - 321 : -67.50 17.50 CARCOOS05 - 322 : -74.68 11.16 41194 - 323 : -66.52 17.86 42085 - 324 : -80.06 19.70 42089 - 325 : -64.76 18.25 41052 - 326 : -65.00 18.26 41051 - 327 : -65.46 18.26 41056 - 328 : -67.28 18.38 41115 - 329 : -81.08 30.00 41117 - 330 : -81.24 24.53 42079 - 331 : -75.04 36.00 42086 - 332 : -81.97 24.41 42095 - 333 : -72.66 34.68 41001 - 334 : -75.36 32.32 41002 - 335 : -79.09 32.50 41004 - 336 : -80.87 31.40 41008 - 337 : -80.17 28.50 41009 - 338 : -78.47 28.95 41010 - 339 : -80.60 30.00 41012 - 340 : -77.74 33.44 41013 - 341 : -75.40 35.01 41025 - 342 : -77.28 34.48 41035 - 343 : -76.95 34.21 41036 - 344 : -65.01 20.99 41043 - 345 : -70.99 24.00 41046 - 346 : -71.49 27.47 41047 - 347 : -69.65 31.98 41048 - 348 : -81.29 30.71 41112 - 349 : -80.53 28.40 41113 - 350 : -80.22 27.55 41114 - 351 : -74.84 36.61 44014 - 352 : -77.00 30.75 CHS01 - 353 : -77.00 34.00 ILM01 - 354 : -78.50 33.25 ILM02 - 355 : -80.25 29.50 JAX02 - 356 : -79.50 27.25 MLB01 - 357 : -79.50 26.25 MIA01 - 358 : -79.75 25.00 MIA02 - 359 : -74.60 36.30 OPCA05 - 360 : -73.80 35.60 OPCA06 - 361 : -70.80 34.90 OPCA07 - 362 : -76.00 33.80 OPCA08 - 363 : -72.30 32.80 OPCA09 - 364 : -77.36 33.99 41037 - 365 : -77.72 34.14 41038 - 366 : -63.00 27.50 41049 - 367 : -58.69 21.65 41044 - 368 : -77.30 34.48 41109 - 369 : -77.71 34.14 41110 - 370 : -67.28 18.38 41111 - 371 : -66.10 18.47 41053 - 372 : -65.16 18.48 41058 - 373 : -78.48 33.84 41024 - 374 : -78.14 33.30 41027 - 375 : -79.62 32.80 41029 - 376 : -79.34 32.52 41030 - 377 : -80.41 32.28 41033 - 378 : -38.00 24.58 41061 - 379 : -75.10 35.78 41062 - 380 : -75.94 34.78 41063 - 381 : -76.95 34.21 41064 - 382 : -78.01 33.72 41108 - 383 : -76.95 34.21 41159 - 384 : -75.71 36.20 44056 - 385 : -53.62 44.26 44138 - 386 : -66.58 41.11 44011 - 387 : -58.00 43.00 44141 - 388 : -64.02 42.50 44142 - 389 : -48.01 46.77 WRB07 - 390 : -62.00 42.26 44137 - 391 : -57.08 44.26 44139 - 392 : -51.74 43.75 44140 - 393 : -64.01 42.50 44150 - 394 : -70.43 38.48 44004 - 395 : -69.16 43.19 44005 - 396 : -69.43 40.50 44008 - 397 : -74.70 38.46 44009 - 398 : -72.10 40.70 44017 - 399 : -69.29 41.26 44018 - 400 : -73.17 40.25 44025 - 401 : -70.25 42.50 BOX01 - 402 : -67.50 44.00 CAR01 - 403 : -71.01 41.38 44070 - 404 : -65.93 42.31 44024 - 405 : -67.31 44.27 44027 - 406 : -67.88 43.49 44037 - 407 : -66.55 43.62 44038 - 408 : -53.39 46.44 44251 - 409 : -57.35 47.28 44255 - 410 : -69.75 43.25 GYX01 - 411 : -67.70 42.35 OPCA01 - 412 : -72.00 39.30 OPCA02 - 413 : -65.70 39.30 OPCA03 - 414 : -70.10 37.30 OPCA04 - 415 : -75.72 36.92 44099 - 416 : -75.59 36.26 44100 - 417 : -72.60 39.58 44066 - 418 : -75.49 36.87 44093 - 419 : -75.33 35.75 44095 - 420 : -75.81 37.02 44096 - 421 : -71.13 40.97 44097 - 422 : -70.17 42.80 44098 - 423 : -70.14 43.53 44007 - 424 : -70.65 42.35 44013 - 425 : -70.19 41.44 44020 - 426 : -70.57 42.52 44029 - 427 : -70.43 43.18 44030 - 428 : -70.06 43.57 44031 - 429 : -69.36 43.72 44032 - 430 : -69.00 44.06 44033 - 431 : -68.11 44.11 44034 - 432 : -72.65 41.14 44039 - 433 : -75.18 38.88 44054 - 434 : -75.26 39.12 44055 - 435 : -76.26 37.57 44058 - 436 : -72.07 41.26 44060 - 437 : -76.42 38.56 44062 - 438 : -76.45 38.96 44063 - 439 : -76.09 37.00 44064 - 440 : -73.70 40.37 44065 - 441 : -76.27 37.20 44072 - 442 : -75.33 37.76 44089 - 443 : -70.33 41.84 44090 - 444 : -73.77 39.77 44091 - 445 : -70.63 42.94 44092 - 446 : -73.11 40.58 44094 - 447 : -63.41 44.50 44172 - 448 : -57.34 47.26 44235 - 449 : -76.15 37.02 44087 - 450 : -48.13 -27.70 31201 - 451 : -48.75 -32.00 RIO_GRANDE - 452 : -46.25 -28.00 FLORIPA - 453 : -43.75 -25.00 SANTOS - 454 : -34.57 -8.15 31052 - 455 : -43.09 -23.03 31260 - 456 : -47.37 -28.50 31374 - 457 : -44.93 -25.28 31051 - 458 : -51.35 -32.60 31053 - 459 : -42.19 -22.99 SIODOC - 460 : -44.27 -23.42 ILHAGRANDE - 461 : -43.46 -23.16 RECREIO - 462 : -43.12 -23.11 CAGARRAS - 463 : -49.81 -31.52 31375 - 464 : -53.08 14.55 41040 - 465 : -46.00 14.53 41041 - 466 : -57.90 15.90 41100 - 467 : -56.20 14.60 41101 - 468 : -55.00 15.00 TPC20 - 469 : -38.75 -21.00 CAMPOS - 470 : -36.25 -13.00 SALVADOR - 471 : -32.50 -8.00 RECIFE - 472 : -36.25 -3.00 FORTALEZA - 473 : -47.50 3.00 AMAZON - 474 : -30.00 1.00 PETER_PAUL - 475 : -5.78 35.90 TANGER - 476 : -9.63 30.42 AGADIR - 477 : -7.62 33.70 CASABLANCA - 478 : 3.50 6.00 LAGOS - 479 : 5.00 5.00 FORCADOS - 480 : 7.00 4.00 NIGERDELTA - 481 : 8.00 4.00 CALABAR - 482 : -50.95 14.75 41060 - 483 : -60.85 11.19 42087 - 484 : -60.52 11.30 42088 - 485 : -83.75 0.00 TPC00 - 486 : -93.75 0.00 TPC04 - 487 : 102.00 -5.00 P_ENGGANO - 488 : 107.00 0.00 P_PENJAN - 489 : 110.00 -5.00 SEMARANG - 490 : 132.00 1.00 P_IGI - 491 : 100.00 4.00 P_PANGKOR - 492 : 123.00 3.00 BORNEO - 493 : 126.00 1.00 P_GUREDA - 494 : -125.03 10.05 43010 - 495 : -144.67 13.73 52009 - 496 : -5.00 45.20 62001 - 497 : -20.00 41.60 62002 - 498 : -12.40 48.70 62029 - 499 : -7.90 51.40 62023 - 500 : -5.60 48.50 62052 - 501 : -13.30 51.00 62081 - 502 : -11.20 53.13 62090 - 503 : -5.42 53.47 62091 - 504 : -10.55 51.22 62092 - 505 : -9.07 54.67 62093 - 506 : -6.70 51.69 62094 - 507 : -15.92 53.06 62095 - 508 : -2.90 49.90 62103 - 509 : -12.36 54.54 62105 - 510 : -9.90 57.00 62106 - 511 : -6.10 50.10 62107 - 512 : -19.50 53.50 62108 - 513 : -8.50 47.50 62163 - 514 : -4.70 52.30 62301 - 515 : -5.10 51.60 62303 - 516 : 0.00 50.40 62305 - 517 : 2.00 51.40 62170 - 518 : -11.40 59.10 64045 - 519 : -4.50 60.70 64046 - 520 : -23.10 64.05 TFGSK - 521 : -15.20 64.00 TFHFN - 522 : -20.35 63.00 TFSRT - 523 : 7.80 64.30 LF3F - 524 : 1.10 55.30 62026 - 525 : 0.00 57.00 62109 - 526 : 0.40 58.10 62111 - 527 : 1.30 58.70 62112 - 528 : 1.40 57.70 62116 - 529 : 0.00 57.90 62117 - 530 : 2.00 57.00 62119 - 531 : 1.40 58.70 62128 - 532 : 2.00 56.40 62132 - 533 : 1.00 57.10 62133 - 534 : 2.10 53.00 62142 - 535 : 1.80 57.70 62143 - 536 : 1.70 53.40 62144 - 537 : 2.80 53.10 62145 - 538 : 1.80 57.00 62152 - 539 : 0.50 57.40 62162 - 540 : 0.50 57.20 62164 - 541 : 1.90 51.10 62304 - 542 : 1.70 60.60 63055 - 543 : 1.60 59.50 63056 - 544 : 1.50 59.20 63057 - 545 : 1.10 61.20 63103 - 546 : 1.70 60.80 63108 - 547 : 1.50 59.50 63110 - 548 : 1.00 61.10 63112 - 549 : 1.70 61.00 63113 - 550 : 1.30 61.60 63115 - 551 : 2.30 61.20 LF3J - 552 : 3.70 60.60 LF4B - 553 : 2.20 59.60 LF4H - 554 : 1.90 58.40 LF4C - 555 : 3.20 56.50 LF5U - 556 : 3.28 51.99 EURO - 557 : 3.22 53.22 K13 - 558 : -3.03 43.63 62024 - 559 : -7.62 44.07 62082 - 560 : -9.40 42.12 62084 - 561 : -6.97 36.48 62085 - 562 : -15.82 28.18 13130 - 563 : -16.58 28.00 13131 - 564 : 0.90 57.70 62118 - 565 : 2.10 57.10 62146 - 566 : 1.50 59.50 63111 - 567 : 6.33 55.00 BSH01 - 568 : 7.89 54.16 BSH02 - 569 : 8.12 54.00 BSH03 - 570 : 6.58 54.00 BSH04 - 571 : 8.22 54.92 BSH05 - 572 : -25.00 65.69 TFBLK - 573 : -18.20 66.50 TFGRS - 574 : -13.50 65.65 TFKGR - 575 : 7.30 65.30 LF3N - 576 : 8.10 66.00 LF5T - 577 : 2.00 66.00 LDWR - 578 : 21.10 71.60 3FYT - 579 : 15.50 73.50 LFB1 - 580 : 30.00 74.00 LFB2 - 581 : -9.26 68.48 64071 - 582 : -166.07 70.03 48012 - 583 : -169.45 65.01 48114 - 584 : -146.04 70.37 48211 - 585 : -150.28 70.87 48212 - 586 : -164.13 71.50 48213 - 587 : -165.25 70.87 48214 - 588 : -167.95 71.76 48216 - 589 : 72.49 17.02 23092 - 590 : 73.75 15.40 23093 - 591 : 74.50 12.94 23094 - 592 : 80.39 13.19 23096 - 593 : 69.24 15.47 23097 - 594 : 72.51 10.65 23098 - 595 : 90.74 12.14 23099 - 596 : 87.56 18.35 23100 - 597 : 83.27 13.97 23101 - 598 : 87.50 15.00 23168 - 599 : 90.14 18.13 23169 - 600 : 72.66 8.33 23170 - 601 : 72.00 12.50 23172 - 602 : 78.57 8.21 23173 - 603 : 81.53 11.57 23174 - 604 : 93.00 6.00 G_NICOBAR - 605 : 116.14 -19.59 56002 - 606 : 115.40 -32.11 56005 - 607 : 114.78 -33.36 56006 - 608 : 114.94 -21.41 56007 - 609 : 22.17 -34.97 AGULHAS_FA - 610 : 104.00 6.00 MALAY01 - 611 : 105.00 3.00 MALAY02 - 612 : 110.00 3.00 MALAY03 - 613 : 113.00 5.00 MALAY04 - 614 : 116.00 7.50 MALAY05 - 615 : 117.00 7.50 MALAY06 - 616 : 121.90 -34.00 56010 - 617 : 114.10 -21.70 56012 - 618 : 136.20 -36.10 CADUCOU - 619 : 139.00 -38.00 SWROBE - 620 : 142.45 -39.20 WBAST1 - 621 : 141.50 -40.00 WBAST2 - 622 : 151.00 -40.00 EBAST - 623 : 146.50 -40.50 CBAST - 624 : 144.60 -42.30 CSORRELL - 625 : 144.50 -40.10 SEKING - 626 : 143.80 -39.20 NKING - 627 : 144.85 -38.60 PNEPEAN - 628 : 147.40 -39.20 EHOGAN - 629 : 147.00 -44.00 STHSEC - 630 : 149.50 -41.50 EBICHENO - 631 : 133.50 -33.50 WCAPYORK - 632 : 114.62 -19.78 JANSZ - 633 : 85.00 12.60 23167 - 634 : 70.00 11.02 23171 - 635 : 91.66 10.52 23451 - 636 : 89.04 10.97 23455 - 637 : 86.98 9.99 23456 - 638 : 70.10 5.16 23491 - 639 : 68.08 13.89 23492 - 640 : 66.98 11.12 23493 - 641 : 75.00 6.46 23494 - 642 : 68.97 7.13 23495 - 643 : 24.99 -34.06 JEFFREYBAY - 644 : 27.93 -33.03 EASTLONDON - 645 : 17.90 -33.06 SALDANHABY - 646 : 18.28 -34.20 CAPEPOINT - 647 : 22.17 -34.97 FAPLATFORM - 648 : 22.17 -34.15 MOSSELBAY - 649 : -160.00 72.00 Alaska_NS1 - 650 : -150.00 72.00 Alaska_NS2 - 651 : -140.00 72.00 Alaska_NS3 - 652 : -116.48 23.62 EFT1 - 653 : -123.75 36.00 DIABLO_01 - 654 : -121.50 35.75 CDIP01 - 655 : -122.00 35.75 CDIP02 - 656 : -122.00 35.25 CDIP03 - 657 : -122.00 34.75 CDIP04 - 658 : -122.00 34.25 CDIP05 - 659 : -122.00 33.75 CDIP06 - 660 : -122.00 33.25 CDIP07 - 661 : -122.00 32.75 CDIP08 - 662 : -122.00 32.25 CDIP09 - 663 : -122.00 31.75 CDIP10 - 664 : -122.00 31.25 CDIP11 - 665 : -122.00 30.75 CDIP12 - 666 : -121.75 30.75 CDIP13 - 667 : -121.25 30.75 CDIP14 - 668 : -120.75 30.75 CDIP15 - 669 : -120.25 30.75 CDIP16 - 670 : -119.75 30.75 CDIP17 - 671 : -119.25 30.75 CDIP18 - 672 : -118.75 30.75 CDIP19 - 673 : -118.25 30.75 CDIP20 - 674 : -117.75 30.75 CDIP21 - 675 : -117.25 30.75 CDIP22 - 676 : -116.75 30.75 CDIP23 - 677 : -116.25 30.75 CDIP24 - 678 : 90.00 -15.00 BKMG01 - 679 : 92.00 -15.00 BKMG02 - 680 : 94.00 -15.00 BKMG03 - 681 : 96.00 -15.00 BKMG04 - 682 : 98.00 -15.00 BKMG05 - 683 : 100.00 -15.00 BKMG06 - 684 : 102.00 -15.00 BKMG07 - 685 : 104.00 -15.00 BKMG08 - 686 : 106.00 -15.00 BKMG09 - 687 : 108.00 -15.00 BKMG10 - 688 : 110.00 -15.00 BKMG11 - 689 : 112.00 -15.00 BKMG12 - 690 : 114.00 -15.00 BKMG13 - 691 : 116.00 -15.00 BKMG14 - 692 : 118.00 -15.00 BKMG15 - 693 : 120.00 -15.00 BKMG16 - 694 : 122.00 -15.00 BKMG17 - 695 : 124.00 -15.00 BKMG18 - 696 : 124.50 -15.00 BKMG19 - 697 : 137.00 -15.00 BKMG20 - 698 : 139.00 -15.00 BKMG21 - 699 : 141.00 -15.00 BKMG22 - 700 : 145.00 -13.50 BKMG23 - 701 : 145.00 -13.00 BKMG24 - 702 : 145.00 -11.00 BKMG25 - 703 : 145.00 -9.00 BKMG26 - 704 : 145.00 -8.00 BKMG27 - 705 : 145.00 -3.00 BKMG28 - 706 : 145.00 -1.00 BKMG29 - 707 : 145.00 1.00 BKMG30 - 708 : 145.00 3.00 BKMG31 - 709 : 145.00 5.00 BKMG32 - 710 : 145.00 7.00 BKMG33 - 711 : 145.00 9.00 BKMG34 - 712 : 145.00 11.00 BKMG35 - 713 : 145.00 13.00 BKMG36 - 714 : 145.00 15.00 BKMG37 - 715 : 143.00 15.00 BKMG38 - 716 : 141.00 15.00 BKMG39 - 717 : 139.00 15.00 BKMG40 - 718 : 137.00 15.00 BKMG41 - 719 : 135.00 15.00 BKMG42 - 720 : 133.00 15.00 BKMG43 - 721 : 131.00 15.00 BKMG44 - 722 : 129.00 15.00 BKMG45 - 723 : 127.00 15.00 BKMG46 - 724 : 125.00 15.00 BKMG47 - 725 : 123.00 15.00 BKMG48 - 726 : 122.50 15.00 BKMG49 - 727 : 120.00 15.00 BKMG50 - 728 : 118.00 15.00 BKMG51 - 729 : 116.00 15.00 BKMG52 - 730 : 114.00 15.00 BKMG53 - 731 : 112.00 15.00 BKMG54 - 732 : 110.00 15.00 BKMG55 - 733 : 97.50 15.00 BKMG56 - 734 : 97.00 15.00 BKMG57 - 735 : 95.00 15.00 BKMG58 - 736 : 93.00 15.00 BKMG59 - 737 : 91.00 15.00 BKMG60 - 738 : 90.00 15.00 BKMG61 - 739 : 90.00 13.00 BKMG62 - 740 : 90.00 11.00 BKMG63 - 741 : 90.00 9.00 BKMG64 - 742 : 90.00 7.00 BKMG65 - 743 : 90.00 5.00 BKMG66 - 744 : 90.00 3.00 BKMG67 - 745 : 90.00 1.00 BKMG68 - 746 : 90.00 -1.00 BKMG69 - 747 : 90.00 -3.00 BKMG70 - 748 : 90.00 -5.00 BKMG71 - 749 : 90.00 -7.00 BKMG72 - 750 : 90.00 -9.00 BKMG73 - 751 : 90.00 -11.00 BKMG74 - 752 : 90.00 -13.00 BKMG75 - 753 : -77.50 6.26 32488 - 754 : -77.74 3.52 32487 - 755 : -72.22 12.35 41193 - 756 : 134.90 28.90 21004 - 757 : 126.30 28.10 22001 - 758 : 129.78 36.35 22106 - 759 : 126.33 33.00 22107 - 760 : 57.70 -20.45 MAUR01 - 761 : 57.75 -20.10 MAUR02 - 762 : 55.00 -25.00 V14003 - 763 : 60.00 -25.00 V14004 - 764 : 57.00 -20.00 V14005 - 765 : 60.00 -20.00 V14006 - 766 : 63.00 -20.00 V14007 - 767 : 64.00 -19.00 V14008 - 768 : 58.00 -18.00 V14009 - 769 : 60.00 -16.00 V14010 - 770 : 56.00 -10.00 V14011 - 771 : 57.00 -11.00 V14012 - 772 : 70.00 -25.00 V14013 - 773 : 80.00 -25.00 V14014 - 774 : 90.00 -25.00 V53015 - 775 : 70.00 -15.00 V14016 - 776 : 80.00 -15.00 V14017 - 777 : 90.00 -15.00 V53018 - 778 : 65.00 -5.00 V23019 - 779 : 75.00 -5.00 V23020 - 780 : 85.00 -5.00 V23021 - 781 : 95.00 -5.00 V23022 - 782 : 55.00 -5.00 V14023 - 783 : 10.93 -17.33 V14039 - 784 : 10.95 -18.10 V14040 - 785 : 11.68 -19.00 V14041 - 786 : 12.25 -20.07 V14042 - 787 : 12.65 -20.97 V14043 - 788 : 13.22 -21.93 V14044 - 789 : 13.42 -22.62 V14045 - 790 : 13.58 -23.61 V14046 - 791 : 13.66 -24.38 V14047 - 792 : 13.95 -25.25 V14048 - 793 : 14.03 -26.08 V14049 - 794 : 14.19 -26.73 V14050 - 795 : 14.54 -27.59 V14051 - 796 : 14.97 -28.34 V14052 - 797 : 15.68 -29.08 V14053 - 798 : 39.75 -5.00 V14054 - 799 : 39.75 -5.58 V14055 - 800 : 39.75 -6.22 V14056 - 801 : 39.80 -6.75 V14057 - 802 : 39.92 -7.08 V14058 - 803 : 39.75 -7.30 V14059 - 804 : 39.75 -8.17 V14060 - 805 : 40.25 -8.83 V14061 - 806 : 40.25 -9.17 V14062 - 807 : 40.25 -9.92 V14063 - 808 : 40.58 -10.33 V14064 - 809 : 42.00 -11.00 V14065 - 810 : 42.00 -12.00 V14066 - 811 : 42.00 -13.00 V14067 - 812 : 42.00 -14.00 V14068 - 813 : 42.00 -15.00 V14069 - 814 : 40.25 -17.00 V14070 - 815 : 39.00 -17.83 V14071 - 816 : 38.25 -18.00 V14072 - 817 : 37.25 -19.00 V14073 - 818 : 35.50 -20.00 V14074 - 819 : 36.00 -21.00 V14075 - 820 : 35.57 -21.92 V14076 - 821 : 35.58 -22.00 V14077 - 822 : 42.83 -22.00 V14078 - 823 : 36.00 -23.00 V14079 - 824 : 36.83 -24.00 V14080 - 825 : 36.00 -25.00 V14081 - 826 : 35.00 -25.50 V14082 - 827 : 34.00 -26.00 V14083 - 828 : 34.00 -27.00 V14084 - 829 : 40.00 -8.00 KNY51 - 830 : 42.00 -8.00 KNY52 - 831 : 44.00 -8.00 KNY53 - 832 : 46.00 -8.00 KNY54 - 833 : 48.00 -8.00 KNY55 - 834 : 50.00 -8.00 KNY56 - 835 : 52.00 -8.00 KNY57 - 836 : 54.00 -8.00 KNY58 - 837 : 56.00 -8.00 KNY59 - 838 : 58.00 -8.00 KNY60 - 839 : 60.00 -8.00 KNY61 - 840 : 62.00 -8.00 KNY62 - 841 : 64.00 -8.00 KNY63 - 842 : 66.00 -8.00 KNY64 - 843 : 68.00 -8.00 KNY65 - 844 : 70.00 -8.00 KNY66 - 845 : 70.00 -6.00 KNY67 - 846 : 70.00 -4.00 KNY68 - 847 : 70.00 -2.00 KNY69 - 848 : 70.00 0.00 KNY70 - 849 : 70.00 2.00 KNY71 - 850 : 70.00 4.00 KNY72 - 851 : 70.00 6.00 KNY73 - 852 : 70.00 8.00 KNY74 - 853 : 68.00 8.00 KNY75 - 854 : 66.00 8.00 KNY76 - 855 : 64.00 8.00 KNY77 - 856 : 62.00 8.00 KNY78 - 857 : 60.00 8.00 KNY79 - 858 : 58.00 8.00 KNY80 - 859 : 56.00 8.00 KNY81 - 860 : 54.00 8.00 KNY82 - 861 : 52.00 8.00 KNY83 - 862 : 42.00 -18.00 MDG51 - 863 : 42.00 -20.00 MDG52 - 864 : 42.00 -22.00 MDG53 - 865 : 42.00 -24.00 MDG54 - 866 : 42.00 -26.00 MDG55 - 867 : 42.00 -28.00 MDG56 - 868 : 42.00 -30.00 MDG57 - 869 : 44.00 -30.00 MDG58 - 870 : 46.00 -30.00 MDG59 - 871 : 48.00 -30.00 MDG60 - 872 : 50.00 -30.00 MDG61 - 873 : 52.00 -30.00 MDG62 - 874 : 54.00 -30.00 MDG63 - 875 : 54.00 -28.00 MDG64 - 876 : 54.00 -26.00 MDG65 - 877 : 54.00 -24.00 MDG66 - 878 : 54.00 -22.00 MDG67 - 879 : 54.00 -20.00 MDG68 - 880 : 54.00 -18.00 MDG69 - 881 : 54.00 -16.00 MDG70 - 882 : 54.00 -14.00 MDG71 - 883 : 54.00 -12.00 MDG72 - 884 : 54.00 -10.00 MDG73 - 885 : 52.00 -10.00 MDG74 - 886 : 50.00 -10.00 MDG75 - 887 : 48.00 -10.00 MDG76 - 888 : 46.00 -10.00 MDG77 - 889 : 44.00 -10.00 MDG78 - 890 : 44.00 -12.00 MDG79 - 891 : 44.00 -14.00 MDG80 - 892 : 42.00 -16.00 MDG81 - 893 : 50.00 -10.00 SYC51 - 894 : 52.00 -10.00 SYC52 - 895 : 54.00 -10.00 SYC53 - 896 : 56.00 -10.00 SYC54 - 897 : 58.00 -10.00 SYC55 - 898 : 60.00 -10.00 SYC56 - 899 : 60.00 -8.00 SYC57 - 900 : 60.00 -6.00 SYC58 - 901 : 60.00 -4.00 SYC59 - 902 : 60.00 -2.00 SYC60 - 903 : 60.00 0.00 SYC61 - 904 : 58.00 0.00 SYC62 - 905 : 56.00 0.00 SYC63 - 906 : 54.00 0.00 SYC64 - 907 : 52.00 0.00 SYC65 - 908 : 50.00 0.00 SYC66 - 909 : 50.00 -2.00 SYC67 - 910 : 50.00 -4.00 SYC68 - 911 : 50.00 -6.00 SYC69 - 912 : 50.00 -8.00 SYC70 - 913 : 50.00 -10.00 SYC71 - 914 : -117.80 32.33 SITE3 - 915 : -117.80 32.80 SITEX - 916 : -80.60 28.61 CAPC - 917 : -75.76 32.52 ATX1 - 918 : -75.57 32.67 ATX2 - 919 : -80.25 28.25 CAP1 - 920 : -80.23 28.85 CAPN - 921 : -80.60 28.80 DAB - 922 : -80.25 30.92 JAX - 923 : -83.75 28.50 TPA - 924 : -84.20 29.25 TLH - 925 : -86.03 29.77 PAN - 926 : -87.50 29.80 PEN - 927 : -64.97 42.05 ECALE - 928 : -54.25 45.88 ECALL - 929 : -9.25 51.23 SHAN - 930 : -53.01 46.35 STJO - 931 : -75.05 33.50 CMFS - 932 : -143.64 -21.94 TAHI - 933 : -157.25 20.75 HNL51 - 934 : -157.81 20.16 HNL52 - 935 : -158.54 20.57 HNL53 - 936 : -159.27 20.98 HNL54 - 937 : -158.74 21.79 HNL55 - 938 : -158.17 22.67 HNL56 - 939 : -157.42 22.26 HNL57 - 940 : -156.67 21.87 HNL58 - 941 : -157.00 21.50 HNL59 - 942 : -161.12 21.34 HNL61 - 943 : -160.71 22.49 HNL62 - 944 : -160.29 23.65 HNL63 - 945 : -159.27 23.20 HNL64 - 946 : -158.25 22.75 HNL65 - 947 : -158.76 21.80 HNL66 - 948 : -159.24 20.82 HNL67 - 949 : -160.18 21.08 HNL68 - 950 : -81.30 30.67 CHS51 - 951 : -80.50 30.67 CHS52 - 952 : -79.80 30.67 CHS53 - 953 : -79.00 30.67 CHS54 - 954 : -78.10 30.67 CHS55 - 955 : -78.10 31.40 CHS56 - 956 : -78.10 32.13 CHS57 - 957 : -78.10 32.86 CHS58 - 958 : -78.10 33.58 CHS59 - 959 : -78.50 33.58 CHS60 - 960 : -78.90 33.58 CHS61 - 961 : -80.25 28.50 JAX51 - 962 : -79.50 28.50 JAX52 - 963 : -78.50 28.50 JAX53 - 964 : -78.50 29.75 JAX54 - 965 : -78.50 31.25 JAX55 - 966 : -78.50 32.50 JAX56 - 967 : -79.25 32.50 JAX57 - 968 : -80.00 32.50 JAX58 - 969 : -80.20 28.70 JXFL51 - 970 : -79.70 28.70 JXFL52 - 971 : -79.30 28.70 JXFL53 - 972 : -79.30 29.20 JXFL54 - 973 : -79.30 29.70 JXFL55 - 974 : -79.30 30.20 JXFL56 - 975 : -79.30 30.70 JXFL57 - 976 : -79.30 31.20 JXFL58 - 977 : -79.30 31.60 JXFL59 - 978 : -79.30 32.00 JXFL60 - 979 : -79.80 32.00 JXFL61 - 980 : -80.20 32.00 JXFL62 - 981 : -80.50 32.00 JXFL63 - 982 : -72.40 40.60 MNE51 - 983 : -72.40 40.20 MNE52 - 984 : -71.60 40.20 MNE53 - 985 : -70.80 40.20 MNE54 - 986 : -70.00 40.20 MNE55 - 987 : -69.20 40.20 MNE56 - 988 : -68.40 40.20 MNE57 - 989 : -67.60 40.20 MNE58 - 990 : -66.80 40.20 MNE59 - 991 : -66.80 41.00 MNE60 - 992 : -66.80 41.80 MNE61 - 993 : -66.80 42.60 MNE62 - 994 : -66.80 43.40 MNE63 - 995 : -66.80 44.20 MNE64 - 996 : -66.80 44.50 MNE65 - 997 : -68.00 17.00 SJU51 - 998 : -68.00 17.50 SJU52 - 999 : -68.00 18.00 SJU53 - 1000 : -68.00 18.50 SJU54 - 1001 : -68.00 19.00 SJU55 - 1002 : -68.00 19.50 SJU56 - 1003 : -67.50 19.50 SJU57 - 1004 : -67.00 19.50 SJU58 - 1005 : -66.50 19.50 SJU59 - 1006 : -66.00 19.50 SJU60 - 1007 : -65.50 19.50 SJU61 - 1008 : -65.00 19.50 SJU62 - 1009 : -64.50 19.50 SJU63 - 1010 : -64.00 19.50 SJU64 - 1011 : -64.00 19.00 SJU65 - 1012 : -64.00 18.50 SJU66 - 1013 : -64.00 18.00 SJU67 - 1014 : -64.00 17.50 SJU68 - 1015 : -64.00 17.00 SJU69 - 1016 : -64.50 17.00 SJU70 - 1017 : -65.00 17.00 SJU71 - 1018 : -65.50 17.00 SJU72 - 1019 : -66.00 17.00 SJU73 - 1020 : -66.50 17.00 SJU74 - 1021 : -67.00 17.00 SJU75 - 1022 : -67.50 17.00 SJU76 - 1023 : -68.00 17.00 SJU77 - 1024 : -65.50 31.50 BER51 - 1025 : -65.25 31.50 BER52 - 1026 : -65.00 31.50 BER53 - 1027 : -64.75 31.50 BER54 - 1028 : -64.50 31.50 BER55 - 1029 : -64.25 31.50 BER56 - 1030 : -64.00 31.50 BER57 - 1031 : -64.00 31.75 BER58 - 1032 : -64.00 32.00 BER59 - 1033 : -64.00 32.25 BER60 - 1034 : -64.00 32.50 BER61 - 1035 : -64.00 32.75 BER62 - 1036 : -64.00 33.00 BER63 - 1037 : -64.25 33.00 BER64 - 1038 : -64.50 33.00 BER65 - 1039 : -64.75 33.00 BER66 - 1040 : -64.00 33.00 BER67 - 1041 : -65.25 33.00 BER68 - 1042 : -65.50 33.00 BER69 - 1043 : -65.50 32.75 BER70 - 1044 : -65.50 32.50 BER71 - 1045 : -65.50 32.25 BER72 - 1046 : -65.50 32.00 BER73 - 1047 : -65.50 31.75 BER74 - 1048 : -97.00 23.00 SRH51 - 1049 : -96.00 23.00 SRH52 - 1050 : -95.00 23.00 SRH53 - 1051 : -94.00 23.00 SRH54 - 1052 : -93.00 23.00 SRH55 - 1053 : -92.00 23.00 SRH56 - 1054 : -91.00 23.00 SRH57 - 1055 : -90.00 23.00 SRH58 - 1056 : -89.00 23.00 SRH59 - 1057 : -88.00 23.00 SRH60 - 1058 : -87.00 23.00 SRH61 - 1059 : -86.00 23.00 SRH62 - 1060 : -85.00 23.00 SRH63 - 1061 : -79.00 23.00 SRH64 - 1062 : -78.00 23.00 SRH65 - 1063 : -78.00 28.00 SRH66 - 1064 : -78.00 29.00 SRH67 - 1065 : -78.00 30.00 SRH68 - 1066 : -78.00 31.00 SRH69 - 1067 : -78.00 32.00 SRH70 - 1068 : -78.00 33.00 SRH71 - 1069 : -96.91 25.26 BRO51 - 1070 : -96.41 25.26 BRO52 - 1071 : -95.91 25.26 BRO53 - 1072 : -95.41 25.26 BRO54 - 1073 : -95.41 25.76 BRO55 - 1074 : -95.41 26.26 BRO56 - 1075 : -95.41 26.76 BRO57 - 1076 : -95.41 27.26 BRO58 - 1077 : -95.41 27.76 BRO59 - 1078 : -95.41 28.26 BRO60 - 1079 : -96.87 27.00 HGX51 - 1080 : -96.37 27.00 HGX52 - 1081 : -95.87 27.00 HGX53 - 1082 : -95.37 27.00 HGX54 - 1083 : -94.87 27.00 HGX55 - 1084 : -94.37 27.00 HGX56 - 1085 : -93.87 27.00 HGX57 - 1086 : -93.37 27.00 HGX58 - 1087 : -93.37 27.50 HGX59 - 1088 : -93.37 28.00 HGX60 - 1089 : -93.37 28.50 HGX61 - 1090 : -93.37 29.00 HGX62 - 1091 : -97.02 25.75 CRP51 - 1092 : -96.73 25.75 CRP52 - 1093 : -96.48 25.75 CRP53 - 1094 : -96.16 25.76 CRP54 - 1095 : -95.77 25.73 CRP55 - 1096 : -95.39 25.73 CRP56 - 1097 : -95.15 25.75 CRP57 - 1098 : -95.14 25.98 CRP58 - 1099 : -95.13 26.35 CRP59 - 1100 : -95.13 26.68 CRP60 - 1101 : -95.11 27.00 CRP61 - 1102 : -95.10 27.35 CRP62 - 1103 : -95.09 27.76 CRP63 - 1104 : -95.09 28.10 CRP64 - 1105 : -95.08 28.36 CRP65 - 1106 : -95.07 28.64 CRP66 - 1107 : -95.06 28.86 CRP67 - 1108 : -95.05 29.09 CRP68 - 1109 : -97.00 26.00 CCTX51 - 1110 : -96.50 26.00 CCTX52 - 1111 : -96.00 26.00 CCTX53 - 1112 : -95.50 26.00 CCTX54 - 1113 : -95.00 26.00 CCTX55 - 1114 : -95.00 26.50 CCTX56 - 1115 : -95.00 27.00 CCTX57 - 1116 : -95.00 27.50 CCTX58 - 1117 : -95.00 28.00 CCTX59 - 1118 : -95.00 28.50 CCTX60 - 1119 : -95.00 28.80 CCTX61 - 1120 : -95.03 28.55 LCH51 - 1121 : -95.03 28.23 LCH52 - 1122 : -95.03 27.79 LCH53 - 1123 : -95.03 27.41 LCH54 - 1124 : -94.45 27.41 LCH55 - 1125 : -93.53 27.41 LCH56 - 1126 : -92.66 27.41 LCH57 - 1127 : -91.69 27.41 LCH58 - 1128 : -90.43 27.41 LCH59 - 1129 : -90.43 27.74 LCH60 - 1130 : -90.43 28.16 LCH61 - 1131 : -90.43 28.52 LCH62 - 1132 : -90.43 28.68 LCH63 - 1133 : -91.80 28.90 LIX51 - 1134 : -91.80 28.50 LIX52 - 1135 : -91.80 28.00 LIX53 - 1136 : -91.80 27.50 LIX54 - 1137 : -91.30 27.50 LIX55 - 1138 : -90.80 27.50 LIX56 - 1139 : -90.30 27.50 LIX57 - 1140 : -89.80 27.50 LIX58 - 1141 : -89.30 27.50 LIX59 - 1142 : -88.80 27.50 LIX60 - 1143 : -88.30 27.50 LIX61 - 1144 : -87.80 27.50 LIX62 - 1145 : -87.40 27.50 LIX63 - 1146 : -87.40 28.00 LIX64 - 1147 : -87.40 28.50 LIX65 - 1148 : -87.40 29.00 LIX66 - 1149 : -87.40 29.50 LIX67 - 1150 : -87.40 30.00 LIX68 - 1151 : -89.10 28.80 MOB51 - 1152 : -89.10 28.50 MOB52 - 1153 : -88.40 28.50 MOB53 - 1154 : -87.90 28.50 MOB54 - 1155 : -87.40 28.50 MOB55 - 1156 : -86.90 28.50 MOB56 - 1157 : -86.40 28.50 MOB57 - 1158 : -85.90 28.50 MOB58 - 1159 : -85.90 29.00 MOB59 - 1160 : -85.90 29.50 MOB60 - 1161 : -85.90 30.00 MOB61 - 1162 : -87.35 28.35 PCB51 - 1163 : -87.35 28.65 PCB52 - 1164 : -87.35 28.95 PCB53 - 1165 : -87.35 29.25 PCB54 - 1166 : -87.35 29.55 PCB55 - 1167 : -87.35 29.85 PCB56 - 1168 : -87.35 30.15 PCB57 - 1169 : -87.05 28.35 PCB58 - 1170 : -86.75 28.35 PCB59 - 1171 : -86.45 28.35 PCB60 - 1172 : -86.15 28.35 PCB61 - 1173 : -85.85 28.35 PCB62 - 1174 : -85.55 28.35 PCB63 - 1175 : -85.25 28.35 PCB64 - 1176 : -84.95 28.35 PCB65 - 1177 : -84.65 28.35 PCB66 - 1178 : -84.35 28.35 PCB67 - 1179 : -84.05 28.35 PCB68 - 1180 : -83.75 28.35 PCB69 - 1181 : -83.45 28.35 PCB70 - 1182 : -83.00 28.35 PCB71 - 1183 : -84.80 29.45 TBW51 - 1184 : -84.80 28.95 TBW52 - 1185 : -84.80 28.45 TBW53 - 1186 : -84.80 27.95 TBW54 - 1187 : -84.80 27.45 TBW55 - 1188 : -84.80 26.95 TBW56 - 1189 : -84.80 26.45 TBW57 - 1190 : -84.80 25.95 TBW58 - 1191 : -84.80 25.45 TBW59 - 1192 : -84.30 25.45 TBW60 - 1193 : -83.80 25.45 TBW61 - 1194 : -83.30 25.45 TBW62 - 1195 : -82.80 25.45 TBW63 - 1196 : -82.30 25.45 TBW64 - 1197 : -83.54 24.10 MIA51 - 1198 : -83.54 24.82 MIA52 - 1199 : -83.54 25.54 MIA53 - 1200 : -83.54 26.26 MIA54 - 1201 : -83.54 26.98 MIA55 - 1202 : -83.54 27.70 MIA56 - 1203 : -83.14 27.70 MIA57 - 1204 : -80.01 27.70 MIA58 - 1205 : -79.61 27.70 MIA59 - 1206 : -79.21 27.70 MIA60 - 1207 : -78.81 27.70 MIA61 - 1208 : -78.41 27.70 MIA62 - 1209 : -78.41 27.40 MIA63 - 1210 : -78.41 26.50 MIA64 - 1211 : -78.41 26.20 MIA65 - 1212 : -78.41 25.90 MIA66 - 1213 : -78.41 24.10 MIA67 - 1214 : -78.92 24.10 MIA68 - 1215 : -79.44 24.10 MIA69 - 1216 : -79.95 24.10 MIA70 - 1217 : -80.46 24.10 MIA71 - 1218 : -80.97 24.10 MIA72 - 1219 : -81.49 24.10 MIA73 - 1220 : -82.00 24.10 MIA74 - 1221 : -82.51 24.10 MIA75 - 1222 : -83.03 24.10 MIA76 - 1223 : -81.00 30.00 MLB51 - 1224 : -80.60 30.00 MLB52 - 1225 : -80.20 30.00 MLB53 - 1226 : -79.80 30.00 MLB54 - 1227 : -79.40 30.00 MLB55 - 1228 : -79.00 30.00 MLB56 - 1229 : -78.51 30.00 MLB57 - 1230 : -78.51 29.70 MLB58 - 1231 : -78.51 29.40 MLB59 - 1232 : -78.51 29.10 MLB60 - 1233 : -78.51 28.80 MLB61 - 1234 : -78.51 28.50 MLB62 - 1235 : -78.51 28.20 MLB63 - 1236 : -78.51 27.90 MLB64 - 1237 : -78.51 27.60 MLB65 - 1238 : -78.51 27.30 MLB66 - 1239 : -79.90 26.50 MLB67 - 1240 : -79.70 26.50 MLB68 - 1241 : -79.50 26.50 MLB69 - 1242 : -79.30 26.50 MLB70 - 1243 : -79.10 26.50 MLB71 - 1244 : -78.90 26.50 MLB72 - 1245 : -83.50 26.00 KEY51 - 1246 : -83.50 25.75 KEY52 - 1247 : -83.50 25.50 KEY53 - 1248 : -83.50 25.25 KEY54 - 1249 : -83.50 25.00 KEY55 - 1250 : -83.50 24.75 KEY56 - 1251 : -83.50 24.50 KEY57 - 1252 : -83.50 24.25 KEY58 - 1253 : -83.50 24.00 KEY59 - 1254 : -83.50 23.75 KEY60 - 1255 : -83.50 23.50 KEY61 - 1256 : -83.50 23.25 KEY62 - 1257 : -83.50 23.00 KEY63 - 1258 : -79.80 23.00 KEY64 - 1259 : -79.60 23.00 KEY65 - 1260 : -79.40 23.00 KEY66 - 1261 : -79.20 23.00 KEY67 - 1262 : -79.00 23.00 KEY68 - 1263 : -79.00 23.20 KEY69 - 1264 : -79.00 23.40 KEY70 - 1265 : -79.00 26.00 KEY71 - 1266 : -79.20 26.00 KEY72 - 1267 : -79.40 26.00 KEY73 - 1268 : -79.60 26.00 KEY74 - 1269 : -79.80 26.00 KEY75 - 1270 : -80.00 26.00 KEY76 - 1271 : -82.20 26.00 KEY77 - 1272 : -82.45 26.00 KEY78 - 1273 : -82.70 26.00 KEY79 - 1274 : -82.95 26.00 KEY80 - 1275 : -83.20 26.00 KEY81 - 1276 : -69.20 43.30 NW-CAR51 - 1277 : -68.70 43.30 NW-CAR52 - 1278 : -68.20 43.30 NW-CAR53 - 1279 : -67.70 43.30 NW-CAR54 - 1280 : -67.20 43.30 NW-CAR55 - 1281 : -66.70 43.30 NW-CAR56 - 1282 : -66.30 43.30 NW-CAR57 - 1283 : -66.30 43.55 NW-CAR58 - 1284 : -66.30 43.80 NW-CAR59 - 1285 : -66.30 44.05 NW-CAR60 - 1286 : -66.30 44.30 NW-CAR61 - 1287 : -66.30 44.55 NW-CAR62 - 1288 : -66.30 44.80 NW-CAR63 - 1289 : -66.30 45.05 NW-CAR64 - 1290 : -69.20 43.80 NW-CAR65 - 1291 : -69.20 43.55 NW-CAR66 - 1292 : -69.20 43.30 NW-CAR67 - 1293 : -72.50 40.45 NW-BOX51 - 1294 : -72.00 40.45 NW-BOX52 - 1295 : -71.50 40.45 NW-BOX53 - 1296 : -71.00 40.45 NW-BOX54 - 1297 : -70.50 40.45 NW-BOX55 - 1298 : -70.00 40.45 NW-BOX56 - 1299 : -69.50 40.45 NW-BOX57 - 1300 : -69.00 40.45 NW-BOX58 - 1301 : -68.85 40.45 NW-BOX59 - 1302 : -68.85 40.95 NW-BOX60 - 1303 : -68.85 41.45 NW-BOX61 - 1304 : -68.85 41.95 NW-BOX62 - 1305 : -68.85 42.45 NW-BOX63 - 1306 : -68.85 42.95 NW-BOX64 - 1307 : -68.85 43.25 NW-BOX65 - 1308 : -69.35 43.25 NW-BOX66 - 1309 : -69.85 43.25 NW-BOX67 - 1310 : -70.35 43.25 NW-BOX68 - 1311 : -70.48 43.25 NW-BOX69 - 1312 : -72.50 40.70 NW-BOX70 - 1313 : -72.50 41.20 NW-BOX71 - 1314 : -71.55 39.75 NW-NYC51 - 1315 : -72.05 39.75 NW-NYC52 - 1316 : -72.55 39.75 NW-NYC53 - 1317 : -73.05 39.75 NW-NYC54 - 1318 : -73.55 39.75 NW-NYC55 - 1319 : -74.05 39.75 NW-NYC56 - 1320 : -71.05 39.75 NW-NYC57 - 1321 : -71.05 40.25 NW-NYC58 - 1322 : -71.05 40.75 NW-NYC59 - 1323 : -71.05 41.25 NW-NYC60 - 1324 : -71.05 41.38 NW-NYC61 - 1325 : -75.10 38.20 NW-PHI51 - 1326 : -74.75 38.20 NW-PHI52 - 1327 : -74.25 38.20 NW-PHI53 - 1328 : -73.75 38.20 NW-PHI54 - 1329 : -73.35 38.20 NW-PHI55 - 1330 : -73.35 38.60 NW-PHI56 - 1331 : -73.35 39.10 NW-PHI57 - 1332 : -73.35 39.60 NW-PHI58 - 1333 : -73.35 40.10 NW-PHI59 - 1334 : -73.35 40.60 NW-PHI60 - 1335 : -75.40 36.75 NW-LWX51 - 1336 : -75.65 36.75 NW-LWX52 - 1337 : -75.90 36.75 NW-LWX53 - 1338 : -75.40 37.00 NW-LWX54 - 1339 : -75.40 37.25 NW-LWX55 - 1340 : -75.40 37.50 NW-LWX56 - 1341 : -75.40 37.75 NW-LWX57 - 1342 : -80.15 32.50 NW-ILM51 - 1343 : -79.90 32.50 NW-ILM52 - 1344 : -79.40 32.50 NW-ILM53 - 1345 : -78.90 32.50 NW-ILM54 - 1346 : -78.40 32.50 NW-ILM55 - 1347 : -77.90 32.50 NW-ILM56 - 1348 : -77.40 32.50 NW-ILM57 - 1349 : -76.90 32.50 NW-ILM58 - 1350 : -76.50 32.50 NW-ILM59 - 1351 : -76.50 33.00 NW-ILM60 - 1352 : -76.50 33.50 NW-ILM61 - 1353 : -76.50 34.00 NW-ILM62 - 1354 : -76.50 34.50 NW-ILM63 - 1355 : -70.80 42.40 NW-GYX51 - 1356 : -70.36 42.40 NW-GYX52 - 1357 : -69.86 42.40 NW-GYX53 - 1358 : -69.36 42.40 NW-GYX54 - 1359 : -68.86 42.40 NW-GYX55 - 1360 : -68.36 42.40 NW-GYX56 - 1361 : -67.86 42.40 NW-GYX57 - 1362 : -67.85 42.40 NW-GYX58 - 1363 : -67.85 42.90 NW-GYX59 - 1364 : -67.85 43.40 NW-GYX60 - 1365 : -67.85 43.90 NW-GYX61 - 1366 : -67.85 44.40 NW-GYX62 - 1367 : -77.91 33.85 NW-MHX51 - 1368 : -77.41 33.85 NW-MHX52 - 1369 : -76.91 33.85 NW-MHX53 - 1370 : -76.41 33.85 NW-MHX54 - 1371 : -75.91 33.85 NW-MHX55 - 1372 : -75.41 33.85 NW-MHX56 - 1373 : -74.91 33.85 NW-MHX57 - 1374 : -74.75 33.85 NW-MHX58 - 1375 : -74.75 34.35 NW-MHX59 - 1376 : -74.75 34.85 NW-MHX60 - 1377 : -74.75 35.35 NW-MHX61 - 1378 : -74.75 35.85 NW-MHX62 - 1379 : -74.75 36.35 NW-MHX63 - 1380 : -74.75 36.60 NW-MHX64 - 1381 : -75.25 36.60 NW-MHX65 - 1382 : -75.75 36.60 NW-MHX66 - 1383 : -75.83 36.60 NW-MHX67 - 1384 : -74.25 39.50 NW-AKQ51 - 1385 : -74.25 39.00 NW-AKQ52 - 1386 : -74.25 38.50 NW-AKQ53 - 1387 : -74.25 38.00 NW-AKQ54 - 1388 : -74.25 37.50 NW-AKQ55 - 1389 : -74.25 37.00 NW-AKQ56 - 1390 : -74.25 36.50 NW-AKQ57 - 1391 : -74.25 36.00 NW-AKQ58 - 1392 : -74.25 35.80 NW-AKQ59 - 1393 : -74.50 35.80 NW-AKQ60 - 1394 : -75.00 35.80 NW-AKQ61 - 1395 : -75.50 35.80 NW-AKQ62 - 1396 : -119.00 32.08 NW-SGX51 - 1397 : -119.00 32.33 NW-SGX52 - 1398 : -119.00 32.58 NW-SGX53 - 1399 : -119.00 32.83 NW-SGX54 - 1400 : -119.00 33.08 NW-SGX55 - 1401 : -119.00 33.33 NW-SGX56 - 1402 : -119.00 33.58 NW-SGX57 - 1403 : -119.00 33.82 NW-SGX58 - 1404 : -118.75 33.82 NW-SGX59 - 1405 : -118.50 33.82 NW-SGX60 - 1406 : -116.93 32.08 NW-SGX61 - 1407 : -117.00 32.08 NW-SGX62 - 1408 : -117.25 32.08 NW-SGX63 - 1409 : -117.50 32.08 NW-SGX64 - 1410 : -117.75 32.08 NW-SGX65 - 1411 : -118.00 32.08 NW-SGX66 - 1412 : -118.25 32.08 NW-SGX67 - 1413 : -118.50 32.08 NW-SGX68 - 1414 : -118.75 32.08 NW-SGX69 - 1415 : -123.36 32.60 NW-LOX51 - 1416 : -123.36 33.10 NW-LOX52 - 1417 : -123.36 33.60 NW-LOX53 - 1418 : -123.36 34.10 NW-LOX54 - 1419 : -123.36 34.60 NW-LOX55 - 1420 : -123.36 35.10 NW-LOX56 - 1421 : -123.36 35.60 NW-LOX57 - 1422 : -123.36 36.10 NW-LOX58 - 1423 : -123.36 36.10 NW-LOX59 - 1424 : -122.86 36.10 NW-LOX60 - 1425 : -122.36 36.10 NW-LOX61 - 1426 : -121.86 36.10 NW-LOX62 - 1427 : -123.36 32.60 NW-LOX63 - 1428 : -122.86 32.60 NW-LOX64 - 1429 : -122.36 32.60 NW-LOX65 - 1430 : -121.86 32.60 NW-LOX66 - 1431 : -121.36 32.60 NW-LOX67 - 1432 : -120.86 32.60 NW-LOX68 - 1433 : -120.36 32.60 NW-LOX69 - 1434 : -119.86 32.60 NW-LOX70 - 1435 : -119.36 32.60 NW-LOX71 - 1436 : -118.86 32.60 NW-LOX72 - 1437 : -118.36 32.60 NW-LOX73 - 1438 : -117.86 32.60 NW-LOX74 - 1439 : -117.36 32.60 NW-LOX75 - 1440 : -125.80 39.40 NW-MTR51 - 1441 : -125.80 38.90 NW-MTR52 - 1442 : -125.80 38.40 NW-MTR53 - 1443 : -125.80 37.90 NW-MTR54 - 1444 : -125.80 37.40 NW-MTR55 - 1445 : -125.80 36.90 NW-MTR56 - 1446 : -125.80 36.40 NW-MTR57 - 1447 : -125.80 35.90 NW-MTR58 - 1448 : -125.80 35.40 NW-MTR59 - 1449 : -125.80 35.00 NW-MTR60 - 1450 : -125.30 35.00 NW-MTR61 - 1451 : -124.80 35.00 NW-MTR62 - 1452 : -124.30 35.00 NW-MTR63 - 1453 : -123.80 35.00 NW-MTR64 - 1454 : -123.30 35.00 NW-MTR65 - 1455 : -122.80 35.00 NW-MTR66 - 1456 : -122.30 35.00 NW-MTR67 - 1457 : -121.80 35.00 NW-MTR68 - 1458 : -121.30 35.00 NW-MTR69 - 1459 : -120.80 35.00 NW-MTR70 - 1460 : -125.80 39.40 NW-MTR71 - 1461 : -125.30 39.40 NW-MTR72 - 1462 : -124.80 39.40 NW-MTR73 - 1463 : -124.30 39.40 NW-MTR74 - 1464 : -124.05 39.40 NW-MTR75 - 1465 : -127.30 41.00 NW-MFR51 - 1466 : -127.30 41.50 NW-MFR52 - 1467 : -127.30 42.00 NW-MFR53 - 1468 : -127.30 42.50 NW-MFR54 - 1469 : -127.30 43.00 NW-MFR55 - 1470 : -127.30 43.50 NW-MFR56 - 1471 : -127.30 44.00 NW-MFR57 - 1472 : -127.30 44.45 NW-MFR58 - 1473 : -126.80 44.45 NW-MFR59 - 1474 : -126.30 44.45 NW-MFR60 - 1475 : -125.80 44.45 NW-MFR61 - 1476 : -125.30 44.45 NW-MFR62 - 1477 : -124.80 44.45 NW-MFR63 - 1478 : -124.30 44.45 NW-MFR64 - 1479 : -124.30 41.00 NW-MFR65 - 1480 : -124.80 41.00 NW-MFR66 - 1481 : -125.30 41.00 NW-MFR67 - 1482 : -125.80 41.00 NW-MFR68 - 1483 : -126.30 41.00 NW-MFR69 - 1484 : -126.80 41.00 NW-MFR70 - 1485 : -127.30 41.00 NW-MFR71 - 1486 : -126.28 43.50 NW-PQR51 - 1487 : -126.28 44.00 NW-PQR52 - 1488 : -126.28 44.50 NW-PQR53 - 1489 : -126.28 45.00 NW-PQR54 - 1490 : -126.28 45.50 NW-PQR55 - 1491 : -126.28 46.00 NW-PQR56 - 1492 : -126.28 46.50 NW-PQR57 - 1493 : -126.28 47.00 NW-PQR58 - 1494 : -126.28 47.15 NW-PQR59 - 1495 : -125.78 47.15 NW-PQR60 - 1496 : -125.28 47.15 NW-PQR61 - 1497 : -124.78 47.15 NW-PQR62 - 1498 : -124.28 47.15 NW-PQR63 - 1499 : -124.30 43.50 NW-PQR64 - 1500 : -124.80 43.50 NW-PQR65 - 1501 : -125.30 43.50 NW-PQR66 - 1502 : -125.80 43.50 NW-PQR67 - 1503 : -127.00 46.10 NW-SEW51 - 1504 : -127.00 46.60 NW-SEW52 - 1505 : -127.00 47.10 NW-SEW53 - 1506 : -127.00 47.60 NW-SEW54 - 1507 : -127.00 48.10 NW-SEW55 - 1508 : -127.00 48.60 NW-SEW56 - 1509 : -127.00 49.10 NW-SEW57 - 1510 : -127.00 49.42 NW-SEW58 - 1511 : -126.75 49.42 NW-SEW59 - 1512 : -124.16 46.10 NW-SEW60 - 1513 : -124.41 46.10 NW-SEW61 - 1514 : -124.91 46.10 NW-SEW62 - 1515 : -125.41 46.10 NW-SEW63 - 1516 : -125.91 46.10 NW-SEW64 - 1517 : -126.41 46.10 NW-SEW65 - 1518 : -126.91 46.10 NW-SEW66 - 1519 : -126.27 38.40 NW-EKA51 - 1520 : -126.27 38.90 NW-EKA52 - 1521 : -126.27 39.40 NW-EKA53 - 1522 : -126.27 39.90 NW-EKA54 - 1523 : -126.27 40.40 NW-EKA55 - 1524 : -126.27 40.90 NW-EKA56 - 1525 : -126.27 41.40 NW-EKA57 - 1526 : -126.27 41.90 NW-EKA58 - 1527 : -126.27 42.20 NW-EKA59 - 1528 : -125.77 42.20 NW-EKA60 - 1529 : -125.27 42.20 NW-EKA61 - 1530 : -124.77 42.20 NW-EKA62 - 1531 : -124.64 42.20 NW-EKA63 - 1532 : -123.70 38.40 NW-EKA64 - 1533 : -124.20 38.40 NW-EKA65 - 1534 : -124.70 38.40 NW-EKA66 - 1535 : -125.20 38.40 NW-EKA67 - 1536 : -125.70 38.40 NW-EKA68 - 1537 : -126.20 38.40 NW-EKA69 - 1538 : -126.27 38.40 NW-EKA70 - 1539 : -161.50 18.07 NW-HFO51 - 1540 : -161.50 18.57 NW-HFO52 - 1541 : -161.50 19.07 NW-HFO53 - 1542 : -161.50 19.57 NW-HFO54 - 1543 : -161.50 20.07 NW-HFO55 - 1544 : -161.50 20.57 NW-HFO56 - 1545 : -161.50 21.07 NW-HFO57 - 1546 : -161.50 21.57 NW-HFO58 - 1547 : -161.50 22.07 NW-HFO59 - 1548 : -161.50 22.57 NW-HFO60 - 1549 : -161.50 23.09 NW-HFO61 - 1550 : -161.50 23.09 NW-HFO62 - 1551 : -161.00 23.09 NW-HFO63 - 1552 : -160.50 23.09 NW-HFO64 - 1553 : -160.00 23.09 NW-HFO65 - 1554 : -159.50 23.09 NW-HFO66 - 1555 : -159.00 23.09 NW-HFO67 - 1556 : -158.50 23.09 NW-HFO68 - 1557 : -158.00 23.09 NW-HFO69 - 1558 : -157.50 23.09 NW-HFO70 - 1559 : -157.00 23.09 NW-HFO71 - 1560 : -156.50 23.09 NW-HFO72 - 1561 : -156.00 23.09 NW-HFO73 - 1562 : -155.50 23.09 NW-HFO74 - 1563 : -155.00 23.09 NW-HFO75 - 1564 : -154.50 23.09 NW-HFO76 - 1565 : -153.90 23.09 NW-HFO77 - 1566 : -153.90 23.00 NW-HFO78 - 1567 : -153.90 22.50 NW-HFO79 - 1568 : -153.90 22.00 NW-HFO80 - 1569 : -153.90 21.50 NW-HFO81 - 1570 : -153.90 21.00 NW-HFO82 - 1571 : -153.90 20.50 NW-HFO83 - 1572 : -153.90 20.00 NW-HFO84 - 1573 : -153.90 19.50 NW-HFO85 - 1574 : -153.90 19.00 NW-HFO86 - 1575 : -153.90 18.50 NW-HFO87 - 1576 : -153.90 18.07 NW-HFO88 - 1577 : -154.00 18.07 NW-HFO89 - 1578 : -154.50 18.07 NW-HFO90 - 1579 : -155.00 18.07 NW-HFO91 - 1580 : -155.50 18.07 NW-HFO92 - 1581 : -156.00 18.07 NW-HFO93 - 1582 : -156.50 18.07 NW-HFO94 - 1583 : -157.00 18.07 NW-HFO95 - 1584 : -157.50 18.07 NW-HFO96 - 1585 : -158.00 18.07 NW-HFO97 - 1586 : -158.50 18.07 NW-HFO98 - 1587 : -159.00 18.07 NW-HFO99 - 1588 : -159.50 18.07 NW-HFO100 - 1589 : -160.00 18.07 NW-HFO101 - 1590 : -160.50 18.07 NW-HFO102 - 1591 : -161.00 18.07 NW-HFO103 - 1592 : -144.50 53.90 NW-AJK51 - 1593 : -144.50 54.00 NW-AJK52 - 1594 : -144.50 54.50 NW-AJK53 - 1595 : -144.50 55.00 NW-AJK54 - 1596 : -144.50 55.50 NW-AJK55 - 1597 : -144.50 56.00 NW-AJK56 - 1598 : -144.50 56.50 NW-AJK57 - 1599 : -144.50 57.00 NW-AJK58 - 1600 : -144.50 57.50 NW-AJK59 - 1601 : -144.50 58.00 NW-AJK60 - 1602 : -144.50 58.50 NW-AJK61 - 1603 : -144.50 59.00 NW-AJK62 - 1604 : -144.50 59.50 NW-AJK63 - 1605 : -144.50 60.00 NW-AJK64 - 1606 : -131.00 53.90 NW-AJK65 - 1607 : -131.50 53.90 NW-AJK66 - 1608 : -133.50 53.90 NW-AJK67 - 1609 : -134.00 53.90 NW-AJK68 - 1610 : -134.50 53.90 NW-AJK69 - 1611 : -135.00 53.90 NW-AJK70 - 1612 : -135.50 53.90 NW-AJK71 - 1613 : -136.00 53.90 NW-AJK72 - 1614 : -136.50 53.90 NW-AJK73 - 1615 : -137.00 53.90 NW-AJK74 - 1616 : -137.50 53.90 NW-AJK75 - 1617 : -138.00 53.90 NW-AJK76 - 1618 : -138.50 53.90 NW-AJK77 - 1619 : -139.00 53.90 NW-AJK78 - 1620 : -139.50 53.90 NW-AJK79 - 1621 : -140.00 53.90 NW-AJK80 - 1622 : -140.50 53.90 NW-AJK81 - 1623 : -141.00 53.90 NW-AJK82 - 1624 : -141.50 53.90 NW-AJK83 - 1625 : -142.00 53.90 NW-AJK84 - 1626 : -142.50 53.90 NW-AJK85 - 1627 : -143.00 53.90 NW-AJK86 - 1628 : -143.50 53.90 NW-AJK87 - 1629 : -144.00 53.90 NW-AJK88 - 1630 : -144.50 53.90 NW-AJK89 - 1631 : -158.00 55.00 NW-AER51 - 1632 : -158.00 55.50 NW-AER52 - 1633 : -158.00 56.00 NW-AER53 - 1634 : -158.00 57.75 NW-AER54 - 1635 : -158.00 58.00 NW-AER55 - 1636 : -142.85 59.85 NW-AER56 - 1637 : -142.85 59.35 NW-AER57 - 1638 : -142.85 58.85 NW-AER58 - 1639 : -142.85 58.35 NW-AER59 - 1640 : -142.85 57.85 NW-AER60 - 1641 : -142.85 57.35 NW-AER61 - 1642 : -142.85 56.85 NW-AER62 - 1643 : -142.85 56.35 NW-AER63 - 1644 : -142.85 55.85 NW-AER64 - 1645 : -142.85 55.35 NW-AER65 - 1646 : -142.85 55.00 NW-AER66 - 1647 : -143.35 55.00 NW-AER67 - 1648 : -143.85 55.00 NW-AER68 - 1649 : -144.35 55.00 NW-AER69 - 1650 : -144.85 55.00 NW-AER70 - 1651 : -145.35 55.00 NW-AER71 - 1652 : -145.85 55.00 NW-AER72 - 1653 : -146.35 55.00 NW-AER73 - 1654 : -146.85 55.00 NW-AER74 - 1655 : -147.35 55.00 NW-AER75 - 1656 : -147.85 55.00 NW-AER76 - 1657 : -148.35 55.00 NW-AER77 - 1658 : -148.85 55.00 NW-AER78 - 1659 : -149.35 55.00 NW-AER79 - 1660 : -149.85 55.00 NW-AER80 - 1661 : -150.35 55.00 NW-AER81 - 1662 : -150.85 55.00 NW-AER82 - 1663 : -151.35 55.00 NW-AER83 - 1664 : -151.85 55.00 NW-AER84 - 1665 : -152.35 55.00 NW-AER85 - 1666 : -152.85 55.00 NW-AER86 - 1667 : -153.35 55.00 NW-AER87 - 1668 : -153.85 55.00 NW-AER88 - 1669 : -154.35 55.00 NW-AER89 - 1670 : -154.85 55.00 NW-AER90 - 1671 : -155.35 55.00 NW-AER91 - 1672 : -155.85 55.00 NW-AER92 - 1673 : -156.35 55.00 NW-AER93 - 1674 : -156.85 55.00 NW-AER94 - 1675 : -157.35 55.00 NW-AER95 - 1676 : -157.85 55.00 NW-AER96 - 1677 : -158.00 55.00 NW-AER97 - 1678 : -177.00 51.00 NW-ALU51 - 1679 : -176.50 51.00 NW-ALU52 - 1680 : -176.00 51.00 NW-ALU53 - 1681 : -175.50 51.00 NW-ALU54 - 1682 : -175.00 51.00 NW-ALU55 - 1683 : -174.50 51.00 NW-ALU56 - 1684 : -174.00 51.00 NW-ALU57 - 1685 : -173.50 51.00 NW-ALU58 - 1686 : -173.00 51.00 NW-ALU59 - 1687 : -172.50 51.00 NW-ALU60 - 1688 : -172.00 51.00 NW-ALU61 - 1689 : -171.50 51.00 NW-ALU62 - 1690 : -171.00 51.00 NW-ALU63 - 1691 : -170.50 51.00 NW-ALU64 - 1692 : -170.00 51.00 NW-ALU65 - 1693 : -169.50 51.00 NW-ALU66 - 1694 : -169.00 51.00 NW-ALU67 - 1695 : -168.50 51.00 NW-ALU68 - 1696 : -168.00 51.00 NW-ALU69 - 1697 : -167.50 51.00 NW-ALU70 - 1698 : -167.00 51.00 NW-ALU71 - 1699 : -166.50 51.00 NW-ALU72 - 1700 : -166.00 51.00 NW-ALU73 - 1701 : -165.50 51.00 NW-ALU74 - 1702 : -165.00 51.00 NW-ALU75 - 1703 : -164.50 51.00 NW-ALU76 - 1704 : -164.00 51.00 NW-ALU77 - 1705 : -163.50 51.00 NW-ALU78 - 1706 : -163.00 51.00 NW-ALU79 - 1707 : -162.50 51.00 NW-ALU80 - 1708 : -162.00 51.00 NW-ALU81 - 1709 : -161.50 51.00 NW-ALU82 - 1710 : -161.00 51.00 NW-ALU83 - 1711 : -160.50 51.00 NW-ALU84 - 1712 : -160.00 51.00 NW-ALU85 - 1713 : -159.50 51.00 NW-ALU86 - 1714 : -159.00 51.00 NW-ALU87 - 1715 : -158.50 51.00 NW-ALU88 - 1716 : -158.00 51.00 NW-ALU89 - 1717 : -157.50 51.00 NW-ALU90 - 1718 : -157.00 51.00 NW-ALU91 - 1719 : -156.50 51.00 NW-ALU92 - 1720 : -156.00 51.00 NW-ALU93 - 1721 : -155.50 51.00 NW-ALU94 - 1722 : -155.00 51.00 NW-ALU95 - 1723 : -154.50 51.00 NW-ALU96 - 1724 : -154.00 51.00 NW-ALU97 - 1725 : -153.50 51.00 NW-ALU98 - 1726 : -153.00 51.00 NW-ALU99 - 1727 : -152.40 51.00 NW-ALU100 - 1728 : -177.00 51.50 NW-ALU101 - 1729 : -177.00 52.00 NW-ALU102 - 1730 : -177.00 52.50 NW-ALU103 - 1731 : -177.00 53.00 NW-ALU104 - 1732 : -177.00 53.50 NW-ALU105 - 1733 : -177.00 54.00 NW-ALU106 - 1734 : -177.00 54.50 NW-ALU107 - 1735 : -177.00 55.00 NW-ALU108 - 1736 : -177.00 55.50 NW-ALU109 - 1737 : -177.00 56.00 NW-ALU110 - 1738 : -177.00 56.50 NW-ALU111 - 1739 : -177.00 57.00 NW-ALU112 - 1740 : -177.00 57.50 NW-ALU113 - 1741 : -177.00 58.00 NW-ALU114 - 1742 : -177.00 58.50 NW-ALU115 - 1743 : -177.00 59.00 NW-ALU116 - 1744 : -177.00 59.50 NW-ALU117 - 1745 : -177.00 60.00 NW-ALU118 - 1746 : -177.00 60.50 NW-ALU119 - 1747 : -177.00 61.00 NW-ALU120 - 1748 : -177.00 61.50 NW-ALU121 - 1749 : -177.00 62.00 NW-ALU122 - 1750 : -176.50 62.00 NW-ALU123 - 1751 : -176.00 62.00 NW-ALU124 - 1752 : -175.50 62.00 NW-ALU125 - 1753 : -175.00 62.00 NW-ALU126 - 1754 : -174.50 62.00 NW-ALU127 - 1755 : -174.00 62.00 NW-ALU128 - 1756 : -173.50 62.00 NW-ALU129 - 1757 : -173.00 62.00 NW-ALU130 - 1758 : -172.50 62.00 NW-ALU131 - 1759 : -172.00 62.00 NW-ALU132 - 1760 : -171.50 62.00 NW-ALU133 - 1761 : -171.00 62.00 NW-ALU134 - 1762 : -170.50 62.00 NW-ALU135 - 1763 : -170.00 62.00 NW-ALU136 - 1764 : -169.50 62.00 NW-ALU137 - 1765 : -169.00 62.00 NW-ALU138 - 1766 : -168.50 62.00 NW-ALU139 - 1767 : -168.00 62.00 NW-ALU140 - 1768 : -167.50 62.00 NW-ALU141 - 1769 : -167.00 62.00 NW-ALU142 - 1770 : -166.50 62.00 NW-ALU143 - 1771 : -166.00 62.00 NW-ALU144 - 1772 : -152.40 60.20 NW-ALU145 - 1773 : -152.40 59.70 NW-ALU146 - 1774 : -152.40 59.20 NW-ALU147 - 1775 : -152.40 58.70 NW-ALU148 - 1776 : -152.40 57.20 NW-ALU149 - 1777 : -152.40 56.70 NW-ALU150 - 1778 : -152.40 56.20 NW-ALU151 - 1779 : -152.40 55.70 NW-ALU152 - 1780 : -152.40 55.20 NW-ALU153 - 1781 : -152.40 54.70 NW-ALU154 - 1782 : -152.40 54.20 NW-ALU155 - 1783 : -152.40 53.70 NW-ALU156 - 1784 : -152.40 53.20 NW-ALU157 - 1785 : -152.40 52.70 NW-ALU158 - 1786 : -152.40 52.20 NW-ALU159 - 1787 : -152.40 51.70 NW-ALU160 - 1788 : -152.40 51.20 NW-ALU161 - 1789 : -152.40 51.00 NW-ALU162 - 1790 : -177.00 53.15 NW-ALU163 - 1791 : -176.50 53.15 NW-ALU164 - 1792 : -176.00 53.15 NW-ALU165 - 1793 : -175.50 53.15 NW-ALU166 - 1794 : -175.00 53.15 NW-ALU167 - 1795 : -174.50 53.15 NW-ALU168 - 1796 : -174.00 53.15 NW-ALU169 - 1797 : -173.50 53.15 NW-ALU170 - 1798 : -173.00 53.15 NW-ALU171 - 1799 : -172.50 53.15 NW-ALU172 - 1800 : -172.00 53.15 NW-ALU173 - 1801 : -171.50 53.15 NW-ALU174 - 1802 : -171.00 53.15 NW-ALU175 - 1803 : -170.50 53.15 NW-ALU176 - 1804 : -170.00 53.15 NW-ALU177 - 1805 : -169.50 53.15 NW-ALU178 - 1806 : -169.00 53.15 NW-ALU179 - 1807 : -168.18 53.15 NW-ALU180 - 1808 : -168.00 53.15 NW-ALU181 - 1809 : -167.50 53.15 NW-ALU182 - 1810 : -167.00 53.15 NW-ALU183 - 1811 : -166.50 53.15 NW-ALU184 - 1812 : -166.00 53.15 NW-ALU185 - 1813 : -165.50 53.15 NW-ALU186 - 1814 : -165.00 53.15 NW-ALU187 - 1815 : -164.50 53.15 NW-ALU188 - 1816 : -164.00 53.15 NW-ALU189 - 1817 : -163.50 53.15 NW-ALU190 - 1818 : -163.00 53.15 NW-ALU191 - 1819 : -162.50 53.15 NW-ALU192 - 1820 : -162.00 53.15 NW-ALU193 - 1821 : -161.50 53.15 NW-ALU194 - 1822 : -161.00 53.15 NW-ALU195 - 1823 : -160.50 53.15 NW-ALU196 - 1824 : -160.00 53.15 NW-ALU197 - 1825 : -159.50 53.15 NW-ALU198 - 1826 : -159.00 53.15 NW-ALU199 - 1827 : -158.50 53.15 NW-ALU200 - 1828 : -158.00 53.15 NW-ALU201 - 1829 : -157.50 53.15 NW-ALU202 - 1830 : -157.00 53.15 NW-ALU203 - 1831 : -156.50 53.15 NW-ALU204 - 1832 : -156.00 53.15 NW-ALU205 - 1833 : -155.50 53.15 NW-ALU206 - 1834 : -155.00 53.15 NW-ALU207 - 1835 : -154.50 53.15 NW-ALU208 - 1836 : -154.00 53.15 NW-ALU209 - 1837 : -153.50 53.15 NW-ALU210 - 1838 : -153.00 53.15 NW-ALU211 - 1839 : -152.40 53.15 NW-ALU212 - 1840 : -176.50 73.00 NW-AFG51 - 1841 : -176.00 73.00 NW-AFG52 - 1842 : -175.50 73.00 NW-AFG53 - 1843 : -175.00 73.00 NW-AFG54 - 1844 : -174.50 73.00 NW-AFG55 - 1845 : -174.00 73.00 NW-AFG56 - 1846 : -173.50 73.00 NW-AFG57 - 1847 : -173.00 73.00 NW-AFG58 - 1848 : -172.50 73.00 NW-AFG59 - 1849 : -172.00 73.00 NW-AFG60 - 1850 : -171.50 73.00 NW-AFG61 - 1851 : -171.00 73.00 NW-AFG62 - 1852 : -170.50 73.00 NW-AFG63 - 1853 : -170.00 73.00 NW-AFG64 - 1854 : -169.50 73.00 NW-AFG65 - 1855 : -169.00 73.00 NW-AFG66 - 1856 : -168.50 73.00 NW-AFG67 - 1857 : -168.00 73.00 NW-AFG68 - 1858 : -167.50 73.00 NW-AFG69 - 1859 : -167.00 73.00 NW-AFG70 - 1860 : -166.50 73.00 NW-AFG71 - 1861 : -166.00 73.00 NW-AFG72 - 1862 : -165.50 73.00 NW-AFG73 - 1863 : -165.00 73.00 NW-AFG74 - 1864 : -164.50 73.00 NW-AFG75 - 1865 : -164.00 73.00 NW-AFG76 - 1866 : -163.50 73.00 NW-AFG77 - 1867 : -163.00 73.00 NW-AFG78 - 1868 : -162.50 73.00 NW-AFG79 - 1869 : -162.00 73.00 NW-AFG80 - 1870 : -161.50 73.00 NW-AFG81 - 1871 : -161.00 73.00 NW-AFG82 - 1872 : -160.50 73.00 NW-AFG83 - 1873 : -160.00 73.00 NW-AFG84 - 1874 : -159.50 73.00 NW-AFG85 - 1875 : -159.00 73.00 NW-AFG86 - 1876 : -158.50 73.00 NW-AFG87 - 1877 : -158.00 73.00 NW-AFG88 - 1878 : -157.50 73.00 NW-AFG89 - 1879 : -157.00 73.00 NW-AFG90 - 1880 : -156.50 73.00 NW-AFG91 - 1881 : -156.00 73.00 NW-AFG92 - 1882 : -155.50 73.00 NW-AFG93 - 1883 : -155.00 73.00 NW-AFG94 - 1884 : -154.50 73.00 NW-AFG95 - 1885 : -154.00 73.00 NW-AFG96 - 1886 : -153.50 73.00 NW-AFG97 - 1887 : -153.00 73.00 NW-AFG98 - 1888 : -152.50 73.00 NW-AFG99 - 1889 : -152.00 73.00 NW-AFG100 - 1890 : -151.50 73.00 NW-AFG101 - 1891 : -151.00 73.00 NW-AFG102 - 1892 : -150.50 73.00 NW-AFG103 - 1893 : -150.00 73.00 NW-AFG104 - 1894 : -149.50 73.00 NW-AFG105 - 1895 : -149.00 73.00 NW-AFG106 - 1896 : -148.50 73.00 NW-AFG107 - 1897 : -148.00 73.00 NW-AFG108 - 1898 : -147.50 73.00 NW-AFG109 - 1899 : -147.00 73.00 NW-AFG110 - 1900 : -146.50 73.00 NW-AFG111 - 1901 : -146.00 73.00 NW-AFG112 - 1902 : -145.50 73.00 NW-AFG113 - 1903 : -145.00 73.00 NW-AFG114 - 1904 : -144.50 73.00 NW-AFG115 - 1905 : -144.00 73.00 NW-AFG116 - 1906 : -143.50 73.00 NW-AFG117 - 1907 : -143.00 73.00 NW-AFG118 - 1908 : -142.50 73.00 NW-AFG119 - 1909 : -142.00 73.00 NW-AFG120 - 1910 : -141.50 73.00 NW-AFG121 - 1911 : -141.00 73.00 NW-AFG122 - 1912 : -140.50 73.00 NW-AFG123 - 1913 : -140.00 73.00 NW-AFG124 - 1914 : -139.50 73.00 NW-AFG125 - 1915 : -139.00 73.00 NW-AFG126 - 1916 : -139.00 72.50 NW-AFG127 - 1917 : -139.00 72.00 NW-AFG128 - 1918 : -139.00 71.50 NW-AFG129 - 1919 : -139.00 71.00 NW-AFG130 - 1920 : -139.00 70.50 NW-AFG131 - 1921 : -139.00 70.00 NW-AFG132 - 1922 : -139.00 69.75 NW-AFG133 - 1923 : -176.50 72.50 NW-AFG134 - 1924 : -176.50 72.00 NW-AFG135 - 1925 : -176.50 71.50 NW-AFG136 - 1926 : -176.50 71.00 NW-AFG137 - 1927 : -176.50 70.50 NW-AFG138 - 1928 : -176.50 70.00 NW-AFG139 - 1929 : -176.50 69.50 NW-AFG140 - 1930 : -176.50 69.00 NW-AFG141 - 1931 : -176.50 68.50 NW-AFG142 - 1932 : -176.50 65.00 NW-AFG143 - 1933 : -176.50 64.50 NW-AFG144 - 1934 : -176.50 64.00 NW-AFG145 - 1935 : -176.50 63.50 NW-AFG146 - 1936 : -176.50 63.00 NW-AFG147 - 1937 : -176.50 62.50 NW-AFG148 - 1938 : -176.50 62.00 NW-AFG149 - 1939 : -176.50 61.50 NW-AFG150 - 1940 : -176.50 61.00 NW-AFG151 - 1941 : -176.00 61.00 NW-AFG152 - 1942 : -175.50 61.00 NW-AFG153 - 1943 : -175.00 61.00 NW-AFG154 - 1944 : -174.50 61.00 NW-AFG155 - 1945 : -174.00 61.00 NW-AFG156 - 1946 : -173.50 61.00 NW-AFG157 - 1947 : -173.00 61.00 NW-AFG158 - 1948 : -172.50 61.00 NW-AFG159 - 1949 : -172.00 61.00 NW-AFG160 - 1950 : -171.50 61.00 NW-AFG161 - 1951 : -171.00 61.00 NW-AFG162 - 1952 : -170.50 61.00 NW-AFG163 - 1953 : -170.00 61.00 NW-AFG164 - 1954 : -169.50 61.00 NW-AFG165 - 1955 : -169.00 61.00 NW-AFG166 - 1956 : -168.50 61.00 NW-AFG167 - 1957 : -168.00 61.00 NW-AFG168 - 1958 : -167.50 61.00 NW-AFG169 - 1959 : -167.00 61.00 NW-AFG170 - 1960 : -166.50 61.00 NW-AFG171 - 1961 : -166.00 61.00 NW-AFG172 - 1962 : 143.67 12.34 NW-GUM51 - 1963 : 143.67 12.84 NW-GUM52 - 1964 : 143.67 13.34 NW-GUM53 - 1965 : 143.67 13.84 NW-GUM54 - 1966 : 143.67 14.34 NW-GUM55 - 1967 : 143.67 14.84 NW-GUM56 - 1968 : 143.67 15.34 NW-GUM57 - 1969 : 143.67 15.84 NW-GUM58 - 1970 : 143.67 16.34 NW-GUM59 - 1971 : 143.67 16.81 NW-GUM60 - 1972 : 144.17 16.81 NW-GUM61 - 1973 : 144.67 16.81 NW-GUM62 - 1974 : 145.17 16.81 NW-GUM63 - 1975 : 145.67 16.81 NW-GUM64 - 1976 : 146.17 16.81 NW-GUM65 - 1977 : 146.67 16.81 NW-GUM66 - 1978 : 147.17 16.81 NW-GUM67 - 1979 : 147.67 16.81 NW-GUM68 - 1980 : 148.17 16.81 NW-GUM69 - 1981 : 148.29 16.81 NW-GUM70 - 1982 : 148.29 16.34 NW-GUM71 - 1983 : 148.29 15.84 NW-GUM72 - 1984 : 148.29 15.34 NW-GUM73 - 1985 : 148.29 14.84 NW-GUM74 - 1986 : 148.29 14.34 NW-GUM75 - 1987 : 148.29 13.84 NW-GUM76 - 1988 : 148.29 13.34 NW-GUM77 - 1989 : 148.29 12.84 NW-GUM78 - 1990 : 148.29 12.34 NW-GUM79 - 1991 : 147.79 12.34 NW-GUM80 - 1992 : 147.29 12.34 NW-GUM81 - 1993 : 146.79 12.34 NW-GUM82 - 1994 : 146.29 12.34 NW-GUM83 - 1995 : 145.79 12.34 NW-GUM84 - 1996 : 145.29 12.34 NW-GUM85 - 1997 : 144.79 12.34 NW-GUM86 - 1998 : 144.29 12.34 NW-GUM87 - 1999 : 143.79 12.34 NW-GUM88 - 2000 : 143.67 16.84 NW-GUM89 - 2001 : 143.67 17.34 NW-GUM90 - 2002 : 143.67 17.84 NW-GUM91 - 2003 : 143.67 18.34 NW-GUM92 - 2004 : 143.67 18.84 NW-GUM93 - 2005 : 143.67 19.34 NW-GUM94 - 2006 : 143.67 19.84 NW-GUM95 - 2007 : 143.67 20.00 NW-GUM96 - 2008 : 144.17 20.00 NW-GUM97 - 2009 : 144.67 20.00 NW-GUM98 - 2010 : 145.17 20.00 NW-GUM99 - 2011 : 145.67 20.00 NW-GUM100 - 2012 : 146.17 20.00 NW-GUM101 - 2013 : 146.67 20.00 NW-GUM102 - 2014 : 147.17 20.00 NW-GUM103 - 2015 : 147.67 20.00 NW-GUM104 - 2016 : 148.17 20.00 NW-GUM105 - 2017 : 148.29 20.00 NW-GUM106 - 2018 : 148.29 19.84 NW-GUM107 - 2019 : 148.29 19.34 NW-GUM108 - 2020 : 148.29 18.84 NW-GUM109 - 2021 : 148.29 18.34 NW-GUM110 - 2022 : 148.29 17.84 NW-GUM111 - 2023 : 148.29 17.34 NW-GUM112 - 2024 : 148.29 16.84 NW-GUM113 - 2025 : -172.68 -15.69 NW-PGG50 - 2026 : -172.68 -15.19 NW-PGG51 - 2027 : -172.68 -14.69 NW-PGG52 - 2028 : -172.68 -14.19 NW-PGG53 - 2029 : -172.68 -13.69 NW-PGG54 - 2030 : -172.68 -13.19 NW-PGG55 - 2031 : -172.68 -12.69 NW-PGG56 - 2032 : -172.68 -12.19 NW-PGG57 - 2033 : -172.68 -11.69 NW-PGG58 - 2034 : -172.68 -11.19 NW-PGG59 - 2035 : -172.68 -10.69 NW-PGG60 - 2036 : -172.68 -10.19 NW-PGG61 - 2037 : -172.68 -9.70 NW-PGG62 - 2038 : -172.18 -9.70 NW-PGG63 - 2039 : -171.68 -9.70 NW-PGG64 - 2040 : -171.18 -9.70 NW-PGG65 - 2041 : -170.68 -9.70 NW-PGG66 - 2042 : -170.18 -9.70 NW-PGG67 - 2043 : -169.68 -9.70 NW-PGG68 - 2044 : -169.18 -9.70 NW-PGG69 - 2045 : -168.68 -9.70 NW-PGG70 - 2046 : -168.18 -9.70 NW-PGG71 - 2047 : -167.68 -9.70 NW-PGG72 - 2048 : -167.00 -9.70 NW-PGG73 - 2049 : -167.00 -10.19 NW-PGG74 - 2050 : -167.00 -10.69 NW-PGG75 - 2051 : -167.00 -11.19 NW-PGG76 - 2052 : -167.00 -11.69 NW-PGG77 - 2053 : -167.00 -12.19 NW-PGG78 - 2054 : -167.00 -12.69 NW-PGG79 - 2055 : -167.00 -13.19 NW-PGG80 - 2056 : -167.00 -13.69 NW-PGG81 - 2057 : -167.00 -14.19 NW-PGG82 - 2058 : -167.00 -14.69 NW-PGG83 - 2059 : -167.00 -15.19 NW-PGG84 - 2060 : -167.00 -15.69 NW-PGG85 - 2061 : -167.68 -15.69 NW-PGG86 - 2062 : -168.18 -15.69 NW-PGG87 - 2063 : -168.68 -15.69 NW-PGG88 - 2064 : -169.18 -15.69 NW-PGG89 - 2065 : -169.68 -15.69 NW-PGG90 - 2066 : -170.18 -15.69 NW-PGG91 - 2067 : -170.68 -15.69 NW-PGG92 - 2068 : -171.18 -15.69 NW-PGG93 - 2069 : -171.68 -15.69 NW-PGG94 - 2070 : -172.18 -15.69 NW-PGG95 - 2071 : -80.00 27.50 NW-OPC51 - 2072 : -79.50 27.50 NW-OPC52 - 2073 : -79.00 27.50 NW-OPC53 - 2074 : -78.50 27.50 NW-OPC54 - 2075 : -78.00 27.50 NW-OPC55 - 2076 : -77.50 27.50 NW-OPC56 - 2077 : -77.00 27.50 NW-OPC57 - 2078 : -76.50 27.50 NW-OPC58 - 2079 : -76.00 27.50 NW-OPC59 - 2080 : -75.50 27.50 NW-OPC60 - 2081 : -75.00 27.50 NW-OPC61 - 2082 : -74.50 27.50 NW-OPC62 - 2083 : -74.00 27.50 NW-OPC63 - 2084 : -73.50 27.50 NW-OPC64 - 2085 : -73.00 27.50 NW-OPC65 - 2086 : -72.50 27.50 NW-OPC66 - 2087 : -72.00 27.50 NW-OPC67 - 2088 : -71.50 27.50 NW-OPC68 - 2089 : -71.00 27.50 NW-OPC69 - 2090 : -70.50 27.50 NW-OPC70 - 2091 : -70.00 27.50 NW-OPC71 - 2092 : -69.50 27.50 NW-OPC72 - 2093 : -69.00 27.50 NW-OPC73 - 2094 : -68.50 27.50 NW-OPC74 - 2095 : -68.00 27.50 NW-OPC75 - 2096 : -67.50 27.50 NW-OPC76 - 2097 : -67.00 27.50 NW-OPC77 - 2098 : -66.50 27.50 NW-OPC78 - 2099 : -66.00 27.50 NW-OPC79 - 2100 : -65.50 27.50 NW-OPC80 - 2101 : -65.00 27.50 NW-OPC81 - 2102 : -64.50 27.50 NW-OPC82 - 2103 : -64.20 27.50 NW-OPC83 - 2104 : -64.20 28.00 NW-OPC84 - 2105 : -64.20 28.50 NW-OPC85 - 2106 : -64.20 29.00 NW-OPC86 - 2107 : -64.20 29.50 NW-OPC87 - 2108 : -64.20 30.00 NW-OPC88 - 2109 : -64.20 30.50 NW-OPC89 - 2110 : -64.20 31.00 NW-OPC90 - 2111 : -64.20 31.50 NW-OPC91 - 2112 : -64.20 32.00 NW-OPC92 - 2113 : -64.20 32.50 NW-OPC93 - 2114 : -64.20 33.00 NW-OPC94 - 2115 : -64.20 33.50 NW-OPC95 - 2116 : -64.20 34.00 NW-OPC96 - 2117 : -64.20 34.50 NW-OPC97 - 2118 : -64.20 35.00 NW-OPC98 - 2119 : -64.20 35.50 NW-OPC99 - 2120 : -64.20 36.00 NW-OPC100 - 2121 : -64.20 36.50 NW-OPC101 - 2122 : -64.20 37.00 NW-OPC102 - 2123 : -64.20 37.50 NW-OPC103 - 2124 : -64.20 38.00 NW-OPC104 - 2125 : -64.20 38.50 NW-OPC105 - 2126 : -64.20 39.00 NW-OPC106 - 2127 : -64.20 39.50 NW-OPC107 - 2128 : -64.20 40.00 NW-OPC108 - 2129 : -64.20 40.50 NW-OPC109 - 2130 : -64.20 41.00 NW-OPC110 - 2131 : -64.20 41.50 NW-OPC111 - 2132 : -64.20 42.00 NW-OPC112 - 2133 : -64.20 42.50 NW-OPC113 - 2134 : -64.20 43.00 NW-OPC114 - 2135 : -64.20 43.50 NW-OPC115 - 2136 : -64.20 44.00 NW-OPC116 - 2137 : -66.00 44.80 NW-OPC117 - 2138 : -66.50 44.80 NW-OPC118 - 2139 : 40.00 -5.00 HWRFa-50 - 2140 : 42.00 -5.00 HWRFa-51 - 2141 : 44.00 -5.00 HWRFa-52 - 2142 : 46.00 -5.00 HWRFa-53 - 2143 : 48.00 -5.00 HWRFa-54 - 2144 : 50.00 -5.00 HWRFa-55 - 2145 : 52.00 -5.00 HWRFa-56 - 2146 : 54.00 -5.00 HWRFa-57 - 2147 : 56.00 -5.00 HWRFa-58 - 2148 : 58.00 -5.00 HWRFa-59 - 2149 : 60.00 -5.00 HWRFa-60 - 2150 : 62.00 -5.00 HWRFa-61 - 2151 : 64.00 -5.00 HWRFa-62 - 2152 : 66.00 -5.00 HWRFa-63 - 2153 : 68.00 -5.00 HWRFa-64 - 2154 : 70.00 -5.00 HWRFa-65 - 2155 : 72.00 -5.00 HWRFa-66 - 2156 : 74.00 -5.00 HWRFa-67 - 2157 : 76.00 -5.00 HWRFa-68 - 2158 : 78.00 -5.00 HWRFa-69 - 2159 : 80.00 -5.00 HWRFa-70 - 2160 : 82.00 -5.00 HWRFa-71 - 2161 : 84.00 -5.00 HWRFa-72 - 2162 : 86.00 -5.00 HWRFa-73 - 2163 : 88.00 -5.00 HWRFa-74 - 2164 : 90.00 -5.00 HWRFa-75 - 2165 : 92.00 -5.00 HWRFa-76 - 2166 : 94.00 -5.00 HWRFa-77 - 2167 : 96.00 -5.00 HWRFa-78 - 2168 : 98.00 -5.00 HWRFa-79 - 2169 : 100.00 -5.00 HWRFa-80 - 2170 : 102.00 -5.00 HWRFa-81 - 2171 : 105.00 -1.00 HWRFa-82 - 2172 : 105.00 1.00 HWRFa-83 - 2173 : 105.00 3.00 HWRFa-84 - 2174 : 105.00 5.00 HWRFa-85 - 2175 : 105.00 7.00 HWRFa-86 - 2176 : 125.35 1.50 HWRFc-50 - 2177 : 126.00 1.50 HWRFc-51 - 2178 : 127.00 1.50 HWRFc-52 - 2179 : 128.00 1.50 HWRFc-53 - 2180 : 129.00 1.50 HWRFc-54 - 2181 : 130.00 1.50 HWRFc-55 - 2182 : 131.00 1.50 HWRFc-56 - 2183 : 132.00 1.50 HWRFc-57 - 2184 : 133.00 1.50 HWRFc-58 - 2185 : 134.00 1.50 HWRFc-59 - 2186 : 135.00 1.50 HWRFc-60 - 2187 : 136.00 1.50 HWRFc-61 - 2188 : 137.00 1.50 HWRFc-62 - 2189 : 138.00 1.50 HWRFc-63 - 2190 : 139.00 1.50 HWRFc-64 - 2191 : 140.00 1.50 HWRFc-65 - 2192 : 141.00 1.50 HWRFc-66 - 2193 : 142.00 1.50 HWRFc-67 - 2194 : 143.00 1.50 HWRFc-68 - 2195 : 144.00 1.50 HWRFc-69 - 2196 : 145.00 1.50 HWRFc-70 - 2197 : 146.00 1.50 HWRFc-71 - 2198 : 147.00 1.50 HWRFc-72 - 2199 : 148.00 1.50 HWRFc-73 - 2200 : 149.00 1.50 HWRFc-74 - 2201 : 150.00 1.50 HWRFc-75 - 2202 : 151.00 1.50 HWRFc-76 - 2203 : 152.00 1.50 HWRFc-77 - 2204 : 153.00 1.50 HWRFc-78 - 2205 : 154.00 1.50 HWRFc-79 - 2206 : 155.00 1.50 HWRFc-80 - 2207 : 156.00 1.50 HWRFc-81 - 2208 : 157.00 1.50 HWRFc-82 - 2209 : 158.00 1.50 HWRFc-83 - 2210 : 159.00 1.50 HWRFc-84 - 2211 : 160.00 1.50 HWRFc-85 - 2212 : 161.00 1.50 HWRFc-86 - 2213 : 162.00 1.50 HWRFc-87 - 2214 : 163.00 1.50 HWRFc-88 - 2215 : 164.00 1.50 HWRFc-89 - 2216 : 165.00 1.50 HWRFc-90 - 2217 : 166.00 1.50 HWRFc-91 - 2218 : 167.00 1.50 HWRFc-92 - 2219 : 168.00 1.50 HWRFc-93 - 2220 : 169.00 1.50 HWRFc-94 - 2221 : 170.00 1.50 HWRFc-95 - 2222 : 171.00 1.50 HWRFc-96 - 2223 : 172.00 1.50 HWRFc-97 - 2224 : 173.00 1.50 HWRFc-98 - 2225 : 174.00 1.50 HWRFc-99 - 2226 : 175.00 1.50 HWRFc-100 - 2227 : 176.00 1.50 HWRFc-101 - 2228 : 177.00 1.50 HWRFc-102 - 2229 : 178.00 1.50 HWRFc-103 - 2230 : 179.00 1.50 HWRFc-104 - 2231 : 180.00 1.50 HWRFc-105 - 2232 : 181.00 1.50 HWRFc-106 - 2233 : 182.00 1.50 HWRFc-107 - 2234 : 183.00 1.50 HWRFc-108 - 2235 : 184.00 1.50 HWRFc-109 - 2236 : 185.00 1.50 HWRFc-110 - 2237 : 186.00 1.50 HWRFc-111 - 2238 : 187.00 1.50 HWRFc-112 - 2239 : 188.00 1.50 HWRFc-113 - 2240 : 189.00 1.50 HWRFc-114 - 2241 : 190.00 1.50 HWRFc-115 - 2242 : 191.00 1.50 HWRFc-116 - 2243 : 192.00 1.50 HWRFc-117 - 2244 : 193.00 1.50 HWRFc-118 - 2245 : 194.00 1.50 HWRFc-119 - 2246 : 195.00 1.50 HWRFc-120 - 2247 : 196.00 1.50 HWRFc-121 - 2248 : 197.00 1.50 HWRFc-122 - 2249 : 198.00 1.50 HWRFc-123 - 2250 : 199.00 1.50 HWRFc-124 - 2251 : 200.00 1.50 HWRFc-125 - 2252 : 201.00 1.50 HWRFc-126 - 2253 : 202.00 1.50 HWRFc-127 - 2254 : 203.00 1.50 HWRFc-128 - 2255 : 204.00 1.50 HWRFc-129 - 2256 : 205.00 1.50 HWRFc-130 - 2257 : 206.00 1.50 HWRFc-131 - 2258 : 207.00 1.50 HWRFc-132 - 2259 : 208.00 1.50 HWRFc-133 - 2260 : 209.00 1.50 HWRFc-134 - 2261 : 210.00 1.50 HWRFc-135 - 2262 : 211.00 1.50 HWRFc-136 - 2263 : 212.00 1.50 HWRFc-137 - 2264 : 213.00 1.50 HWRFc-138 - 2265 : 214.00 1.50 HWRFc-139 - 2266 : 215.00 1.50 HWRFc-140 - 2267 : 216.00 1.50 HWRFc-141 - 2268 : 217.00 1.50 HWRFc-142 - 2269 : 218.00 1.50 HWRFc-143 - 2270 : 219.00 1.50 HWRFc-144 - 2271 : 220.00 1.50 HWRFc-145 - 2272 : 220.00 2.50 HWRFc-146 - 2273 : 220.00 3.50 HWRFc-147 - 2274 : 220.00 4.50 HWRFc-148 - 2275 : 220.00 5.50 HWRFc-149 - 2276 : 220.00 6.50 HWRFc-150 - 2277 : 220.00 7.50 HWRFc-151 - 2278 : 220.00 8.50 HWRFc-152 - 2279 : 220.00 9.50 HWRFc-153 - 2280 : 220.00 10.50 HWRFc-154 - 2281 : 220.00 11.50 HWRFc-155 - 2282 : 220.00 12.50 HWRFc-156 - 2283 : 220.00 13.50 HWRFc-157 - 2284 : 220.00 14.50 HWRFc-158 - 2285 : 220.00 15.50 HWRFc-159 - 2286 : 220.00 16.50 HWRFc-160 - 2287 : 220.00 17.50 HWRFc-161 - 2288 : 220.00 18.50 HWRFc-162 - 2289 : 220.00 19.50 HWRFc-163 - 2290 : 220.00 20.50 HWRFc-164 - 2291 : 220.00 21.50 HWRFc-165 - 2292 : 220.00 22.50 HWRFc-166 - 2293 : 220.00 23.50 HWRFc-167 - 2294 : 220.00 24.50 HWRFc-168 - 2295 : 220.00 25.50 HWRFc-169 - 2296 : 220.00 26.50 HWRFc-170 - 2297 : 220.00 27.50 HWRFc-171 - 2298 : 220.00 28.50 HWRFc-172 - 2299 : 220.00 29.50 HWRFc-173 - 2300 : 220.00 30.50 HWRFc-174 - 2301 : 220.00 31.50 HWRFc-175 - 2302 : 220.00 32.50 HWRFc-176 - 2303 : 220.00 33.50 HWRFc-177 - 2304 : 220.00 34.50 HWRFc-178 - 2305 : 220.00 35.50 HWRFc-179 - 2306 : 220.00 36.50 HWRFc-180 - 2307 : 220.00 37.50 HWRFc-181 - 2308 : 220.00 38.50 HWRFc-182 - 2309 : 220.00 39.50 HWRFc-183 - 2310 : 220.00 40.50 HWRFc-184 - 2311 : 220.00 41.50 HWRFc-185 - 2312 : 219.00 41.50 HWRFc-186 - 2313 : 218.00 41.50 HWRFc-187 - 2314 : 217.00 41.50 HWRFc-188 - 2315 : 216.00 41.50 HWRFc-189 - 2316 : 215.00 41.50 HWRFc-190 - 2317 : 214.00 41.50 HWRFc-191 - 2318 : 213.00 41.50 HWRFc-192 - 2319 : 212.00 41.50 HWRFc-193 - 2320 : 211.00 41.50 HWRFc-194 - 2321 : 210.00 41.50 HWRFc-195 - 2322 : 209.00 41.50 HWRFc-196 - 2323 : 208.00 41.50 HWRFc-197 - 2324 : 207.00 41.50 HWRFc-198 - 2325 : 206.00 41.50 HWRFc-199 - 2326 : 205.00 41.50 HWRFc-200 - 2327 : 204.00 41.50 HWRFc-201 - 2328 : 203.00 41.50 HWRFc-202 - 2329 : 202.00 41.50 HWRFc-203 - 2330 : 201.00 41.50 HWRFc-204 - 2331 : 200.00 41.50 HWRFc-205 - 2332 : 199.00 41.50 HWRFc-206 - 2333 : 198.00 41.50 HWRFc-207 - 2334 : 197.00 41.50 HWRFc-208 - 2335 : 196.00 41.50 HWRFc-209 - 2336 : 195.00 41.50 HWRFc-210 - 2337 : 194.00 41.50 HWRFc-211 - 2338 : 193.00 41.50 HWRFc-212 - 2339 : 192.00 41.50 HWRFc-213 - 2340 : 191.00 41.50 HWRFc-214 - 2341 : 190.00 41.50 HWRFc-215 - 2342 : 189.00 41.50 HWRFc-216 - 2343 : 188.00 41.50 HWRFc-217 - 2344 : 187.00 41.50 HWRFc-218 - 2345 : 186.00 41.50 HWRFc-219 - 2346 : 185.00 41.50 HWRFc-220 - 2347 : 184.00 41.50 HWRFc-221 - 2348 : 183.00 41.50 HWRFc-222 - 2349 : 182.00 41.50 HWRFc-223 - 2350 : 181.00 41.50 HWRFc-224 - 2351 : 180.00 41.50 HWRFc-225 - 2352 : 179.00 41.50 HWRFc-226 - 2353 : 178.00 41.50 HWRFc-227 - 2354 : 177.00 41.50 HWRFc-228 - 2355 : 176.00 41.50 HWRFc-229 - 2356 : 175.00 41.50 HWRFc-230 - 2357 : 174.00 41.50 HWRFc-231 - 2358 : 173.00 41.50 HWRFc-232 - 2359 : 172.00 41.50 HWRFc-233 - 2360 : 171.00 41.50 HWRFc-234 - 2361 : 170.00 41.50 HWRFc-235 - 2362 : 169.00 41.50 HWRFc-236 - 2363 : 168.00 41.50 HWRFc-237 - 2364 : 167.00 41.50 HWRFc-238 - 2365 : 166.00 41.50 HWRFc-239 - 2366 : 165.00 41.50 HWRFc-240 - 2367 : 164.00 41.50 HWRFc-241 - 2368 : 163.00 41.50 HWRFc-242 - 2369 : 162.00 41.50 HWRFc-243 - 2370 : 161.00 41.50 HWRFc-244 - 2371 : 160.00 41.50 HWRFc-245 - 2372 : 159.00 41.50 HWRFc-246 - 2373 : 158.00 41.50 HWRFc-247 - 2374 : 157.00 41.50 HWRFc-248 - 2375 : 156.00 41.50 HWRFc-249 - 2376 : 155.00 41.50 HWRFc-250 - 2377 : 154.00 41.50 HWRFc-251 - 2378 : 153.00 41.50 HWRFc-252 - 2379 : 152.00 41.50 HWRFc-253 - 2380 : 151.00 41.50 HWRFc-254 - 2381 : 150.00 41.50 HWRFc-255 - 2382 : 149.00 41.50 HWRFc-256 - 2383 : 148.00 41.50 HWRFc-257 - 2384 : 147.00 41.50 HWRFc-258 - 2385 : 146.00 41.50 HWRFc-259 - 2386 : 145.00 41.50 HWRFc-260 - 2387 : 144.00 41.50 HWRFc-261 - 2388 : 143.00 41.50 HWRFc-262 - 2389 : 142.00 41.50 HWRFc-263 - 2390 : 141.00 41.50 HWRFc-264 - 2391 : 140.00 41.50 HWRFc-265 - 2392 : 139.00 41.50 HWRFc-266 - 2393 : 138.00 41.50 HWRFc-267 - 2394 : 137.00 41.50 HWRFc-268 - 2395 : 136.00 41.50 HWRFc-269 - 2396 : 135.00 41.50 HWRFc-270 - 2397 : 134.00 41.50 HWRFc-271 - 2398 : 133.00 41.50 HWRFc-272 - 2399 : 132.00 41.50 HWRFc-273 - 2400 : 125.00 39.50 HWRFc-274 - 2401 : 125.00 38.70 HWRFc-275 - 2402 : 125.00 37.50 HWRFc-276 - 2403 : 125.00 36.50 HWRFc-277 - 2404 : 125.00 35.50 HWRFc-278 - 2405 : 125.00 34.50 HWRFc-279 - 2406 : 125.00 33.50 HWRFc-280 - 2407 : 125.00 32.50 HWRFc-281 - 2408 : 125.00 31.50 HWRFc-282 - 2409 : 125.00 30.50 HWRFc-283 - 2410 : 125.00 29.50 HWRFc-284 - 2411 : 125.00 28.50 HWRFc-285 - 2412 : 125.00 27.50 HWRFc-286 - 2413 : 125.00 26.50 HWRFc-287 - 2414 : 125.00 25.50 HWRFc-288 - 2415 : 125.00 24.50 HWRFc-289 - 2416 : 125.00 23.50 HWRFc-290 - 2417 : 125.00 22.50 HWRFc-291 - 2418 : 125.00 21.50 HWRFc-292 - 2419 : 125.00 20.50 HWRFc-293 - 2420 : 125.00 19.50 HWRFc-294 - 2421 : 125.00 18.50 HWRFc-295 - 2422 : 125.00 17.50 HWRFc-296 - 2423 : 125.00 16.50 HWRFc-297 - 2424 : 125.00 15.50 HWRFc-298 - 2425 : 125.00 14.50 HWRFc-299 - 2426 : 125.00 13.50 HWRFc-300 - 2427 : 125.00 13.00 HWRFc-301 - 2428 : 125.00 12.60 HWRFc-302 - 2429 : 125.00 10.00 HWRFc-303 - 2430 : 125.00 9.50 HWRFc-304 - 2431 : 125.00 9.00 HWRFc-305 - 2432 : 125.00 5.75 HWRFc-306 - 2433 : 125.00 5.50 HWRFc-307 - 2434 : 125.00 5.00 HWRFc-308 - 2435 : 125.00 4.50 HWRFc-309 - 2436 : 125.00 3.50 HWRFc-310 - 2437 : 125.00 2.50 HWRFc-311 - 2438 : 180.00 1.50 HWRFe-50 - 2439 : 180.50 1.50 HWRFe-51 - 2440 : 181.00 1.50 HWRFe-52 - 2441 : 181.50 1.50 HWRFe-53 - 2442 : 182.00 1.50 HWRFe-54 - 2443 : 182.50 1.50 HWRFe-55 - 2444 : 183.00 1.50 HWRFe-56 - 2445 : 183.50 1.50 HWRFe-57 - 2446 : 184.00 1.50 HWRFe-58 - 2447 : 184.50 1.50 HWRFe-59 - 2448 : 185.00 1.50 HWRFe-60 - 2449 : 185.50 1.50 HWRFe-61 - 2450 : 186.00 1.50 HWRFe-62 - 2451 : 186.50 1.50 HWRFe-63 - 2452 : 187.00 1.50 HWRFe-64 - 2453 : 187.50 1.50 HWRFe-65 - 2454 : 188.00 1.50 HWRFe-66 - 2455 : 188.50 1.50 HWRFe-67 - 2456 : 189.00 1.50 HWRFe-68 - 2457 : 189.50 1.50 HWRFe-69 - 2458 : 190.00 1.50 HWRFe-70 - 2459 : 190.50 1.50 HWRFe-71 - 2460 : 191.00 1.50 HWRFe-72 - 2461 : 191.50 1.50 HWRFe-73 - 2462 : 192.00 1.50 HWRFe-74 - 2463 : 192.50 1.50 HWRFe-75 - 2464 : 193.00 1.50 HWRFe-76 - 2465 : 193.50 1.50 HWRFe-77 - 2466 : 194.00 1.50 HWRFe-78 - 2467 : 194.50 1.50 HWRFe-79 - 2468 : 195.00 1.50 HWRFe-80 - 2469 : 195.50 1.50 HWRFe-81 - 2470 : 196.00 1.50 HWRFe-82 - 2471 : 196.50 1.50 HWRFe-83 - 2472 : 197.00 1.50 HWRFe-84 - 2473 : 197.50 1.50 HWRFe-85 - 2474 : 198.00 1.50 HWRFe-86 - 2475 : 198.50 1.50 HWRFe-87 - 2476 : 199.00 1.50 HWRFe-88 - 2477 : 199.50 1.50 HWRFe-89 - 2478 : 200.00 1.50 HWRFe-90 - 2479 : 200.50 1.50 HWRFe-91 - 2480 : 201.00 1.50 HWRFe-92 - 2481 : 201.50 1.50 HWRFe-93 - 2482 : 202.00 1.50 HWRFe-94 - 2483 : 202.50 1.50 HWRFe-95 - 2484 : 203.00 1.50 HWRFe-96 - 2485 : 203.50 1.50 HWRFe-97 - 2486 : 204.00 1.50 HWRFe-98 - 2487 : 204.50 1.50 HWRFe-99 - 2488 : 205.00 1.50 HWRFe-100 - 2489 : 205.50 1.50 HWRFe-101 - 2490 : 206.00 1.50 HWRFe-102 - 2491 : 206.50 1.50 HWRFe-103 - 2492 : 207.00 1.50 HWRFe-104 - 2493 : 207.50 1.50 HWRFe-105 - 2494 : 208.00 1.50 HWRFe-106 - 2495 : 208.50 1.50 HWRFe-107 - 2496 : 209.00 1.50 HWRFe-108 - 2497 : 209.50 1.50 HWRFe-109 - 2498 : 210.00 1.50 HWRFe-110 - 2499 : 210.50 1.50 HWRFe-111 - 2500 : 211.00 1.50 HWRFe-112 - 2501 : 211.50 1.50 HWRFe-113 - 2502 : 212.00 1.50 HWRFe-114 - 2503 : 212.50 1.50 HWRFe-115 - 2504 : 213.00 1.50 HWRFe-116 - 2505 : 213.50 1.50 HWRFe-117 - 2506 : 214.00 1.50 HWRFe-118 - 2507 : 214.50 1.50 HWRFe-119 - 2508 : 215.00 1.50 HWRFe-120 - 2509 : 215.50 1.50 HWRFe-121 - 2510 : 216.00 1.50 HWRFe-122 - 2511 : 216.50 1.50 HWRFe-123 - 2512 : 217.00 1.50 HWRFe-124 - 2513 : 217.50 1.50 HWRFe-125 - 2514 : 218.00 1.50 HWRFe-126 - 2515 : 218.50 1.50 HWRFe-127 - 2516 : 219.00 1.50 HWRFe-128 - 2517 : 219.50 1.50 HWRFe-129 - 2518 : 220.00 1.50 HWRFe-130 - 2519 : 220.50 1.50 HWRFe-131 - 2520 : 221.00 1.50 HWRFe-132 - 2521 : 221.50 1.50 HWRFe-133 - 2522 : 222.00 1.50 HWRFe-134 - 2523 : 222.50 1.50 HWRFe-135 - 2524 : 223.00 1.50 HWRFe-136 - 2525 : 223.50 1.50 HWRFe-137 - 2526 : 224.00 1.50 HWRFe-138 - 2527 : 224.50 1.50 HWRFe-139 - 2528 : 225.00 1.50 HWRFe-140 - 2529 : 225.50 1.50 HWRFe-141 - 2530 : 226.00 1.50 HWRFe-142 - 2531 : 226.50 1.50 HWRFe-143 - 2532 : 227.00 1.50 HWRFe-144 - 2533 : 227.50 1.50 HWRFe-145 - 2534 : 228.00 1.50 HWRFe-146 - 2535 : 228.50 1.50 HWRFe-147 - 2536 : 229.00 1.50 HWRFe-148 - 2537 : 229.50 1.50 HWRFe-149 - 2538 : 230.00 1.50 HWRFe-150 - 2539 : 230.50 1.50 HWRFe-151 - 2540 : 231.00 1.50 HWRFe-152 - 2541 : 231.50 1.50 HWRFe-153 - 2542 : 232.00 1.50 HWRFe-154 - 2543 : 232.50 1.50 HWRFe-155 - 2544 : 233.00 1.50 HWRFe-156 - 2545 : 233.50 1.50 HWRFe-157 - 2546 : 234.00 1.50 HWRFe-158 - 2547 : 234.50 1.50 HWRFe-159 - 2548 : 235.00 1.50 HWRFe-160 - 2549 : 235.50 1.50 HWRFe-161 - 2550 : 236.00 1.50 HWRFe-162 - 2551 : 236.50 1.50 HWRFe-163 - 2552 : 237.00 1.50 HWRFe-164 - 2553 : 237.50 1.50 HWRFe-165 - 2554 : 238.00 1.50 HWRFe-166 - 2555 : 238.50 1.50 HWRFe-167 - 2556 : 239.00 1.50 HWRFe-168 - 2557 : 239.50 1.50 HWRFe-169 - 2558 : 240.00 1.50 HWRFe-170 - 2559 : 240.50 1.50 HWRFe-171 - 2560 : 241.00 1.50 HWRFe-172 - 2561 : 241.50 1.50 HWRFe-173 - 2562 : 242.00 1.50 HWRFe-174 - 2563 : 242.50 1.50 HWRFe-175 - 2564 : 243.00 1.50 HWRFe-176 - 2565 : 243.50 1.50 HWRFe-177 - 2566 : 244.00 1.50 HWRFe-178 - 2567 : 244.50 1.50 HWRFe-179 - 2568 : 245.00 1.50 HWRFe-180 - 2569 : 245.50 1.50 HWRFe-181 - 2570 : 246.00 1.50 HWRFe-182 - 2571 : 246.50 1.50 HWRFe-183 - 2572 : 247.00 1.50 HWRFe-184 - 2573 : 247.50 1.50 HWRFe-185 - 2574 : 248.00 1.50 HWRFe-186 - 2575 : 248.50 1.50 HWRFe-187 - 2576 : 249.00 1.50 HWRFe-188 - 2577 : 249.50 1.50 HWRFe-189 - 2578 : 250.00 1.50 HWRFe-190 - 2579 : 250.50 1.50 HWRFe-191 - 2580 : 251.00 1.50 HWRFe-192 - 2581 : 251.50 1.50 HWRFe-193 - 2582 : 252.00 1.50 HWRFe-194 - 2583 : 252.50 1.50 HWRFe-195 - 2584 : 253.00 1.50 HWRFe-196 - 2585 : 253.50 1.50 HWRFe-197 - 2586 : 254.00 1.50 HWRFe-198 - 2587 : 254.50 1.50 HWRFe-199 - 2588 : 255.00 1.50 HWRFe-200 - 2589 : 255.50 1.50 HWRFe-201 - 2590 : 256.00 1.50 HWRFe-202 - 2591 : 256.50 1.50 HWRFe-203 - 2592 : 257.00 1.50 HWRFe-204 - 2593 : 257.50 1.50 HWRFe-205 - 2594 : 258.00 1.50 HWRFe-206 - 2595 : 258.50 1.50 HWRFe-207 - 2596 : 259.00 1.50 HWRFe-208 - 2597 : 259.50 1.50 HWRFe-209 - 2598 : 260.00 1.50 HWRFe-210 - 2599 : 260.50 1.50 HWRFe-211 - 2600 : 261.00 1.50 HWRFe-212 - 2601 : 261.50 1.50 HWRFe-213 - 2602 : 262.00 1.50 HWRFe-214 - 2603 : 262.50 1.50 HWRFe-215 - 2604 : 263.00 1.50 HWRFe-216 - 2605 : 263.50 1.50 HWRFe-217 - 2606 : 264.00 1.50 HWRFe-218 - 2607 : 264.50 1.50 HWRFe-219 - 2608 : 265.00 1.50 HWRFe-220 - 2609 : 265.50 1.50 HWRFe-221 - 2610 : 266.00 1.50 HWRFe-222 - 2611 : 266.50 1.50 HWRFe-223 - 2612 : 267.00 1.50 HWRFe-224 - 2613 : 267.50 1.50 HWRFe-225 - 2614 : 268.00 1.50 HWRFe-226 - 2615 : 268.50 1.50 HWRFe-227 - 2616 : 269.00 1.50 HWRFe-228 - 2617 : 269.50 1.50 HWRFe-229 - 2618 : 270.00 1.50 HWRFe-230 - 2619 : 270.50 1.50 HWRFe-231 - 2620 : 271.00 1.50 HWRFe-232 - 2621 : 271.50 1.50 HWRFe-233 - 2622 : 272.00 1.50 HWRFe-234 - 2623 : 272.50 1.50 HWRFe-235 - 2624 : 273.00 1.50 HWRFe-236 - 2625 : 273.50 1.50 HWRFe-237 - 2626 : 274.00 1.50 HWRFe-238 - 2627 : 274.50 1.50 HWRFe-239 - 2628 : 275.00 1.50 HWRFe-240 - 2629 : 275.50 1.50 HWRFe-241 - 2630 : 276.00 1.50 HWRFe-242 - 2631 : 276.50 1.50 HWRFe-243 - 2632 : 277.00 1.50 HWRFe-244 - 2633 : 277.50 1.50 HWRFe-245 - 2634 : 278.00 1.50 HWRFe-246 - 2635 : 234.00 41.50 HWRFe-247 - 2636 : 233.50 41.50 HWRFe-248 - 2637 : 233.00 41.50 HWRFe-249 - 2638 : 232.50 41.50 HWRFe-250 - 2639 : 232.00 41.50 HWRFe-251 - 2640 : 231.50 41.50 HWRFe-252 - 2641 : 231.00 41.50 HWRFe-253 - 2642 : 230.50 41.50 HWRFe-254 - 2643 : 230.00 41.50 HWRFe-255 - 2644 : 229.50 41.50 HWRFe-256 - 2645 : 229.00 41.50 HWRFe-257 - 2646 : 228.50 41.50 HWRFe-258 - 2647 : 228.00 41.50 HWRFe-259 - 2648 : 227.50 41.50 HWRFe-260 - 2649 : 227.00 41.50 HWRFe-261 - 2650 : 226.50 41.50 HWRFe-262 - 2651 : 226.00 41.50 HWRFe-263 - 2652 : 225.50 41.50 HWRFe-264 - 2653 : 225.00 41.50 HWRFe-265 - 2654 : 224.50 41.50 HWRFe-266 - 2655 : 224.00 41.50 HWRFe-267 - 2656 : 223.50 41.50 HWRFe-268 - 2657 : 223.00 41.50 HWRFe-269 - 2658 : 222.50 41.50 HWRFe-270 - 2659 : 222.00 41.50 HWRFe-271 - 2660 : 221.50 41.50 HWRFe-272 - 2661 : 221.00 41.50 HWRFe-273 - 2662 : 220.50 41.50 HWRFe-274 - 2663 : 220.00 41.50 HWRFe-275 - 2664 : 219.50 41.50 HWRFe-276 - 2665 : 219.00 41.50 HWRFe-277 - 2666 : 218.50 41.50 HWRFe-278 - 2667 : 218.00 41.50 HWRFe-279 - 2668 : 217.50 41.50 HWRFe-280 - 2669 : 217.00 41.50 HWRFe-281 - 2670 : 216.50 41.50 HWRFe-282 - 2671 : 216.00 41.50 HWRFe-283 - 2672 : 215.50 41.50 HWRFe-284 - 2673 : 215.00 41.50 HWRFe-285 - 2674 : 214.50 41.50 HWRFe-286 - 2675 : 214.00 41.50 HWRFe-287 - 2676 : 213.50 41.50 HWRFe-288 - 2677 : 213.00 41.50 HWRFe-289 - 2678 : 212.50 41.50 HWRFe-290 - 2679 : 212.00 41.50 HWRFe-291 - 2680 : 211.50 41.50 HWRFe-292 - 2681 : 211.00 41.50 HWRFe-293 - 2682 : 210.50 41.50 HWRFe-294 - 2683 : 210.00 41.50 HWRFe-295 - 2684 : 209.50 41.50 HWRFe-296 - 2685 : 209.00 41.50 HWRFe-297 - 2686 : 208.50 41.50 HWRFe-298 - 2687 : 208.00 41.50 HWRFe-299 - 2688 : 207.50 41.50 HWRFe-300 - 2689 : 207.00 41.50 HWRFe-301 - 2690 : 206.50 41.50 HWRFe-302 - 2691 : 206.00 41.50 HWRFe-303 - 2692 : 205.50 41.50 HWRFe-304 - 2693 : 205.00 41.50 HWRFe-305 - 2694 : 204.50 41.50 HWRFe-306 - 2695 : 204.00 41.50 HWRFe-307 - 2696 : 203.50 41.50 HWRFe-308 - 2697 : 203.00 41.50 HWRFe-309 - 2698 : 202.50 41.50 HWRFe-310 - 2699 : 202.00 41.50 HWRFe-311 - 2700 : 201.50 41.50 HWRFe-312 - 2701 : 201.00 41.50 HWRFe-313 - 2702 : 200.50 41.50 HWRFe-314 - 2703 : 200.00 41.50 HWRFe-315 - 2704 : 199.50 41.50 HWRFe-316 - 2705 : 199.00 41.50 HWRFe-317 - 2706 : 198.50 41.50 HWRFe-318 - 2707 : 198.00 41.50 HWRFe-319 - 2708 : 197.50 41.50 HWRFe-320 - 2709 : 197.00 41.50 HWRFe-321 - 2710 : 196.50 41.50 HWRFe-322 - 2711 : 196.00 41.50 HWRFe-323 - 2712 : 195.50 41.50 HWRFe-324 - 2713 : 195.00 41.50 HWRFe-325 - 2714 : 194.50 41.50 HWRFe-326 - 2715 : 194.00 41.50 HWRFe-327 - 2716 : 193.50 41.50 HWRFe-328 - 2717 : 193.00 41.50 HWRFe-329 - 2718 : 192.50 41.50 HWRFe-330 - 2719 : 192.00 41.50 HWRFe-331 - 2720 : 191.50 41.50 HWRFe-332 - 2721 : 191.00 41.50 HWRFe-333 - 2722 : 190.50 41.50 HWRFe-334 - 2723 : 190.00 41.50 HWRFe-335 - 2724 : 189.50 41.50 HWRFe-336 - 2725 : 189.00 41.50 HWRFe-337 - 2726 : 188.50 41.50 HWRFe-338 - 2727 : 188.00 41.50 HWRFe-339 - 2728 : 187.50 41.50 HWRFe-340 - 2729 : 187.00 41.50 HWRFe-341 - 2730 : 186.50 41.50 HWRFe-342 - 2731 : 186.00 41.50 HWRFe-343 - 2732 : 185.50 41.50 HWRFe-344 - 2733 : 185.00 41.50 HWRFe-345 - 2734 : 184.50 41.50 HWRFe-346 - 2735 : 184.00 41.50 HWRFe-347 - 2736 : 183.50 41.50 HWRFe-348 - 2737 : 183.00 41.50 HWRFe-349 - 2738 : 182.50 41.50 HWRFe-350 - 2739 : 182.00 41.50 HWRFe-351 - 2740 : 181.50 41.50 HWRFe-352 - 2741 : 181.00 41.50 HWRFe-353 - 2742 : 180.50 41.50 HWRFe-354 - 2743 : 180.00 41.50 HWRFe-355 - 2744 : 180.00 41.00 HWRFe-356 - 2745 : 180.00 40.50 HWRFe-357 - 2746 : 180.00 40.00 HWRFe-358 - 2747 : 180.00 39.50 HWRFe-359 - 2748 : 180.00 39.00 HWRFe-360 - 2749 : 180.00 38.50 HWRFe-361 - 2750 : 180.00 38.00 HWRFe-362 - 2751 : 180.00 37.50 HWRFe-363 - 2752 : 180.00 37.00 HWRFe-364 - 2753 : 180.00 36.50 HWRFe-365 - 2754 : 180.00 36.00 HWRFe-366 - 2755 : 180.00 35.50 HWRFe-367 - 2756 : 180.00 35.00 HWRFe-368 - 2757 : 180.00 34.50 HWRFe-369 - 2758 : 180.00 34.00 HWRFe-370 - 2759 : 180.00 33.50 HWRFe-371 - 2760 : 180.00 33.00 HWRFe-372 - 2761 : 180.00 32.50 HWRFe-373 - 2762 : 180.00 32.00 HWRFe-374 - 2763 : 180.00 31.50 HWRFe-375 - 2764 : 180.00 31.00 HWRFe-376 - 2765 : 180.00 30.50 HWRFe-377 - 2766 : 180.00 30.00 HWRFe-378 - 2767 : 180.00 29.50 HWRFe-379 - 2768 : 180.00 29.00 HWRFe-380 - 2769 : 180.00 28.50 HWRFe-381 - 2770 : 180.00 28.00 HWRFe-382 - 2771 : 180.00 27.50 HWRFe-383 - 2772 : 180.00 27.00 HWRFe-384 - 2773 : 180.00 26.50 HWRFe-385 - 2774 : 180.00 26.00 HWRFe-386 - 2775 : 180.00 25.50 HWRFe-387 - 2776 : 180.00 25.00 HWRFe-388 - 2777 : 180.00 24.50 HWRFe-389 - 2778 : 180.00 24.00 HWRFe-390 - 2779 : 180.00 23.50 HWRFe-391 - 2780 : 180.00 23.00 HWRFe-392 - 2781 : 180.00 22.50 HWRFe-393 - 2782 : 180.00 22.00 HWRFe-394 - 2783 : 180.00 21.50 HWRFe-395 - 2784 : 180.00 21.00 HWRFe-396 - 2785 : 180.00 20.50 HWRFe-397 - 2786 : 180.00 20.00 HWRFe-398 - 2787 : 180.00 19.50 HWRFe-399 - 2788 : 180.00 19.00 HWRFe-400 - 2789 : 180.00 18.50 HWRFe-401 - 2790 : 180.00 18.00 HWRFe-402 - 2791 : 180.00 17.50 HWRFe-403 - 2792 : 180.00 17.00 HWRFe-404 - 2793 : 180.00 16.50 HWRFe-405 - 2794 : 180.00 16.00 HWRFe-406 - 2795 : 180.00 15.50 HWRFe-407 - 2796 : 180.00 15.00 HWRFe-408 - 2797 : 180.00 14.50 HWRFe-409 - 2798 : 180.00 14.00 HWRFe-410 - 2799 : 180.00 13.50 HWRFe-411 - 2800 : 180.00 13.00 HWRFe-412 - 2801 : 180.00 12.50 HWRFe-413 - 2802 : 180.00 12.00 HWRFe-414 - 2803 : 180.00 11.50 HWRFe-415 - 2804 : 180.00 11.00 HWRFe-416 - 2805 : 180.00 10.50 HWRFe-417 - 2806 : 180.00 10.00 HWRFe-418 - 2807 : 180.00 9.50 HWRFe-419 - 2808 : 180.00 9.00 HWRFe-420 - 2809 : 180.00 8.50 HWRFe-421 - 2810 : 180.00 8.00 HWRFe-422 - 2811 : 180.00 7.50 HWRFe-423 - 2812 : 180.00 7.00 HWRFe-424 - 2813 : 180.00 6.50 HWRFe-425 - 2814 : 180.00 6.00 HWRFe-426 - 2815 : 180.00 5.50 HWRFe-427 - 2816 : 180.00 5.00 HWRFe-428 - 2817 : 180.00 4.50 HWRFe-429 - 2818 : 180.00 4.00 HWRFe-430 - 2819 : 180.00 3.50 HWRFe-431 - 2820 : 180.00 3.00 HWRFe-432 - 2821 : 180.00 2.50 HWRFe-433 - 2822 : 180.00 2.00 HWRFe-434 - 2823 : 312.00 1.50 HWRFl-50 - 2824 : 312.50 1.50 HWRFl-51 - 2825 : 313.00 1.50 HWRFl-52 - 2826 : 313.50 1.50 HWRFl-53 - 2827 : 314.00 1.50 HWRFl-54 - 2828 : 314.50 1.50 HWRFl-55 - 2829 : 315.00 1.50 HWRFl-56 - 2830 : 315.50 1.50 HWRFl-57 - 2831 : 316.00 1.50 HWRFl-58 - 2832 : 316.50 1.50 HWRFl-59 - 2833 : 317.00 1.50 HWRFl-60 - 2834 : 317.50 1.50 HWRFl-61 - 2835 : 318.00 1.50 HWRFl-62 - 2836 : 318.50 1.50 HWRFl-63 - 2837 : 319.00 1.50 HWRFl-64 - 2838 : 319.50 1.50 HWRFl-65 - 2839 : 320.00 1.50 HWRFl-66 - 2840 : 320.50 1.50 HWRFl-67 - 2841 : 321.00 1.50 HWRFl-68 - 2842 : 321.50 1.50 HWRFl-69 - 2843 : 322.00 1.50 HWRFl-70 - 2844 : 322.50 1.50 HWRFl-71 - 2845 : 323.00 1.50 HWRFl-72 - 2846 : 323.50 1.50 HWRFl-73 - 2847 : 324.00 1.50 HWRFl-74 - 2848 : 324.50 1.50 HWRFl-75 - 2849 : 325.00 1.50 HWRFl-76 - 2850 : 325.50 1.50 HWRFl-77 - 2851 : 326.00 1.50 HWRFl-78 - 2852 : 326.50 1.50 HWRFl-79 - 2853 : 327.00 1.50 HWRFl-80 - 2854 : 327.50 1.50 HWRFl-81 - 2855 : 328.00 1.50 HWRFl-82 - 2856 : 328.50 1.50 HWRFl-83 - 2857 : 329.00 1.50 HWRFl-84 - 2858 : 329.50 1.50 HWRFl-85 - 2859 : 330.00 1.50 HWRFl-86 - 2860 : 330.50 1.50 HWRFl-87 - 2861 : 331.00 1.50 HWRFl-88 - 2862 : 331.50 1.50 HWRFl-89 - 2863 : 332.00 1.50 HWRFl-90 - 2864 : 332.50 1.50 HWRFl-91 - 2865 : 333.00 1.50 HWRFl-92 - 2866 : 333.50 1.50 HWRFl-93 - 2867 : 334.00 1.50 HWRFl-94 - 2868 : 334.50 1.50 HWRFl-95 - 2869 : 335.00 1.50 HWRFl-96 - 2870 : 335.50 1.50 HWRFl-97 - 2871 : 336.00 1.50 HWRFl-98 - 2872 : 336.50 1.50 HWRFl-99 - 2873 : 337.00 1.50 HWRFl-100 - 2874 : 337.50 1.50 HWRFl-101 - 2875 : 338.00 1.50 HWRFl-102 - 2876 : 338.50 1.50 HWRFl-103 - 2877 : 339.00 1.50 HWRFl-104 - 2878 : 339.50 1.50 HWRFl-105 - 2879 : 340.00 1.50 HWRFl-106 - 2880 : 340.50 1.50 HWRFl-107 - 2881 : 341.00 1.50 HWRFl-108 - 2882 : 341.50 1.50 HWRFl-109 - 2883 : 342.00 1.50 HWRFl-110 - 2884 : 342.50 1.50 HWRFl-111 - 2885 : 343.00 1.50 HWRFl-112 - 2886 : 343.50 1.50 HWRFl-113 - 2887 : 344.00 1.50 HWRFl-114 - 2888 : 344.50 1.50 HWRFl-115 - 2889 : 345.00 1.50 HWRFl-116 - 2890 : 345.50 1.50 HWRFl-117 - 2891 : 346.00 1.50 HWRFl-118 - 2892 : 346.50 1.50 HWRFl-119 - 2893 : 347.00 1.50 HWRFl-120 - 2894 : 347.50 1.50 HWRFl-121 - 2895 : 348.00 1.50 HWRFl-122 - 2896 : 348.50 1.50 HWRFl-123 - 2897 : 349.00 1.50 HWRFl-124 - 2898 : 349.50 1.50 HWRFl-125 - 2899 : 350.00 1.50 HWRFl-126 - 2900 : 350.50 1.50 HWRFl-127 - 2901 : 351.00 1.50 HWRFl-128 - 2902 : 351.50 1.50 HWRFl-129 - 2903 : 352.00 1.50 HWRFl-130 - 2904 : 352.00 2.00 HWRFl-131 - 2905 : 352.00 2.50 HWRFl-132 - 2906 : 352.00 3.00 HWRFl-133 - 2907 : 352.00 3.50 HWRFl-134 - 2908 : 352.00 45.50 HWRFl-135 - 2909 : 351.50 45.50 HWRFl-136 - 2910 : 351.00 45.50 HWRFl-137 - 2911 : 350.50 45.50 HWRFl-138 - 2912 : 350.00 45.50 HWRFl-139 - 2913 : 349.50 45.50 HWRFl-140 - 2914 : 349.00 45.50 HWRFl-141 - 2915 : 348.50 45.50 HWRFl-142 - 2916 : 348.00 45.50 HWRFl-143 - 2917 : 347.50 45.50 HWRFl-144 - 2918 : 347.00 45.50 HWRFl-145 - 2919 : 346.50 45.50 HWRFl-146 - 2920 : 346.00 45.50 HWRFl-147 - 2921 : 345.50 45.50 HWRFl-148 - 2922 : 345.00 45.50 HWRFl-149 - 2923 : 344.50 45.50 HWRFl-150 - 2924 : 344.00 45.50 HWRFl-151 - 2925 : 343.50 45.50 HWRFl-152 - 2926 : 343.00 45.50 HWRFl-153 - 2927 : 342.50 45.50 HWRFl-154 - 2928 : 342.00 45.50 HWRFl-155 - 2929 : 341.50 45.50 HWRFl-156 - 2930 : 341.00 45.50 HWRFl-157 - 2931 : 340.50 45.50 HWRFl-158 - 2932 : 340.00 45.50 HWRFl-159 - 2933 : 339.50 45.50 HWRFl-160 - 2934 : 339.00 45.50 HWRFl-161 - 2935 : 338.50 45.50 HWRFl-162 - 2936 : 338.00 45.50 HWRFl-163 - 2937 : 337.50 45.50 HWRFl-164 - 2938 : 337.00 45.50 HWRFl-165 - 2939 : 336.50 45.50 HWRFl-166 - 2940 : 336.00 45.50 HWRFl-167 - 2941 : 335.50 45.50 HWRFl-168 - 2942 : 335.00 45.50 HWRFl-169 - 2943 : 334.50 45.50 HWRFl-170 - 2944 : 334.00 45.50 HWRFl-171 - 2945 : 333.50 45.50 HWRFl-172 - 2946 : 333.00 45.50 HWRFl-173 - 2947 : 332.50 45.50 HWRFl-174 - 2948 : 332.00 45.50 HWRFl-175 - 2949 : 331.50 45.50 HWRFl-176 - 2950 : 331.00 45.50 HWRFl-177 - 2951 : 330.50 45.50 HWRFl-178 - 2952 : 330.00 45.50 HWRFl-179 - 2953 : 329.50 45.50 HWRFl-180 - 2954 : 329.00 45.50 HWRFl-181 - 2955 : 328.50 45.50 HWRFl-182 - 2956 : 328.00 45.50 HWRFl-183 - 2957 : 327.50 45.50 HWRFl-184 - 2958 : 327.00 45.50 HWRFl-185 - 2959 : 326.50 45.50 HWRFl-186 - 2960 : 326.00 45.50 HWRFl-187 - 2961 : 325.50 45.50 HWRFl-188 - 2962 : 325.00 45.50 HWRFl-189 - 2963 : 324.50 45.50 HWRFl-190 - 2964 : 324.00 45.50 HWRFl-191 - 2965 : 323.50 45.50 HWRFl-192 - 2966 : 323.00 45.50 HWRFl-193 - 2967 : 322.50 45.50 HWRFl-194 - 2968 : 322.00 45.50 HWRFl-195 - 2969 : 321.50 45.50 HWRFl-196 - 2970 : 321.00 45.50 HWRFl-197 - 2971 : 320.50 45.50 HWRFl-198 - 2972 : 320.00 45.50 HWRFl-199 - 2973 : 319.50 45.50 HWRFl-200 - 2974 : 319.00 45.50 HWRFl-201 - 2975 : 318.50 45.50 HWRFl-202 - 2976 : 318.00 45.50 HWRFl-203 - 2977 : 317.50 45.50 HWRFl-204 - 2978 : 317.00 45.50 HWRFl-205 - 2979 : 316.50 45.50 HWRFl-206 - 2980 : 316.00 45.50 HWRFl-207 - 2981 : 315.50 45.50 HWRFl-208 - 2982 : 315.00 45.50 HWRFl-209 - 2983 : 314.50 45.50 HWRFl-210 - 2984 : 314.00 45.50 HWRFl-211 - 2985 : 313.50 45.50 HWRFl-212 - 2986 : 313.00 45.50 HWRFl-213 - 2987 : 312.50 45.50 HWRFl-214 - 2988 : 312.00 45.50 HWRFl-215 - 2989 : 311.50 45.50 HWRFl-216 - 2990 : 311.00 45.50 HWRFl-217 - 2991 : 310.50 45.50 HWRFl-218 - 2992 : 310.00 45.50 HWRFl-219 - 2993 : 309.50 45.50 HWRFl-220 - 2994 : 309.00 45.50 HWRFl-221 - 2995 : 308.50 45.50 HWRFl-222 - 2996 : 308.00 45.50 HWRFl-223 - 2997 : 307.50 45.50 HWRFl-224 - 2998 : 307.00 45.50 HWRFl-225 - 2999 : 306.50 45.50 HWRFl-226 - 3000 : 306.00 45.50 HWRFl-227 - 3001 : 305.50 45.50 HWRFl-228 - 3002 : 305.00 45.50 HWRFl-229 - 3003 : 304.50 45.50 HWRFl-230 - 3004 : 304.00 45.50 HWRFl-231 - 3005 : 303.50 45.50 HWRFl-232 - 3006 : 303.00 45.50 HWRFl-233 - 3007 : 302.50 45.50 HWRFl-234 - 3008 : 302.00 45.50 HWRFl-235 - 3009 : 301.50 45.50 HWRFl-236 - 3010 : 301.00 45.50 HWRFl-237 - 3011 : 300.50 45.50 HWRFl-238 - 3012 : 300.00 45.50 HWRFl-239 - 3013 : 125.00 -42.50 HWRFp-50 - 3014 : 127.00 -42.50 HWRFp-51 - 3015 : 129.00 -42.50 HWRFp-52 - 3016 : 131.00 -42.50 HWRFp-53 - 3017 : 133.00 -42.50 HWRFp-54 - 3018 : 135.00 -42.50 HWRFp-55 - 3019 : 137.00 -42.50 HWRFp-56 - 3020 : 139.00 -42.50 HWRFp-57 - 3021 : 141.00 -42.50 HWRFp-58 - 3022 : 143.00 -42.50 HWRFp-59 - 3023 : 145.00 -42.50 HWRFp-60 - 3024 : 148.20 -42.50 HWRFp-61 - 3025 : 149.00 -42.50 HWRFp-62 - 3026 : 151.00 -42.50 HWRFp-63 - 3027 : 153.00 -42.50 HWRFp-64 - 3028 : 155.00 -42.50 HWRFp-65 - 3029 : 157.00 -42.50 HWRFp-66 - 3030 : 159.00 -42.50 HWRFp-67 - 3031 : 161.00 -42.50 HWRFp-68 - 3032 : 163.00 -42.50 HWRFp-69 - 3033 : 165.00 -42.50 HWRFp-70 - 3034 : 167.00 -42.50 HWRFp-71 - 3035 : 169.00 -42.50 HWRFp-72 - 3036 : 175.00 -42.50 HWRFp-73 - 3037 : 177.00 -42.50 HWRFp-74 - 3038 : 179.00 -42.50 HWRFp-75 - 3039 : 181.00 -42.50 HWRFp-76 - 3040 : 183.00 -42.50 HWRFp-77 - 3041 : 185.00 -42.50 HWRFp-78 - 3042 : 187.00 -42.50 HWRFp-79 - 3043 : 189.00 -42.50 HWRFp-80 - 3044 : 191.00 -42.50 HWRFp-81 - 3045 : 193.00 -42.50 HWRFp-82 - 3046 : 195.00 -42.50 HWRFp-83 - 3047 : 197.00 -42.50 HWRFp-84 - 3048 : 199.00 -42.50 HWRFp-85 - 3049 : 201.00 -42.50 HWRFp-86 - 3050 : 203.00 -42.50 HWRFp-87 - 3051 : 205.00 -42.50 HWRFp-88 - 3052 : 207.00 -42.50 HWRFp-89 - 3053 : 209.00 -42.50 HWRFp-90 - 3054 : 211.00 -42.50 HWRFp-91 - 3055 : 213.00 -42.50 HWRFp-92 - 3056 : 215.00 -42.50 HWRFp-93 - 3057 : 217.00 -42.50 HWRFp-94 - 3058 : 219.00 -42.50 HWRFp-95 - 3059 : 221.00 -42.50 HWRFp-96 - 3060 : 223.00 -42.50 HWRFp-97 - 3061 : 225.00 -42.50 HWRFp-98 - 3062 : 225.00 -40.50 HWRFp-99 - 3063 : 225.00 -38.50 HWRFp-100 - 3064 : 225.00 -36.50 HWRFp-101 - 3065 : 225.00 -34.50 HWRFp-102 - 3066 : 225.00 -32.50 HWRFp-103 - 3067 : 225.00 -30.50 HWRFp-104 - 3068 : 225.00 -28.50 HWRFp-105 - 3069 : 225.00 -26.50 HWRFp-106 - 3070 : 225.00 -24.50 HWRFp-107 - 3071 : 225.00 -22.50 HWRFp-108 - 3072 : 225.00 -20.50 HWRFp-109 - 3073 : 225.00 -18.50 HWRFp-110 - 3074 : 225.00 -16.50 HWRFp-111 - 3075 : 225.00 -14.50 HWRFp-112 - 3076 : 225.00 -12.50 HWRFp-113 - 3077 : 225.00 -10.50 HWRFp-114 - 3078 : 225.00 -8.50 HWRFp-115 - 3079 : 225.00 -6.50 HWRFp-116 - 3080 : 225.00 -4.50 HWRFp-117 - 3081 : 225.00 -2.50 HWRFp-118 - 3082 : 223.00 -2.50 HWRFp-119 - 3083 : 221.00 -2.50 HWRFp-120 - 3084 : 219.00 -2.50 HWRFp-121 - 3085 : 217.00 -2.50 HWRFp-122 - 3086 : 215.00 -2.50 HWRFp-123 - 3087 : 213.00 -2.50 HWRFp-124 - 3088 : 211.00 -2.50 HWRFp-125 - 3089 : 209.00 -2.50 HWRFp-126 - 3090 : 207.00 -2.50 HWRFp-127 - 3091 : 205.00 -2.50 HWRFp-128 - 3092 : 203.00 -2.50 HWRFp-129 - 3093 : 201.00 -2.50 HWRFp-130 - 3094 : 199.00 -2.50 HWRFp-131 - 3095 : 197.00 -2.50 HWRFp-132 - 3096 : 195.00 -2.50 HWRFp-133 - 3097 : 193.00 -2.50 HWRFp-134 - 3098 : 191.00 -2.50 HWRFp-135 - 3099 : 189.00 -2.50 HWRFp-136 - 3100 : 187.00 -2.50 HWRFp-137 - 3101 : 185.00 -2.50 HWRFp-138 - 3102 : 183.00 -2.50 HWRFp-139 - 3103 : 181.00 -2.50 HWRFp-140 - 3104 : 179.00 -2.50 HWRFp-141 - 3105 : 177.00 -2.50 HWRFp-142 - 3106 : 175.00 -2.50 HWRFp-143 - 3107 : 173.00 -2.50 HWRFp-144 - 3108 : 171.00 -2.50 HWRFp-145 - 3109 : 169.00 -2.50 HWRFp-146 - 3110 : 167.00 -2.50 HWRFp-147 - 3111 : 165.00 -2.50 HWRFp-148 - 3112 : 163.00 -2.50 HWRFp-149 - 3113 : 161.00 -2.50 HWRFp-150 - 3114 : 159.00 -2.50 HWRFp-151 - 3115 : 157.00 -2.50 HWRFp-152 - 3116 : 155.00 -2.50 HWRFp-153 - 3117 : 153.00 -2.50 HWRFp-154 - 3118 : 151.00 -2.50 HWRFp-155 - 3119 : 149.00 -2.50 HWRFp-156 - 3120 : 147.00 -2.50 HWRFp-157 - 3121 : 145.00 -2.50 HWRFp-158 - 3122 : 131.00 -2.50 HWRFp-159 - 3123 : 127.00 -2.50 HWRFp-160 - 3124 : 125.00 -2.50 HWRFp-161 - 3125 : 125.00 -4.50 HWRFp-162 - 3126 : 125.00 -6.50 HWRFp-163 - 3127 : 125.00 -10.50 HWRFp-164 - 3128 : 125.00 -12.50 HWRFp-165 - 3129 : 125.00 -34.50 HWRFp-166 - 3130 : 125.00 -36.50 HWRFp-167 - 3131 : 125.00 -38.50 HWRFp-168 - 3132 : 125.00 -40.50 HWRFp-169 - 3133 : 31.00 -42.50 HWRFs-50 - 3134 : 33.00 -42.50 HWRFs-51 - 3135 : 35.00 -42.50 HWRFs-52 - 3136 : 37.00 -42.50 HWRFs-53 - 3137 : 39.00 -42.50 HWRFs-54 - 3138 : 41.00 -42.50 HWRFs-55 - 3139 : 43.00 -42.50 HWRFs-56 - 3140 : 45.00 -42.50 HWRFs-57 - 3141 : 47.00 -42.50 HWRFs-58 - 3142 : 49.00 -42.50 HWRFs-59 - 3143 : 51.00 -42.50 HWRFs-60 - 3144 : 53.00 -42.50 HWRFs-61 - 3145 : 55.00 -42.50 HWRFs-62 - 3146 : 57.00 -42.50 HWRFs-63 - 3147 : 59.00 -42.50 HWRFs-64 - 3148 : 61.00 -42.50 HWRFs-65 - 3149 : 63.00 -42.50 HWRFs-66 - 3150 : 65.00 -42.50 HWRFs-67 - 3151 : 67.00 -42.50 HWRFs-68 - 3152 : 69.00 -42.50 HWRFs-69 - 3153 : 71.00 -42.50 HWRFs-70 - 3154 : 73.00 -42.50 HWRFs-71 - 3155 : 75.00 -42.50 HWRFs-72 - 3156 : 77.00 -42.50 HWRFs-73 - 3157 : 79.00 -42.50 HWRFs-74 - 3158 : 81.00 -42.50 HWRFs-75 - 3159 : 83.00 -42.50 HWRFs-76 - 3160 : 85.00 -42.50 HWRFs-77 - 3161 : 87.00 -42.50 HWRFs-78 - 3162 : 89.00 -42.50 HWRFs-79 - 3163 : 91.00 -42.50 HWRFs-80 - 3164 : 93.00 -42.50 HWRFs-81 - 3165 : 95.00 -42.50 HWRFs-82 - 3166 : 97.00 -42.50 HWRFs-83 - 3167 : 99.00 -42.50 HWRFs-84 - 3168 : 101.00 -42.50 HWRFs-85 - 3169 : 103.00 -42.50 HWRFs-86 - 3170 : 105.00 -42.50 HWRFs-87 - 3171 : 107.00 -42.50 HWRFs-88 - 3172 : 109.00 -42.50 HWRFs-89 - 3173 : 111.00 -42.50 HWRFs-90 - 3174 : 113.00 -42.50 HWRFs-91 - 3175 : 115.00 -42.50 HWRFs-92 - 3176 : 117.00 -42.50 HWRFs-93 - 3177 : 119.00 -42.50 HWRFs-94 - 3178 : 121.00 -42.50 HWRFs-95 - 3179 : 123.00 -42.50 HWRFs-96 - 3180 : 125.00 -42.50 HWRFs-97 - 3181 : 127.00 -42.50 HWRFs-98 - 3182 : 128.00 -42.50 HWRFs-99 - 3183 : 128.00 -40.50 HWRFs-100 - 3184 : 128.00 -38.50 HWRFs-101 - 3185 : 128.00 -36.50 HWRFs-102 - 3186 : 128.00 -34.50 HWRFs-103 - 3187 : 128.00 -12.50 HWRFs-104 - 3188 : 128.00 -10.50 HWRFs-105 - 3189 : 128.00 -8.50 HWRFs-106 - 3190 : 128.00 -6.50 HWRFs-107 - 3191 : 128.00 -4.50 HWRFs-108 - 3192 : 128.00 -2.50 HWRFs-109 - 3193 : 127.00 -2.50 HWRFs-110 - 3194 : 125.00 -2.50 HWRFs-111 - 3195 : 123.00 -2.50 HWRFs-112 - 3196 : 117.00 -2.50 HWRFs-113 - 3197 : 109.00 -2.50 HWRFs-114 - 3198 : 107.00 -2.50 HWRFs-115 - 3199 : 101.00 -2.50 HWRFs-116 - 3200 : 99.00 -2.50 HWRFs-117 - 3201 : 97.00 -2.50 HWRFs-118 - 3202 : 95.00 -2.50 HWRFs-119 - 3203 : 93.00 -2.50 HWRFs-120 - 3204 : 91.00 -2.50 HWRFs-121 - 3205 : 89.00 -2.50 HWRFs-122 - 3206 : 87.00 -2.50 HWRFs-123 - 3207 : 85.00 -2.50 HWRFs-124 - 3208 : 83.00 -2.50 HWRFs-125 - 3209 : 81.00 -2.50 HWRFs-126 - 3210 : 79.00 -2.50 HWRFs-127 - 3211 : 77.00 -2.50 HWRFs-128 - 3212 : 75.00 -2.50 HWRFs-129 - 3213 : 73.00 -2.50 HWRFs-130 - 3214 : 71.00 -2.50 HWRFs-131 - 3215 : 69.00 -2.50 HWRFs-132 - 3216 : 67.00 -2.50 HWRFs-133 - 3217 : 65.00 -2.50 HWRFs-134 - 3218 : 63.00 -2.50 HWRFs-135 - 3219 : 61.00 -2.50 HWRFs-136 - 3220 : 59.00 -2.50 HWRFs-137 - 3221 : 57.00 -2.50 HWRFs-138 - 3222 : 55.00 -2.50 HWRFs-139 - 3223 : 53.00 -2.50 HWRFs-140 - 3224 : 51.00 -2.50 HWRFs-141 - 3225 : 49.00 -2.50 HWRFs-142 - 3226 : 47.00 -2.50 HWRFs-143 - 3227 : 45.00 -2.50 HWRFs-144 - 3228 : 43.00 -2.50 HWRFs-145 - 3229 : 31.00 -32.50 HWRFs-146 - 3230 : 31.00 -34.50 HWRFs-147 - 3231 : 31.00 -36.50 HWRFs-148 - 3232 : 31.00 -38.50 HWRFs-149 - 3233 : 31.00 -40.50 HWRFs-150 - 3234 : 105.00 1.50 HWRFw-50 - 3235 : 107.00 1.50 HWRFw-51 - 3236 : 123.00 1.50 HWRFw-52 - 3237 : 127.00 1.50 HWRFw-53 - 3238 : 131.00 1.50 HWRFw-54 - 3239 : 133.00 1.50 HWRFw-55 - 3240 : 135.00 1.50 HWRFw-56 - 3241 : 137.00 1.50 HWRFw-57 - 3242 : 139.00 1.50 HWRFw-58 - 3243 : 141.00 1.50 HWRFw-59 - 3244 : 143.00 1.50 HWRFw-60 - 3245 : 145.00 1.50 HWRFw-61 - 3246 : 147.00 1.50 HWRFw-62 - 3247 : 149.00 1.50 HWRFw-63 - 3248 : 151.00 1.50 HWRFw-64 - 3249 : 153.00 1.50 HWRFw-65 - 3250 : 155.00 1.50 HWRFw-66 - 3251 : 157.00 1.50 HWRFw-67 - 3252 : 159.00 1.50 HWRFw-68 - 3253 : 161.00 1.50 HWRFw-69 - 3254 : 163.00 1.50 HWRFw-70 - 3255 : 165.00 1.50 HWRFw-71 - 3256 : 167.00 1.50 HWRFw-72 - 3257 : 169.00 1.50 HWRFw-73 - 3258 : 171.00 1.50 HWRFw-74 - 3259 : 173.00 1.50 HWRFw-75 - 3260 : 175.00 1.50 HWRFw-76 - 3261 : 177.00 1.50 HWRFw-77 - 3262 : 179.00 1.50 HWRFw-78 - 3263 : 180.00 1.50 HWRFw-79 - 3264 : 180.00 3.50 HWRFw-80 - 3265 : 180.00 5.50 HWRFw-81 - 3266 : 180.00 7.50 HWRFw-82 - 3267 : 180.00 9.50 HWRFw-83 - 3268 : 180.00 11.50 HWRFw-84 - 3269 : 180.00 13.50 HWRFw-85 - 3270 : 180.00 15.50 HWRFw-86 - 3271 : 180.00 17.50 HWRFw-87 - 3272 : 180.00 19.50 HWRFw-88 - 3273 : 180.00 21.50 HWRFw-89 - 3274 : 180.00 23.50 HWRFw-90 - 3275 : 180.00 25.50 HWRFw-91 - 3276 : 180.00 27.50 HWRFw-92 - 3277 : 180.00 29.50 HWRFw-93 - 3278 : 180.00 31.50 HWRFw-94 - 3279 : 180.00 33.50 HWRFw-95 - 3280 : 180.00 35.50 HWRFw-96 - 3281 : 180.00 37.50 HWRFw-97 - 3282 : 180.00 39.50 HWRFw-98 - 3283 : 180.00 41.50 HWRFw-99 - 3284 : 179.00 41.50 HWRFw-100 - 3285 : 177.00 41.50 HWRFw-101 - 3286 : 175.00 41.50 HWRFw-102 - 3287 : 173.00 41.50 HWRFw-103 - 3288 : 171.00 41.50 HWRFw-104 - 3289 : 169.00 41.50 HWRFw-105 - 3290 : 167.00 41.50 HWRFw-106 - 3291 : 165.00 41.50 HWRFw-107 - 3292 : 163.00 41.50 HWRFw-108 - 3293 : 161.00 41.50 HWRFw-109 - 3294 : 159.00 41.50 HWRFw-110 - 3295 : 157.00 41.50 HWRFw-111 - 3296 : 155.00 41.50 HWRFw-112 - 3297 : 153.00 41.50 HWRFw-113 - 3298 : 151.00 41.50 HWRFw-114 - 3299 : 149.00 41.50 HWRFw-115 - 3300 : 147.00 41.50 HWRFw-116 - 3301 : 145.00 41.50 HWRFw-117 - 3302 : 143.00 41.50 HWRFw-118 - 3303 : 139.00 41.50 HWRFw-119 - 3304 : 137.00 41.50 HWRFw-120 - 3305 : 135.00 41.50 HWRFw-121 - 3306 : 133.00 41.50 HWRFw-122 - 3307 : 131.00 41.50 HWRFw-123 - 3308 : -98.00 3.00 RW-NH2-51 - 3309 : -98.00 3.50 RW-NH2-52 - 3310 : -98.00 4.00 RW-NH2-53 - 3311 : -98.00 4.50 RW-NH2-54 - 3312 : -98.00 5.00 RW-NH2-55 - 3313 : -98.00 5.50 RW-NH2-56 - 3314 : -98.00 6.00 RW-NH2-57 - 3315 : -98.00 6.50 RW-NH2-58 - 3316 : -98.00 7.00 RW-NH2-59 - 3317 : -98.00 7.50 RW-NH2-60 - 3318 : -98.00 8.00 RW-NH2-61 - 3319 : -98.00 8.50 RW-NH2-62 - 3320 : -98.00 9.00 RW-NH2-63 - 3321 : -98.00 9.50 RW-NH2-64 - 3322 : -98.00 10.00 RW-NH2-65 - 3323 : -98.00 10.50 RW-NH2-66 - 3324 : -98.00 11.00 RW-NH2-67 - 3325 : -98.00 11.50 RW-NH2-68 - 3326 : -98.00 12.00 RW-NH2-69 - 3327 : -98.00 12.50 RW-NH2-70 - 3328 : -98.00 13.00 RW-NH2-71 - 3329 : -98.00 13.50 RW-NH2-72 - 3330 : -98.00 14.00 RW-NH2-73 - 3331 : -98.00 14.50 RW-NH2-74 - 3332 : -98.00 15.00 RW-NH2-75 - 3333 : -98.00 15.50 RW-NH2-76 - 3334 : -98.00 16.00 RW-NH2-77 - 3335 : -80.75 32.00 RW-NH2-78 - 3336 : -80.50 32.00 RW-NH2-79 - 3337 : -80.00 32.00 RW-NH2-80 - 3338 : -79.50 32.00 RW-NH2-81 - 3339 : -79.00 32.00 RW-NH2-82 - 3340 : -78.50 32.00 RW-NH2-83 - 3341 : -78.00 32.00 RW-NH2-84 - 3342 : -77.50 32.00 RW-NH2-85 - 3343 : -77.00 32.00 RW-NH2-86 - 3344 : -76.50 32.00 RW-NH2-87 - 3345 : -76.00 32.00 RW-NH2-88 - 3346 : -75.50 32.00 RW-NH2-89 - 3347 : -75.00 32.00 RW-NH2-90 - 3348 : -74.50 32.00 RW-NH2-91 - 3349 : -74.00 32.00 RW-NH2-92 - 3350 : -73.50 32.00 RW-NH2-93 - 3351 : -73.00 32.00 RW-NH2-94 - 3352 : -72.50 32.00 RW-NH2-95 - 3353 : -72.00 32.00 RW-NH2-96 - 3354 : -71.50 32.00 RW-NH2-97 - 3355 : -71.00 32.00 RW-NH2-98 - 3356 : -70.50 32.00 RW-NH2-99 - 3357 : -70.00 32.00 RW-NH2-100 - 3358 : -69.50 32.00 RW-NH2-101 - 3359 : -69.00 32.00 RW-NH2-102 - 3360 : -68.50 32.00 RW-NH2-103 - 3361 : -68.00 32.00 RW-NH2-104 - 3362 : -67.50 32.00 RW-NH2-105 - 3363 : -67.00 32.00 RW-NH2-106 - 3364 : -66.50 32.00 RW-NH2-107 - 3365 : -66.00 32.00 RW-NH2-108 - 3366 : -65.50 32.00 RW-NH2-109 - 3367 : -65.00 32.00 RW-NH2-110 - 3368 : -64.50 32.00 RW-NH2-111 - 3369 : -64.00 32.00 RW-NH2-112 - 3370 : -63.50 32.00 RW-NH2-113 - 3371 : -63.00 32.00 RW-NH2-114 - 3372 : -62.50 32.00 RW-NH2-115 - 3373 : -62.00 32.00 RW-NH2-116 - 3374 : -61.50 32.00 RW-NH2-117 - 3375 : -61.00 32.00 RW-NH2-118 - 3376 : -60.50 32.00 RW-NH2-119 - 3377 : -60.00 32.00 RW-NH2-120 - 3378 : -59.50 32.00 RW-NH2-121 - 3379 : -59.00 32.00 RW-NH2-122 - 3380 : -58.50 32.00 RW-NH2-123 - 3381 : -58.00 32.00 RW-NH2-124 - 3382 : -57.50 32.00 RW-NH2-125 - 3383 : -57.00 32.00 RW-NH2-126 - 3384 : -56.50 32.00 RW-NH2-127 - 3385 : -56.00 32.00 RW-NH2-128 - 3386 : -55.50 32.00 RW-NH2-129 - 3387 : -55.00 32.00 RW-NH2-130 - 3388 : -54.50 32.00 RW-NH2-131 - 3389 : -54.00 32.00 RW-NH2-132 - 3390 : -53.50 32.00 RW-NH2-133 - 3391 : -53.00 32.00 RW-NH2-134 - 3392 : -52.50 32.00 RW-NH2-135 - 3393 : -52.00 32.00 RW-NH2-136 - 3394 : -51.50 32.00 RW-NH2-137 - 3395 : -51.00 32.00 RW-NH2-138 - 3396 : -50.50 32.00 RW-NH2-139 - 3397 : -50.00 32.00 RW-NH2-140 - 3398 : -49.50 32.00 RW-NH2-141 - 3399 : -49.00 32.00 RW-NH2-142 - 3400 : -48.50 32.00 RW-NH2-143 - 3401 : -48.00 32.00 RW-NH2-144 - 3402 : -47.50 32.00 RW-NH2-145 - 3403 : -47.00 32.00 RW-NH2-146 - 3404 : -46.50 32.00 RW-NH2-147 - 3405 : -46.00 32.00 RW-NH2-148 - 3406 : -45.50 32.00 RW-NH2-149 - 3407 : -45.00 32.00 RW-NH2-150 - 3408 : -44.50 32.00 RW-NH2-151 - 3409 : -44.00 32.00 RW-NH2-152 - 3410 : -43.50 32.00 RW-NH2-153 - 3411 : -43.00 32.00 RW-NH2-154 - 3412 : -42.50 32.00 RW-NH2-155 - 3413 : -42.00 32.00 RW-NH2-156 - 3414 : -41.50 32.00 RW-NH2-157 - 3415 : -41.00 32.00 RW-NH2-158 - 3416 : -40.50 32.00 RW-NH2-159 - 3417 : -40.00 32.00 RW-NH2-160 - 3418 : -39.50 32.00 RW-NH2-161 - 3419 : -39.00 32.00 RW-NH2-162 - 3420 : -38.50 32.00 RW-NH2-163 - 3421 : -38.00 32.00 RW-NH2-164 - 3422 : -37.50 32.00 RW-NH2-165 - 3423 : -37.00 32.00 RW-NH2-166 - 3424 : -36.50 32.00 RW-NH2-167 - 3425 : -36.00 32.00 RW-NH2-168 - 3426 : -35.50 32.00 RW-NH2-169 - 3427 : -35.00 32.00 RW-NH2-170 - 3428 : -34.50 32.00 RW-NH2-171 - 3429 : -34.00 32.00 RW-NH2-172 - 3430 : -33.50 32.00 RW-NH2-173 - 3431 : -33.00 32.00 RW-NH2-174 - 3432 : -32.50 32.00 RW-NH2-175 - 3433 : -32.00 32.00 RW-NH2-176 - 3434 : -31.50 32.00 RW-NH2-177 - 3435 : -31.00 32.00 RW-NH2-178 - 3436 : -30.50 32.00 RW-NH2-179 - 3437 : -30.00 32.00 RW-NH2-180 - 3438 : -29.50 32.00 RW-NH2-181 - 3439 : -29.00 32.00 RW-NH2-182 - 3440 : -28.50 32.00 RW-NH2-183 - 3441 : -28.00 32.00 RW-NH2-184 - 3442 : -27.50 32.00 RW-NH2-185 - 3443 : -27.00 32.00 RW-NH2-186 - 3444 : -26.50 32.00 RW-NH2-187 - 3445 : -26.00 32.00 RW-NH2-188 - 3446 : -25.50 32.00 RW-NH2-189 - 3447 : -25.00 32.00 RW-NH2-190 - 3448 : -24.50 32.00 RW-NH2-191 - 3449 : -24.00 32.00 RW-NH2-192 - 3450 : -23.50 32.00 RW-NH2-193 - 3451 : -23.00 32.00 RW-NH2-194 - 3452 : -22.50 32.00 RW-NH2-195 - 3453 : -22.00 32.00 RW-NH2-196 - 3454 : -21.50 32.00 RW-NH2-197 - 3455 : -21.00 32.00 RW-NH2-198 - 3456 : -20.50 32.00 RW-NH2-199 - 3457 : -20.00 32.00 RW-NH2-200 - 3458 : -19.50 32.00 RW-NH2-201 - 3459 : -19.00 32.00 RW-NH2-202 - 3460 : -18.50 32.00 RW-NH2-203 - 3461 : -18.00 32.00 RW-NH2-204 - 3462 : -17.50 32.00 RW-NH2-205 - 3463 : -17.00 32.00 RW-NH2-206 - 3464 : -16.50 32.00 RW-NH2-207 - 3465 : -16.00 32.00 RW-NH2-208 - 3466 : -15.50 32.00 RW-NH2-209 - 3467 : -15.00 32.00 RW-NH2-210 - 3468 : -14.50 32.00 RW-NH2-211 - 3469 : -14.00 32.00 RW-NH2-212 - 3470 : -13.50 32.00 RW-NH2-213 - 3471 : -13.00 32.00 RW-NH2-214 - 3472 : -12.50 32.00 RW-NH2-215 - 3473 : -12.00 32.00 RW-NH2-216 - 3474 : -11.50 32.00 RW-NH2-217 - 3475 : -11.00 32.00 RW-NH2-218 - 3476 : -10.50 32.00 RW-NH2-219 - 3477 : -10.00 32.00 RW-NH2-220 - 3478 : -9.50 32.00 RW-NH2-221 - 3479 : -10.00 5.50 RW-NH2-222 - 3480 : -10.00 5.00 RW-NH2-223 - 3481 : -10.00 4.50 RW-NH2-224 - 3482 : -10.00 4.00 RW-NH2-225 - 3483 : -10.00 3.50 RW-NH2-226 - 3484 : -10.00 3.00 RW-NH2-227 - 3485 : -10.00 3.00 RW-NH2-228 - 3486 : -10.50 3.00 RW-NH2-229 - 3487 : -11.00 3.00 RW-NH2-230 - 3488 : -11.50 3.00 RW-NH2-231 - 3489 : -12.00 3.00 RW-NH2-232 - 3490 : -12.50 3.00 RW-NH2-233 - 3491 : -13.00 3.00 RW-NH2-234 - 3492 : -13.50 3.00 RW-NH2-235 - 3493 : -14.00 3.00 RW-NH2-236 - 3494 : -14.50 3.00 RW-NH2-237 - 3495 : -15.00 3.00 RW-NH2-238 - 3496 : -15.50 3.00 RW-NH2-239 - 3497 : -16.00 3.00 RW-NH2-240 - 3498 : -16.50 3.00 RW-NH2-241 - 3499 : -17.00 3.00 RW-NH2-242 - 3500 : -17.50 3.00 RW-NH2-243 - 3501 : -18.00 3.00 RW-NH2-244 - 3502 : -18.50 3.00 RW-NH2-245 - 3503 : -19.00 3.00 RW-NH2-246 - 3504 : -19.50 3.00 RW-NH2-247 - 3505 : -20.00 3.00 RW-NH2-248 - 3506 : -20.50 3.00 RW-NH2-249 - 3507 : -21.00 3.00 RW-NH2-250 - 3508 : -21.50 3.00 RW-NH2-251 - 3509 : -22.00 3.00 RW-NH2-252 - 3510 : -22.50 3.00 RW-NH2-253 - 3511 : -23.00 3.00 RW-NH2-254 - 3512 : -23.50 3.00 RW-NH2-255 - 3513 : -24.00 3.00 RW-NH2-256 - 3514 : -24.50 3.00 RW-NH2-257 - 3515 : -25.00 3.00 RW-NH2-258 - 3516 : -25.50 3.00 RW-NH2-259 - 3517 : -26.00 3.00 RW-NH2-260 - 3518 : -26.50 3.00 RW-NH2-261 - 3519 : -27.00 3.00 RW-NH2-262 - 3520 : -27.50 3.00 RW-NH2-263 - 3521 : -28.00 3.00 RW-NH2-264 - 3522 : -28.50 3.00 RW-NH2-265 - 3523 : -29.00 3.00 RW-NH2-266 - 3524 : -29.50 3.00 RW-NH2-267 - 3525 : -30.00 3.00 RW-NH2-268 - 3526 : -30.50 3.00 RW-NH2-269 - 3527 : -31.00 3.00 RW-NH2-270 - 3528 : -31.50 3.00 RW-NH2-271 - 3529 : -32.00 3.00 RW-NH2-272 - 3530 : -32.50 3.00 RW-NH2-273 - 3531 : -33.00 3.00 RW-NH2-274 - 3532 : -33.50 3.00 RW-NH2-275 - 3533 : -34.00 3.00 RW-NH2-276 - 3534 : -34.50 3.00 RW-NH2-277 - 3535 : -35.00 3.00 RW-NH2-278 - 3536 : -35.50 3.00 RW-NH2-279 - 3537 : -36.00 3.00 RW-NH2-280 - 3538 : -36.50 3.00 RW-NH2-281 - 3539 : -37.00 3.00 RW-NH2-282 - 3540 : -37.50 3.00 RW-NH2-283 - 3541 : -38.00 3.00 RW-NH2-284 - 3542 : -38.50 3.00 RW-NH2-285 - 3543 : -39.00 3.00 RW-NH2-286 - 3544 : -39.50 3.00 RW-NH2-287 - 3545 : -40.00 3.00 RW-NH2-288 - 3546 : -40.50 3.00 RW-NH2-289 - 3547 : -41.00 3.00 RW-NH2-290 - 3548 : -41.50 3.00 RW-NH2-291 - 3549 : -42.00 3.00 RW-NH2-292 - 3550 : -42.50 3.00 RW-NH2-293 - 3551 : -43.00 3.00 RW-NH2-294 - 3552 : -43.50 3.00 RW-NH2-295 - 3553 : -44.00 3.00 RW-NH2-296 - 3554 : -44.50 3.00 RW-NH2-297 - 3555 : -42.00 3.00 RW-NH2-298 - 3556 : -42.50 3.00 RW-NH2-299 - 3557 : -43.00 3.00 RW-NH2-300 - 3558 : -43.50 3.00 RW-NH2-301 - 3559 : -44.00 3.00 RW-NH2-302 - 3560 : -44.50 3.00 RW-NH2-303 - 3561 : -45.00 3.00 RW-NH2-304 - 3562 : -42.50 3.00 RW-NH2-305 - 3563 : -43.00 3.00 RW-NH2-306 - 3564 : -43.50 3.00 RW-NH2-307 - 3565 : -44.00 3.00 RW-NH2-308 - 3566 : -44.50 3.00 RW-NH2-309 - 3567 : -45.00 3.00 RW-NH2-310 - 3568 : -45.50 3.00 RW-NH2-311 - 3569 : -46.00 3.00 RW-NH2-312 - 3570 : -46.50 3.00 RW-NH2-313 - 3571 : -47.00 3.00 RW-NH2-314 - 3572 : -47.50 3.00 RW-NH2-315 - 3573 : -48.00 3.00 RW-NH2-316 - 3574 : -45.50 3.00 RW-NH2-317 - 3575 : -46.00 3.00 RW-NH2-318 - 3576 : -46.50 3.00 RW-NH2-319 - 3577 : -47.00 3.00 RW-NH2-320 - 3578 : -47.50 3.00 RW-NH2-321 - 3579 : -48.00 3.00 RW-NH2-322 - 3580 : -48.50 3.00 RW-NH2-323 - 3581 : -49.00 3.00 RW-NH2-324 - 3582 : -49.50 3.00 RW-NH2-325 - 3583 : -50.00 3.00 RW-NH2-326 - 3584 : -50.50 3.00 RW-NH2-327 - 3585 : -50.75 3.00 RW-NH2-328 - 3586 : -77.75 3.00 RW-NH2-329 - 3587 : -78.00 3.00 RW-NH2-330 - 3588 : -78.50 3.00 RW-NH2-331 - 3589 : -79.00 3.00 RW-NH2-332 - 3590 : -79.50 3.00 RW-NH2-333 - 3591 : -80.00 3.00 RW-NH2-334 - 3592 : -80.50 3.00 RW-NH2-335 - 3593 : -81.00 3.00 RW-NH2-336 - 3594 : -81.50 3.00 RW-NH2-337 - 3595 : -82.00 3.00 RW-NH2-338 - 3596 : -82.50 3.00 RW-NH2-339 - 3597 : -83.00 3.00 RW-NH2-340 - 3598 : -83.50 3.00 RW-NH2-341 - 3599 : -84.00 3.00 RW-NH2-342 - 3600 : -84.50 3.00 RW-NH2-343 - 3601 : -85.00 3.00 RW-NH2-344 - 3602 : -85.50 3.00 RW-NH2-345 - 3603 : -86.00 3.00 RW-NH2-346 - 3604 : -86.50 3.00 RW-NH2-347 - 3605 : -87.00 3.00 RW-NH2-348 - 3606 : -87.50 3.00 RW-NH2-349 - 3607 : -88.00 3.00 RW-NH2-350 - 3608 : -88.50 3.00 RW-NH2-351 - 3609 : -89.00 3.00 RW-NH2-352 - 3610 : -89.50 3.00 RW-NH2-353 - 3611 : -90.00 3.00 RW-NH2-354 - 3612 : -90.50 3.00 RW-NH2-355 - 3613 : -91.00 3.00 RW-NH2-356 - 3614 : -91.50 3.00 RW-NH2-357 - 3615 : -92.00 3.00 RW-NH2-358 - 3616 : -92.50 3.00 RW-NH2-359 - 3617 : -93.00 3.00 RW-NH2-360 - 3618 : -93.50 3.00 RW-NH2-361 - 3619 : -94.00 3.00 RW-NH2-362 - 3620 : -94.50 3.00 RW-NH2-363 - 3621 : -95.00 3.00 RW-NH2-364 - 3622 : -95.50 3.00 RW-NH2-365 - 3623 : -96.00 3.00 RW-NH2-366 - 3624 : -96.50 3.00 RW-NH2-367 - 3625 : -97.00 3.00 RW-NH2-368 - 3626 : -97.50 3.00 RW-NH2-369 - 3627 : -98.00 3.00 RW-NH2-370 - 3628 : -30.00 3.00 RW-NH2-371 - 3629 : -30.00 3.50 RW-NH2-372 - 3630 : -30.00 4.00 RW-NH2-373 - 3631 : -30.00 4.50 RW-NH2-374 - 3632 : -30.00 5.00 RW-NH2-375 - 3633 : -30.00 5.50 RW-NH2-376 - 3634 : -30.00 6.00 RW-NH2-377 - 3635 : -30.00 6.50 RW-NH2-378 - 3636 : -30.00 7.00 RW-NH2-379 - 3637 : -30.00 7.50 RW-NH2-380 - 3638 : -30.00 8.00 RW-NH2-381 - 3639 : -30.00 8.50 RW-NH2-382 - 3640 : -30.00 9.00 RW-NH2-383 - 3641 : -30.00 9.50 RW-NH2-384 - 3642 : -30.00 10.00 RW-NH2-385 - 3643 : -30.00 10.50 RW-NH2-386 - 3644 : -30.00 11.00 RW-NH2-387 - 3645 : -30.00 11.50 RW-NH2-388 - 3646 : -30.00 12.00 RW-NH2-389 - 3647 : -30.00 12.50 RW-NH2-390 - 3648 : -30.00 13.00 RW-NH2-391 - 3649 : -30.00 13.50 RW-NH2-392 - 3650 : -30.00 14.00 RW-NH2-393 - 3651 : -30.00 14.50 RW-NH2-394 - 3652 : -30.00 15.00 RW-NH2-395 - 3653 : -30.00 15.50 RW-NH2-396 - 3654 : -30.00 16.00 RW-NH2-397 - 3655 : -30.00 16.50 RW-NH2-398 - 3656 : -30.00 17.00 RW-NH2-399 - 3657 : -30.00 17.50 RW-NH2-400 - 3658 : -30.00 18.00 RW-NH2-401 - 3659 : -30.00 18.50 RW-NH2-402 - 3660 : -30.00 19.00 RW-NH2-403 - 3661 : -30.00 19.50 RW-NH2-404 - 3662 : -30.00 20.00 RW-NH2-405 - 3663 : -30.00 20.50 RW-NH2-406 - 3664 : -30.00 21.00 RW-NH2-407 - 3665 : -30.00 21.50 RW-NH2-408 - 3666 : -30.00 22.00 RW-NH2-409 - 3667 : -30.00 22.50 RW-NH2-410 - 3668 : -30.00 23.00 RW-NH2-411 - 3669 : -30.00 23.50 RW-NH2-412 - 3670 : -30.00 24.00 RW-NH2-413 - 3671 : -30.00 24.50 RW-NH2-414 - 3672 : -30.00 25.00 RW-NH2-415 - 3673 : -30.00 25.60 RW-NH2-416 - 3674 : -30.00 26.00 RW-NH2-417 - 3675 : -30.00 26.50 RW-NH2-418 - 3676 : -30.00 27.00 RW-NH2-419 - 3677 : -30.00 27.50 RW-NH2-420 - 3678 : -30.00 28.00 RW-NH2-421 - 3679 : -30.00 28.50 RW-NH2-422 - 3680 : -30.00 29.00 RW-NH2-423 - 3681 : -30.00 29.50 RW-NH2-424 - 3682 : -30.00 30.00 RW-NH2-425 - 3683 : -30.00 30.50 RW-NH2-426 - 3684 : -30.00 31.00 RW-NH2-427 - 3685 : -30.00 31.50 RW-NH2-428 - 3686 : -30.00 32.00 RW-NH2-429 - 3687 : -141.00 -19.50 RW-NH1-51 - 3688 : -141.00 -19.00 RW-NH1-52 - 3689 : -141.00 -18.50 RW-NH1-53 - 3690 : -141.00 -18.00 RW-NH1-54 - 3691 : -141.00 -17.50 RW-NH1-55 - 3692 : -141.00 -17.00 RW-NH1-56 - 3693 : -141.00 -16.50 RW-NH1-57 - 3694 : -141.00 -16.00 RW-NH1-58 - 3695 : -141.00 -15.50 RW-NH1-59 - 3696 : -141.00 -15.00 RW-NH1-60 - 3697 : -141.00 -14.50 RW-NH1-61 - 3698 : -141.00 -14.00 RW-NH1-62 - 3699 : -141.00 -13.50 RW-NH1-63 - 3700 : -141.00 -13.00 RW-NH1-64 - 3701 : -141.00 -12.50 RW-NH1-65 - 3702 : -141.00 -12.00 RW-NH1-66 - 3703 : -141.00 -11.50 RW-NH1-67 - 3704 : -141.00 -11.00 RW-NH1-68 - 3705 : -141.00 -10.50 RW-NH1-69 - 3706 : -141.00 -10.00 RW-NH1-70 - 3707 : -141.00 -9.50 RW-NH1-71 - 3708 : -141.00 -9.00 RW-NH1-72 - 3709 : -141.00 -8.50 RW-NH1-73 - 3710 : -141.00 -8.00 RW-NH1-74 - 3711 : -141.00 -7.50 RW-NH1-75 - 3712 : -141.00 -7.00 RW-NH1-76 - 3713 : -141.00 -6.50 RW-NH1-77 - 3714 : -141.00 -6.00 RW-NH1-78 - 3715 : -141.00 -5.50 RW-NH1-79 - 3716 : -141.00 -5.00 RW-NH1-80 - 3717 : -141.00 -4.50 RW-NH1-81 - 3718 : -141.00 -4.00 RW-NH1-82 - 3719 : -141.00 -3.50 RW-NH1-83 - 3720 : -141.00 -3.00 RW-NH1-84 - 3721 : -141.00 -2.50 RW-NH1-85 - 3722 : -141.00 -2.00 RW-NH1-86 - 3723 : -141.00 -1.50 RW-NH1-87 - 3724 : -141.00 -1.00 RW-NH1-88 - 3725 : -141.00 -0.50 RW-NH1-89 - 3726 : -141.00 0.00 RW-NH1-90 - 3727 : -141.00 0.50 RW-NH1-91 - 3728 : -141.00 1.00 RW-NH1-92 - 3729 : -141.00 1.50 RW-NH1-93 - 3730 : -141.00 2.00 RW-NH1-94 - 3731 : -141.00 2.50 RW-NH1-95 - 3732 : -141.00 3.00 RW-NH1-96 - 3733 : -141.00 3.50 RW-NH1-97 - 3734 : -141.00 4.00 RW-NH1-98 - 3735 : -141.00 4.50 RW-NH1-99 - 3736 : -141.00 5.00 RW-NH1-100 - 3737 : -141.00 5.50 RW-NH1-101 - 3738 : -141.00 6.00 RW-NH1-102 - 3739 : -141.00 6.50 RW-NH1-103 - 3740 : -141.00 7.00 RW-NH1-104 - 3741 : -141.00 7.50 RW-NH1-105 - 3742 : -141.00 8.00 RW-NH1-106 - 3743 : -141.00 8.50 RW-NH1-107 - 3744 : -141.00 9.00 RW-NH1-108 - 3745 : -141.00 9.50 RW-NH1-109 - 3746 : -141.00 10.00 RW-NH1-110 - 3747 : -141.00 10.50 RW-NH1-111 - 3748 : -141.00 11.00 RW-NH1-112 - 3749 : -141.00 11.50 RW-NH1-113 - 3750 : -141.00 12.00 RW-NH1-114 - 3751 : -141.00 12.50 RW-NH1-115 - 3752 : -141.00 13.00 RW-NH1-116 - 3753 : -141.00 13.50 RW-NH1-117 - 3754 : -141.00 14.00 RW-NH1-118 - 3755 : -141.00 14.50 RW-NH1-119 - 3756 : -141.00 15.00 RW-NH1-120 - 3757 : -141.00 15.50 RW-NH1-121 - 3758 : -141.00 16.00 RW-NH1-122 - 3759 : -141.00 16.50 RW-NH1-123 - 3760 : -141.00 17.00 RW-NH1-124 - 3761 : -141.00 17.50 RW-NH1-125 - 3762 : -141.00 18.00 RW-NH1-126 - 3763 : -141.00 18.50 RW-NH1-127 - 3764 : -141.00 19.00 RW-NH1-128 - 3765 : -141.00 19.50 RW-NH1-129 - 3766 : -141.00 20.00 RW-NH1-130 - 3767 : -141.00 20.50 RW-NH1-131 - 3768 : -141.00 21.00 RW-NH1-132 - 3769 : -141.00 21.50 RW-NH1-133 - 3770 : -141.00 22.00 RW-NH1-134 - 3771 : -141.00 22.50 RW-NH1-135 - 3772 : -141.00 23.00 RW-NH1-136 - 3773 : -141.00 23.50 RW-NH1-137 - 3774 : -141.00 24.00 RW-NH1-138 - 3775 : -141.00 24.50 RW-NH1-139 - 3776 : -141.00 25.00 RW-NH1-140 - 3777 : -141.00 25.50 RW-NH1-141 - 3778 : -141.00 26.00 RW-NH1-142 - 3779 : -141.00 26.50 RW-NH1-143 - 3780 : -141.00 27.00 RW-NH1-144 - 3781 : -141.00 27.50 RW-NH1-145 - 3782 : -141.00 28.00 RW-NH1-146 - 3783 : -141.00 28.50 RW-NH1-147 - 3784 : -141.00 29.00 RW-NH1-148 - 3785 : -141.00 29.50 RW-NH1-149 - 3786 : -141.00 30.00 RW-NH1-150 - 3787 : -141.00 30.50 RW-NH1-151 - 3788 : -141.00 31.00 RW-NH1-152 - 3789 : -141.00 31.50 RW-NH1-153 - 3790 : -141.00 32.00 RW-NH1-154 - 3791 : -141.00 32.50 RW-NH1-155 - 3792 : -141.00 32.80 RW-NH1-156 - 3793 : -141.00 32.80 RW-NH1-157 - 3794 : -140.50 32.80 RW-NH1-158 - 3795 : -140.00 32.80 RW-NH1-159 - 3796 : -139.50 32.80 RW-NH1-160 - 3797 : -139.00 32.80 RW-NH1-161 - 3798 : -138.50 32.80 RW-NH1-162 - 3799 : -138.00 32.80 RW-NH1-163 - 3800 : -137.50 32.80 RW-NH1-164 - 3801 : -137.00 32.80 RW-NH1-165 - 3802 : -136.50 32.80 RW-NH1-166 - 3803 : -136.00 32.80 RW-NH1-167 - 3804 : -135.50 32.80 RW-NH1-168 - 3805 : -135.00 32.80 RW-NH1-169 - 3806 : -134.50 32.80 RW-NH1-170 - 3807 : -134.00 32.80 RW-NH1-171 - 3808 : -133.50 32.80 RW-NH1-172 - 3809 : -133.00 32.80 RW-NH1-173 - 3810 : -132.50 32.80 RW-NH1-174 - 3811 : -132.00 32.80 RW-NH1-175 - 3812 : -131.50 32.80 RW-NH1-176 - 3813 : -131.00 32.80 RW-NH1-177 - 3814 : -130.50 32.80 RW-NH1-178 - 3815 : -130.00 32.80 RW-NH1-179 - 3816 : -129.50 32.80 RW-NH1-180 - 3817 : -129.00 32.80 RW-NH1-181 - 3818 : -128.50 32.80 RW-NH1-182 - 3819 : -128.00 32.80 RW-NH1-183 - 3820 : -127.50 32.80 RW-NH1-184 - 3821 : -127.00 32.80 RW-NH1-185 - 3822 : -126.50 32.80 RW-NH1-186 - 3823 : -126.00 32.80 RW-NH1-187 - 3824 : -125.50 32.80 RW-NH1-188 - 3825 : -125.00 32.80 RW-NH1-189 - 3826 : -124.50 32.80 RW-NH1-190 - 3827 : -124.00 32.80 RW-NH1-191 - 3828 : -123.50 32.80 RW-NH1-192 - 3829 : -123.00 32.80 RW-NH1-193 - 3830 : -122.50 32.80 RW-NH1-194 - 3831 : -122.00 32.80 RW-NH1-195 - 3832 : -121.50 32.80 RW-NH1-196 - 3833 : -121.00 32.80 RW-NH1-197 - 3834 : -120.50 32.80 RW-NH1-198 - 3835 : -120.00 32.80 RW-NH1-199 - 3836 : -119.50 32.80 RW-NH1-200 - 3837 : -119.00 32.80 RW-NH1-201 - 3838 : -118.50 32.80 RW-NH1-202 - 3839 : -118.00 32.80 RW-NH1-203 - 3840 : -117.50 32.80 RW-NH1-204 - 3841 : -113.50 31.00 RW-NH1-205 - 3842 : -81.25 31.00 RW-NH1-206 - 3843 : -81.00 31.00 RW-NH1-207 - 3844 : -80.50 31.00 RW-NH1-208 - 3845 : -80.00 31.00 RW-NH1-209 - 3846 : -79.50 31.00 RW-NH1-210 - 3847 : -79.00 31.00 RW-NH1-211 - 3848 : -78.50 31.00 RW-NH1-212 - 3849 : -78.00 31.00 RW-NH1-213 - 3850 : -77.50 31.00 RW-NH1-214 - 3851 : -77.00 31.00 RW-NH1-215 - 3852 : -76.50 31.00 RW-NH1-216 - 3853 : -76.00 31.00 RW-NH1-217 - 3854 : -75.50 31.00 RW-NH1-218 - 3855 : -75.00 31.00 RW-NH1-219 - 3856 : -74.50 31.00 RW-NH1-220 - 3857 : -74.00 31.00 RW-NH1-221 - 3858 : -73.50 31.00 RW-NH1-222 - 3859 : -73.00 31.00 RW-NH1-223 - 3860 : -72.50 31.00 RW-NH1-224 - 3861 : -72.00 31.00 RW-NH1-225 - 3862 : -71.50 31.00 RW-NH1-226 - 3863 : -71.00 31.00 RW-NH1-227 - 3864 : -70.50 31.00 RW-NH1-228 - 3865 : -70.00 31.00 RW-NH1-229 - 3866 : -70.00 31.00 RW-NH1-230 - 3867 : -70.00 30.50 RW-NH1-231 - 3868 : -70.00 30.00 RW-NH1-232 - 3869 : -70.00 29.50 RW-NH1-233 - 3870 : -70.00 29.00 RW-NH1-234 - 3871 : -70.00 28.50 RW-NH1-235 - 3872 : -70.00 28.00 RW-NH1-236 - 3873 : -70.00 27.50 RW-NH1-237 - 3874 : -70.00 27.00 RW-NH1-238 - 3875 : -70.00 26.50 RW-NH1-239 - 3876 : -70.00 26.00 RW-NH1-240 - 3877 : -70.00 25.50 RW-NH1-241 - 3878 : -70.00 25.00 RW-NH1-242 - 3879 : -70.00 24.50 RW-NH1-243 - 3880 : -70.00 24.00 RW-NH1-244 - 3881 : -70.00 23.50 RW-NH1-245 - 3882 : -70.00 23.00 RW-NH1-246 - 3883 : -70.00 22.50 RW-NH1-247 - 3884 : -70.00 22.00 RW-NH1-248 - 3885 : -70.00 21.50 RW-NH1-249 - 3886 : -70.00 21.00 RW-NH1-250 - 3887 : -70.00 20.50 RW-NH1-251 - 3888 : -70.00 20.00 RW-NH1-252 - 3889 : -70.00 18.00 RW-NH1-253 - 3890 : -70.00 17.50 RW-NH1-254 - 3891 : -70.00 17.00 RW-NH1-255 - 3892 : -70.00 16.50 RW-NH1-256 - 3893 : -70.00 16.00 RW-NH1-257 - 3894 : -70.00 15.50 RW-NH1-258 - 3895 : -70.00 15.00 RW-NH1-259 - 3896 : -70.00 14.50 RW-NH1-260 - 3897 : -70.00 14.00 RW-NH1-261 - 3898 : -70.00 13.50 RW-NH1-262 - 3899 : -70.00 13.00 RW-NH1-263 - 3900 : -70.00 12.50 RW-NH1-264 - 3901 : -70.50 -19.50 RW-NH1-265 - 3902 : -71.00 -19.50 RW-NH1-266 - 3903 : -71.50 -19.50 RW-NH1-267 - 3904 : -72.00 -19.50 RW-NH1-268 - 3905 : -72.50 -19.50 RW-NH1-269 - 3906 : -73.00 -19.50 RW-NH1-270 - 3907 : -73.50 -19.50 RW-NH1-271 - 3908 : -74.00 -19.50 RW-NH1-272 - 3909 : -74.50 -19.50 RW-NH1-273 - 3910 : -75.00 -19.50 RW-NH1-274 - 3911 : -75.50 -19.50 RW-NH1-275 - 3912 : -76.00 -19.50 RW-NH1-276 - 3913 : -76.50 -19.50 RW-NH1-277 - 3914 : -77.00 -19.50 RW-NH1-278 - 3915 : -77.50 -19.50 RW-NH1-279 - 3916 : -78.00 -19.50 RW-NH1-280 - 3917 : -78.50 -19.50 RW-NH1-281 - 3918 : -79.00 -19.50 RW-NH1-282 - 3919 : -79.50 -19.50 RW-NH1-283 - 3920 : -80.00 -19.50 RW-NH1-284 - 3921 : -80.50 -19.50 RW-NH1-285 - 3922 : -81.00 -19.50 RW-NH1-286 - 3923 : -81.50 -19.50 RW-NH1-287 - 3924 : -82.00 -19.50 RW-NH1-288 - 3925 : -82.50 -19.50 RW-NH1-289 - 3926 : -83.00 -19.50 RW-NH1-290 - 3927 : -83.50 -19.50 RW-NH1-291 - 3928 : -84.00 -19.50 RW-NH1-292 - 3929 : -84.50 -19.50 RW-NH1-293 - 3930 : -85.00 -19.50 RW-NH1-294 - 3931 : -85.50 -19.50 RW-NH1-295 - 3932 : -86.00 -19.50 RW-NH1-296 - 3933 : -86.50 -19.50 RW-NH1-297 - 3934 : -87.00 -19.50 RW-NH1-298 - 3935 : -87.50 -19.50 RW-NH1-299 - 3936 : -88.00 -19.50 RW-NH1-300 - 3937 : -88.50 -19.50 RW-NH1-301 - 3938 : -89.00 -19.50 RW-NH1-302 - 3939 : -89.50 -19.50 RW-NH1-303 - 3940 : -90.00 -19.50 RW-NH1-304 - 3941 : -90.50 -19.50 RW-NH1-305 - 3942 : -91.00 -19.50 RW-NH1-306 - 3943 : -91.50 -19.50 RW-NH1-307 - 3944 : -92.00 -19.50 RW-NH1-308 - 3945 : -92.50 -19.50 RW-NH1-309 - 3946 : -93.00 -19.50 RW-NH1-310 - 3947 : -93.50 -19.50 RW-NH1-311 - 3948 : -94.00 -19.50 RW-NH1-312 - 3949 : -94.50 -19.50 RW-NH1-313 - 3950 : -95.00 -19.50 RW-NH1-314 - 3951 : -95.50 -19.50 RW-NH1-315 - 3952 : -96.00 -19.50 RW-NH1-316 - 3953 : -96.50 -19.50 RW-NH1-317 - 3954 : -97.00 -19.50 RW-NH1-318 - 3955 : -97.50 -19.50 RW-NH1-319 - 3956 : -98.00 -19.50 RW-NH1-320 - 3957 : -98.50 -19.50 RW-NH1-321 - 3958 : -99.00 -19.50 RW-NH1-322 - 3959 : -99.50 -19.50 RW-NH1-323 - 3960 : -100.00 -19.50 RW-NH1-324 - 3961 : -100.50 -19.50 RW-NH1-325 - 3962 : -101.00 -19.50 RW-NH1-326 - 3963 : -101.50 -19.50 RW-NH1-327 - 3964 : -102.00 -19.50 RW-NH1-328 - 3965 : -102.50 -19.50 RW-NH1-329 - 3966 : -103.00 -19.50 RW-NH1-330 - 3967 : -103.50 -19.50 RW-NH1-331 - 3968 : -104.00 -19.50 RW-NH1-332 - 3969 : -104.50 -19.50 RW-NH1-333 - 3970 : -102.00 -19.50 RW-NH1-334 - 3971 : -102.50 -19.50 RW-NH1-335 - 3972 : -103.00 -19.50 RW-NH1-336 - 3973 : -103.50 -19.50 RW-NH1-337 - 3974 : -104.00 -19.50 RW-NH1-338 - 3975 : -104.50 -19.50 RW-NH1-339 - 3976 : -105.00 -19.50 RW-NH1-340 - 3977 : -102.50 -19.50 RW-NH1-341 - 3978 : -103.00 -19.50 RW-NH1-342 - 3979 : -103.50 -19.50 RW-NH1-343 - 3980 : -104.00 -19.50 RW-NH1-344 - 3981 : -104.50 -19.50 RW-NH1-345 - 3982 : -105.00 -19.50 RW-NH1-346 - 3983 : -105.50 -19.50 RW-NH1-347 - 3984 : -106.00 -19.50 RW-NH1-348 - 3985 : -106.50 -19.50 RW-NH1-349 - 3986 : -107.00 -19.50 RW-NH1-350 - 3987 : -107.50 -19.50 RW-NH1-351 - 3988 : -108.00 -19.50 RW-NH1-352 - 3989 : -105.50 -19.50 RW-NH1-353 - 3990 : -106.00 -19.50 RW-NH1-354 - 3991 : -106.50 -19.50 RW-NH1-355 - 3992 : -107.00 -19.50 RW-NH1-356 - 3993 : -107.50 -19.50 RW-NH1-357 - 3994 : -108.00 -19.50 RW-NH1-358 - 3995 : -108.50 -19.50 RW-NH1-359 - 3996 : -109.00 -19.50 RW-NH1-360 - 3997 : -109.50 -19.50 RW-NH1-361 - 3998 : -110.00 -19.50 RW-NH1-362 - 3999 : -110.50 -19.50 RW-NH1-363 - 4000 : -111.00 -19.50 RW-NH1-364 - 4001 : -111.50 -19.50 RW-NH1-365 - 4002 : -112.00 -19.50 RW-NH1-366 - 4003 : -112.50 -19.50 RW-NH1-367 - 4004 : -113.00 -19.50 RW-NH1-368 - 4005 : -113.50 -19.50 RW-NH1-369 - 4006 : -114.00 -19.50 RW-NH1-370 - 4007 : -114.50 -19.50 RW-NH1-371 - 4008 : -115.00 -19.50 RW-NH1-372 - 4009 : -115.50 -19.50 RW-NH1-373 - 4010 : -116.00 -19.50 RW-NH1-374 - 4011 : -116.50 -19.50 RW-NH1-375 - 4012 : -117.00 -19.50 RW-NH1-376 - 4013 : -117.50 -19.50 RW-NH1-377 - 4014 : -118.00 -19.50 RW-NH1-378 - 4015 : -118.50 -19.50 RW-NH1-379 - 4016 : -119.00 -19.50 RW-NH1-380 - 4017 : -119.50 -19.50 RW-NH1-381 - 4018 : -120.00 -19.50 RW-NH1-382 - 4019 : -120.50 -19.50 RW-NH1-383 - 4020 : -121.00 -19.50 RW-NH1-384 - 4021 : -121.50 -19.50 RW-NH1-385 - 4022 : -122.00 -19.50 RW-NH1-386 - 4023 : -122.50 -19.50 RW-NH1-387 - 4024 : -123.00 -19.50 RW-NH1-388 - 4025 : -123.50 -19.50 RW-NH1-389 - 4026 : -124.00 -19.50 RW-NH1-390 - 4027 : -124.50 -19.50 RW-NH1-391 - 4028 : -125.00 -19.50 RW-NH1-392 - 4029 : -125.50 -19.50 RW-NH1-393 - 4030 : -126.00 -19.50 RW-NH1-394 - 4031 : -126.50 -19.50 RW-NH1-395 - 4032 : -127.00 -19.50 RW-NH1-396 - 4033 : -127.50 -19.50 RW-NH1-397 - 4034 : -128.00 -19.50 RW-NH1-398 - 4035 : -128.50 -19.50 RW-NH1-399 - 4036 : -129.00 -19.50 RW-NH1-400 - 4037 : -129.50 -19.50 RW-NH1-401 - 4038 : -130.00 -19.50 RW-NH1-402 - 4039 : -130.50 -19.50 RW-NH1-403 - 4040 : -131.00 -19.50 RW-NH1-404 - 4041 : -131.50 -19.50 RW-NH1-405 - 4042 : -132.00 -19.50 RW-NH1-406 - 4043 : -132.50 -19.50 RW-NH1-407 - 4044 : -133.00 -19.50 RW-NH1-408 - 4045 : -133.50 -19.50 RW-NH1-409 - 4046 : -134.00 -19.50 RW-NH1-410 - 4047 : -134.50 -19.50 RW-NH1-411 - 4048 : -135.00 -19.50 RW-NH1-412 - 4049 : -135.50 -19.50 RW-NH1-413 - 4050 : -136.00 -19.50 RW-NH1-414 - 4051 : -136.50 -19.50 RW-NH1-415 - 4052 : -137.00 -19.50 RW-NH1-416 - 4053 : -137.50 -19.50 RW-NH1-417 - 4054 : -138.00 -19.50 RW-NH1-418 - 4055 : -138.50 -19.50 RW-NH1-419 - 4056 : -139.00 -19.50 RW-NH1-420 - 4057 : -139.50 -19.50 RW-NH1-421 - 4058 : -140.00 -19.50 RW-NH1-422 - 4059 : -140.50 -19.50 RW-NH1-423 - 4060 : -141.00 -19.50 RW-NH1-424 - 4061 : 0.00 0.00 STOPSTRING - Fields : Ice concentration - - Time interval : - -------------------------------------------------- - Starting time : 2021/03/23 12:00:00 UTC - Ending time : 2021/03/28 12:00:00 UTC - - - Output requests : - -------------------------------------------------- - No dedicated output process, any file system. - - Type 1 : Fields of mean wave parameters - ----------------------------------------- - From : 2021/03/23 12:00:00 UTC - To : 2021/03/28 12:00:00 UTC - Interval : 01:00:00 - - - Type 2 : Point output - ----------------------------------------- - From : 2021/03/23 12:00:00 UTC - To : 2021/03/28 12:00:00 UTC - Interval : 01:00:00 - - output dates out of run dates : Restart files second request deactivated - Wave model ... -(wav_comp_nuopc:wavinit_ufs) call w3init - - WAVEWATCH III log file version 7.14 - ===================================================================== - date : 2025/07/28 - time : 18:56:25 - - Model definition file read. - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -146.830 60.220 46061 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -152.090 59.760 46106 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -151.829 59.597 46108 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -146.805 60.584 46060 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -154.175 57.910 46077 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -138.000 58.250 46083 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -131.100 53.620 46183 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -129.810 52.420 46185 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -129.795 52.437 46138 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -120.870 34.880 46011 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -123.320 38.230 46013 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -123.970 39.220 46014 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -120.970 34.710 46023 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.380 41.850 46027 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -119.080 33.750 46025 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -119.850 34.240 46053 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -121.010 35.100 46062 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -118.000 32.500 46086 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -119.800 34.330 46216 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -119.430 34.170 46217 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -120.780 34.450 46218 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -118.641 33.860 46221 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -118.320 33.620 46222 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.770 33.460 46223 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.470 33.180 46224 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.390 32.930 46225 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.440 32.630 46227 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.550 43.770 46229 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.370 32.750 46231 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.425 32.517 46232 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.750 32.640 SGX01 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -120.860 35.200 46215 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -121.950 36.760 46236 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -122.634 37.787 46237 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -121.910 36.620 46240 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.130 46.220 46243 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.360 40.890 46244 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.644 46.133 46248 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -121.899 36.835 46091 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -122.030 36.750 46092 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.300 44.642 46094 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.304 44.639 46097 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.313 40.753 46212 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.353 32.848 46226 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.320 32.936 46233 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.167 32.572 46235 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.439 33.220 46242 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -119.708 33.821 46249 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -119.090 34.034 46250 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -119.564 33.769 46251 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -119.257 33.953 46252 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -118.181 33.576 46253 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.267 32.868 46254 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -118.201 33.700 46256 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -120.766 34.439 46257 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.500 32.750 46258 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -119.004 33.704 46262 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -126.000 48.840 46206 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.510 46.120 46029 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.750 47.340 46041 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.730 48.490 46087 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.240 46.860 46211 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -123.165 48.334 46088 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.127 46.173 46096 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.566 46.986 46099 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.950 47.967 46119 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.063 46.215 46127 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -126.010 48.844 46139 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 126.020 37.230 22101 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 128.900 34.770 22104 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 125.750 36.250 22108 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 129.840 35.350 22189 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 151.070 -23.310 55033 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -88.770 30.090 42007 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -94.400 29.220 42035 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -87.550 30.060 42012 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -88.647 30.042 42067 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -66.500 19.000 PUERTO_R_N - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -74.681 11.161 41194 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -66.524 17.860 42085 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -67.280 18.379 41115 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -81.080 30.000 41117 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -80.870 31.400 41008 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -80.170 28.500 41009 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -81.292 30.709 41112 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -80.530 28.400 41113 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -80.220 27.550 41114 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -67.280 18.380 41111 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -66.099 18.474 41053 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -78.484 33.837 41024 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -79.624 32.803 41029 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -80.410 32.279 41033 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -75.714 36.200 44056 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -75.720 36.915 44099 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -75.590 36.260 44100 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -75.809 37.023 44096 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -70.141 43.525 44007 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -70.428 43.181 44030 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -70.060 43.570 44031 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -69.355 43.716 44032 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -75.183 38.883 44054 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -75.256 39.122 44055 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -76.257 37.567 44058 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -76.415 38.556 44062 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -76.448 38.963 44063 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -76.087 36.998 44064 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -73.703 40.369 44065 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -76.266 37.201 44072 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -73.106 40.585 44094 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -63.408 44.500 44172 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -76.149 37.024 44087 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -48.130 -27.700 31201 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -5.783 35.900 TANGER - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -9.633 30.417 AGADIR - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -7.617 33.700 CASABLANCA - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 100.000 4.000 P_PANGKOR - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -5.420 53.470 62091 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -9.070 54.670 62093 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -4.700 52.300 62301 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 0.000 50.400 62305 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -15.200 64.000 TFHFN - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 1.900 51.100 62304 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -7.620 44.070 62082 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -9.400 42.120 62084 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -6.970 36.480 62085 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -13.500 65.650 TFKGR - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -146.040 70.370 48211 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -150.279 70.874 48212 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 73.750 15.400 23093 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 74.500 12.940 23094 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 80.390 13.190 23096 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 121.900 -34.000 56010 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 144.850 -38.600 PNEPEAN - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 27.933 -33.033 EASTLONDON - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 17.900 -33.055 SALDANHABY - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 22.170 -34.150 MOSSELBAY - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -121.500 35.750 CDIP01 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -116.250 30.750 CDIP24 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 120.000 15.000 BKMG50 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -77.500 6.260 32488 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -77.740 3.520 32487 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -72.220 12.350 41193 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 129.780 36.350 22106 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 42.833 -22.000 V14078 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.800 32.333 SITE3 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.800 32.800 SITEX - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -80.595 28.615 CAPC - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -80.250 28.250 CAP1 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -80.600 28.800 DAB - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -86.028 29.766 PAN - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -81.300 30.670 CHS51 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -78.500 33.580 CHS60 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -78.900 33.580 CHS61 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -80.250 28.500 JAX51 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -80.000 32.500 JAX58 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -66.800 44.500 MNE65 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -96.910 25.260 BRO51 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -97.020 25.750 CRP51 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -95.050 29.090 CRP68 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -97.000 26.000 CCTX51 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -85.900 30.000 MOB61 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -87.350 30.150 PCB57 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -83.000 28.350 PCB71 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -79.800 23.000 KEY64 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -79.600 23.000 KEY65 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -79.400 23.000 KEY66 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -80.000 26.000 KEY76 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -82.200 26.000 KEY77 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -70.350 43.250 NW-BOX68 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -70.480 43.250 NW-BOX69 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -73.350 40.600 NW-PHI60 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -75.650 36.750 NW-LWX52 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -75.900 36.750 NW-LWX53 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -80.150 32.500 NW-ILM51 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -79.900 32.500 NW-ILM52 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -67.850 44.400 NW-GYX62 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -75.750 36.600 NW-MHX66 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -75.830 36.600 NW-MHX67 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -119.000 33.580 NW-SGX57 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -119.000 33.820 NW-SGX58 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -118.750 33.820 NW-SGX59 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -118.500 33.820 NW-SGX60 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -116.930 32.080 NW-SGX61 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.000 32.080 NW-SGX62 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.250 32.080 NW-SGX63 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.500 32.080 NW-SGX64 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.750 32.080 NW-SGX65 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -121.860 36.100 NW-LOX62 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.860 32.600 NW-LOX74 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.360 32.600 NW-LOX75 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -121.300 35.000 NW-MTR69 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -120.800 35.000 NW-MTR70 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.050 39.400 NW-MTR75 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.300 44.450 NW-MFR64 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.300 41.000 NW-MFR65 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.780 47.150 NW-PQR62 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.280 47.150 NW-PQR63 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.300 43.500 NW-PQR64 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.800 43.500 NW-PQR65 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -126.750 49.420 NW-SEW59 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.160 46.100 NW-SEW60 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.410 46.100 NW-SEW61 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.770 42.200 NW-EKA62 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.640 42.200 NW-EKA63 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -123.700 38.400 NW-EKA64 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -124.200 38.400 NW-EKA65 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -144.500 60.000 NW-AJK64 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -131.000 53.900 NW-AJK65 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -158.000 57.750 NW-AER54 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -142.850 59.850 NW-AER56 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -166.500 62.000 NW-ALU143 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -166.000 62.000 NW-ALU144 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -152.400 60.200 NW-ALU145 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -152.400 59.700 NW-ALU146 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -139.000 69.750 NW-AFG133 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -166.500 61.000 NW-AFG171 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -166.000 61.000 NW-AFG172 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 127.000 1.500 HWRFc-52 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 128.000 1.500 HWRFc-53 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 125.000 39.500 HWRFc-274 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 125.000 38.700 HWRFc-275 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 125.000 9.000 HWRFc-305 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 125.000 5.750 HWRFc-306 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 148.200 -42.500 HWRFp-61 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 128.000 -2.500 HWRFs-109 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 107.000 -2.500 HWRFs-115 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 101.000 -2.500 HWRFs-116 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : 127.000 1.500 HWRFw-53 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -98.000 15.500 RW-NH2-76 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -98.000 16.000 RW-NH2-77 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -80.750 32.000 RW-NH2-78 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -10.000 32.000 RW-NH2-220 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -9.500 32.000 RW-NH2-221 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -10.000 5.500 RW-NH2-222 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -10.000 5.000 RW-NH2-223 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -50.750 3.000 RW-NH2-328 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -77.750 3.000 RW-NH2-329 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -118.000 32.800 RW-NH1-203 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -117.500 32.800 RW-NH1-204 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -113.500 31.000 RW-NH1-205 - POINT SKIPPPED - - - *** WAVEWATCH-III WARNING : - OUTPUT POINT OUT OF GRID : -70.500 -19.500 RW-NH1-265 - POINT SKIPPPED - - - Grid name : Global Unstructured 100km - - No water levels. - Prescribed curents. - Prescribed winds. - Prescribed ice fields. - No momentum - No air density - No mud density. - No mud thickness. - No mud viscosity. - No ice parameter 1 - No ice parameter 2 - No ice parameter 3 - No ice parameter 4 - No ice parameter 5 - - Gridded output fields : --------------------------------------------------- - Current vel. - Wind speed - Ice concentration - Wave height - Mean wave period(+2) - Mean wave period(+1) - Peak frequency - Mean wave dir. a1b1 - Peak direction - Part. wave height - Part. peak period - Part. mean direction - Charnock parameter - - Point output requested for 3820 points : ------------------------------------------- - - point | longitude | latitude | name - --------|-------------|-------------|---------------- - 1 | -148.02 | 56.31 | 46001 - 2 | -154.98 | 52.70 | 46066 - 3 | -160.81 | 53.93 | 46075 - 4 | -148.00 | 59.50 | 46076 - 5 | -152.45 | 56.05 | 46078 - 6 | -150.00 | 58.00 | 46080 - 7 | -160.00 | 57.70 | 46021 - 8 | -152.23 | 59.05 | 46079 - 9 | -152.23 | 59.05 | 46105 - 10 | -147.99 | 59.92 | 46107 - 11 | -165.48 | 64.46 | 46265 - 12 | -136.10 | 50.93 | 46004 - 13 | -138.85 | 53.91 | 46184 - 14 | -143.42 | 59.69 | 46082 - 15 | -136.16 | 56.59 | 46084 - 16 | -142.56 | 56.85 | 46085 - 17 | -134.28 | 54.16 | 46205 - 18 | -132.45 | 54.38 | 46145 - 19 | -131.22 | 51.83 | 46147 - 20 | -128.75 | 51.37 | 46204 - 21 | -129.92 | 50.87 | 46207 - 22 | -132.68 | 52.52 | 46208 - 23 | -130.27 | 42.60 | 46002 - 24 | -137.48 | 40.80 | 46006 - 25 | -130.00 | 37.98 | 46059 - 26 | -122.88 | 37.36 | 46012 - 27 | -124.54 | 40.78 | 46022 - 28 | -122.82 | 37.75 | 46026 - 29 | -124.85 | 42.75 | 46015 - 30 | -121.89 | 35.74 | 46028 - 31 | -124.53 | 40.42 | 46030 - 32 | -122.42 | 36.75 | 46042 - 33 | -119.53 | 32.43 | 46047 - 34 | -124.53 | 44.62 | 46050 - 35 | -120.45 | 34.27 | 46054 - 36 | -120.70 | 34.27 | 46063 - 37 | -120.20 | 33.65 | 46069 - 38 | -125.77 | 45.88 | 46089 - 39 | -124.74 | 40.29 | 46213 - 40 | -123.46 | 37.94 | 46214 - 41 | -119.88 | 33.22 | 46219 - 42 | -118.00 | 30.00 | TPC50 - 43 | -135.00 | 20.00 | TPC51 - 44 | -117.00 | 20.00 | TPC52 - 45 | -120.00 | 6.00 | TPC53 - 46 | -95.00 | 15.00 | TPC54 - 47 | -88.00 | 9.00 | TPC55 - 48 | -80.00 | 6.00 | TPC56 - 49 | -130.50 | 48.10 | OPCP01 - 50 | -126.60 | 48.10 | OPCP02 - 51 | -129.70 | 45.30 | OPCP03 - 52 | -125.60 | 45.30 | OPCP04 - 53 | -129.90 | 41.75 | OPCP05 - 54 | -125.80 | 41.90 | OPCP06 - 55 | -129.00 | 38.50 | OPCP07 - 56 | -125.50 | 39.20 | OPCP08 - 57 | -125.40 | 36.40 | OPCP09 - 58 | -125.00 | 33.30 | OPCP10 - 59 | -122.30 | 34.60 | OPCP11 - 60 | -121.50 | 30.90 | OPCP12 - 61 | -117.00 | 29.60 | OPCP13 - 62 | -119.47 | 33.40 | 46238 - 63 | -122.10 | 36.34 | 46239 - 64 | -145.20 | 50.03 | 46246 - 65 | -119.20 | 33.00 | 46024 - 66 | -122.41 | 36.69 | 46093 - 67 | -124.96 | 44.38 | 46098 - 68 | -122.33 | 36.69 | 46114 - 69 | -122.83 | 37.75 | 46247 - 70 | -119.65 | 33.40 | 46255 - 71 | -121.50 | 34.77 | 46259 - 72 | -131.02 | 46.05 | 46005 - 73 | -133.94 | 48.35 | 46036 - 74 | -127.93 | 49.74 | 46132 - 75 | -124.97 | 46.85 | 46100 - 76 | -151.70 | 57.48 | 46264 - 77 | -177.58 | 57.05 | 46035 - 78 | 175.28 | 55.00 | 46070 - 79 | -172.03 | 54.94 | 46073 - 80 | 179.05 | 51.16 | 46071 - 81 | -171.73 | 52.25 | 46072 - 82 | -168.00 | 55.88 | 46020 - 83 | -162.21 | 23.43 | 51001 - 84 | -157.78 | 17.19 | 51002 - 85 | -160.82 | 19.22 | 51003 - 86 | -152.48 | 17.52 | 51004 - 87 | -158.00 | 24.00 | HNL01 - 88 | -153.00 | 22.50 | HNL02 - 89 | -157.75 | 22.00 | HNL10 - 90 | -158.25 | 21.00 | HNL11 - 91 | -156.50 | 19.75 | HNL12 - 92 | -158.12 | 21.67 | 51201 - 93 | -157.68 | 21.42 | 51202 - 94 | -177.40 | 28.20 | MIDWAY - 95 | -169.50 | 16.70 | JOHNSTON - 96 | -166.30 | 23.90 | FF_SHOALS - 97 | -154.06 | 23.55 | 51000 - 98 | -153.90 | 23.56 | 51100 - 99 | -162.06 | 24.32 | 51101 - 100 | -157.00 | 20.79 | 51203 - 101 | -158.12 | 21.28 | 51204 - 102 | -156.42 | 21.02 | 51205 - 103 | -154.97 | 19.78 | 51206 - 104 | -153.87 | 0.02 | 51028 - 105 | -158.30 | 21.10 | 51200 - 106 | -159.57 | 22.28 | 51208 - 107 | -170.50 | -14.27 | 51209 - 108 | 134.67 | 7.69 | 52212 - 109 | -157.96 | 21.30 | 51211 - 110 | -158.15 | 21.32 | 51212 - 111 | -157.00 | 20.75 | 51213 - 112 | 147.50 | 16.00 | SAIPAN - 113 | 143.75 | 12.00 | GUAM - 114 | 166.50 | 19.50 | WAKE - 115 | 136.25 | 9.00 | PALAU - 116 | 138.00 | 9.60 | YAP - 117 | 152.50 | 8.00 | CHUUK - 118 | 157.50 | 7.00 | POHNPEI - 119 | 163.00 | 5.10 | KOSRAE - 120 | 171.25 | 8.00 | MAJURO - 121 | 163.75 | 13.00 | ENEWETAK - 122 | 174.00 | 1.00 | TARAWA - 123 | 144.79 | 13.35 | 52200 - 124 | 125.77 | 34.80 | 22102 - 125 | 127.50 | 34.00 | 22103 - 126 | 130.00 | 37.53 | 22105 - 127 | 171.39 | 7.04 | 52201 - 128 | 144.80 | 13.68 | 52202 - 129 | 145.66 | 15.27 | 52211 - 130 | 145.00 | 14.70 | SAIPAN_W - 131 | 145.60 | 14.30 | SAIPAN_E - 132 | 133.90 | 7.50 | PALAU_W - 133 | 134.90 | 7.70 | PALAU_E - 134 | 138.40 | 9.60 | YAP_E - 135 | 151.20 | 7.40 | CHUUK_W - 136 | 152.10 | 7.60 | CHUUK_E - 137 | 158.40 | 7.10 | POHNPEI_E - 138 | 162.80 | 5.90 | KOSRAE_W - 139 | 167.80 | 9.50 | KWAJALEIN - 140 | 167.50 | 8.67 | KWAJ_W1 - 141 | 167.00 | 8.67 | KWAJ_W2 - 142 | 168.17 | 8.67 | KWAJ_E - 143 | 166.33 | 9.17 | WOTHO - 144 | 168.00 | 9.17 | ROI_NAMUR - 145 | 171.50 | 9.17 | WOTJE_E - 146 | 171.40 | 7.60 | MAJURO_02 - 147 | 171.50 | 6.60 | MAJURO_03 - 148 | 133.62 | 33.19 | 21178 - 149 | 131.11 | 37.46 | 21229 - 150 | 126.14 | 33.79 | 22184 - 151 | 125.43 | 37.09 | 22185 - 152 | 125.81 | 35.66 | 22186 - 153 | 127.02 | 33.13 | 22187 - 154 | 128.23 | 34.39 | 22188 - 155 | 129.87 | 36.91 | 22190 - 156 | 146.25 | -12.00 | NEWGUINE_S - 157 | -168.75 | -15.00 | PAGO_PAGO - 158 | 176.25 | -18.00 | NADI - 159 | 179.20 | -8.50 | FUNAFUTI - 160 | -175.00 | -22.00 | TONGATAPU - 161 | -159.80 | -21.20 | RAROTONGA - 162 | 167.50 | -24.00 | NOUMEA - 163 | 167.50 | -18.00 | PORT_VILA - 164 | -149.60 | -19.00 | PAPEETE - 165 | -169.90 | -19.10 | NIUE - 166 | -171.90 | -9.20 | NUKUNONO - 167 | 167.00 | -0.50 | NAURU - 168 | 160.00 | -12.00 | SOLOMON_SW - 169 | 165.00 | -12.00 | SOLOMON_SE - 170 | 160.00 | -5.00 | SOLOMON_N - 171 | -85.00 | -15.00 | TPC01 - 172 | -110.00 | -15.00 | TPC02 - 173 | -135.00 | -15.00 | TPC03 - 174 | 150.18 | -37.29 | 55020 - 175 | 153.63 | -27.49 | 55035 - 176 | 148.19 | -38.60 | 55039 - 177 | 118.00 | -1.00 | SULAWESI - 178 | 120.00 | -7.50 | P_BONARAT - 179 | 125.00 | -5.00 | P_RUNDUMA - 180 | 133.00 | -8.00 | P_JAMDENA - 181 | 122.00 | -11.00 | P_SAWA - 182 | -89.67 | 25.90 | 42001 - 183 | -94.42 | 25.17 | 42002 - 184 | -85.94 | 26.07 | 42003 - 185 | -95.36 | 27.91 | 42019 - 186 | -96.70 | 26.94 | 42020 - 187 | -84.52 | 28.50 | 42036 - 188 | -86.02 | 28.79 | 42039 - 189 | -88.21 | 29.18 | 42040 - 190 | -90.46 | 27.50 | 42041 - 191 | -92.55 | 27.42 | 42038 - 192 | -94.05 | 22.01 | 42055 - 193 | -84.28 | 27.35 | 42099 - 194 | -86.00 | 23.00 | TPC26 - 195 | -88.49 | 28.19 | 42887 - 196 | -87.73 | 26.00 | 42054 - 197 | -82.92 | 27.17 | 42013 - 198 | -82.22 | 25.25 | 42014 - 199 | -83.31 | 28.31 | 42021 - 200 | -83.74 | 27.50 | 42022 - 201 | -83.09 | 26.01 | 42023 - 202 | -94.90 | 28.98 | 42043 - 203 | -97.05 | 26.19 | 42044 - 204 | -96.50 | 26.22 | 42045 - 205 | -94.04 | 27.89 | 42046 - 206 | -93.60 | 27.90 | 42047 - 207 | -83.65 | 25.70 | 42097 - 208 | -82.93 | 27.59 | 42098 - 209 | -90.47 | 26.67 | 42360 - 210 | -92.49 | 27.55 | 42361 - 211 | -90.65 | 27.80 | 42362 - 212 | -89.22 | 28.16 | 42363 - 213 | -88.09 | 29.06 | 42364 - 214 | -89.12 | 28.20 | 42365 - 215 | -90.28 | 27.21 | 42369 - 216 | -90.54 | 27.32 | 42370 - 217 | -88.06 | 28.87 | 42374 - 218 | -88.29 | 28.52 | 42375 - 219 | -87.94 | 29.11 | 42376 - 220 | -94.90 | 26.13 | 42390 - 221 | -90.03 | 27.20 | 42392 - 222 | -89.24 | 28.16 | 42394 - 223 | -90.79 | 26.40 | 42395 - 224 | -85.06 | 19.87 | 42056 - 225 | -81.50 | 16.83 | 42057 - 226 | -75.06 | 15.09 | 42058 - 227 | -81.95 | 24.39 | 42080 - 228 | -81.75 | 24.00 | EYW01 - 229 | -82.25 | 25.00 | EYW02 - 230 | -66.50 | 17.50 | PUERTO_R_S - 231 | -67.50 | 15.01 | 42059 - 232 | -85.38 | -19.62 | 32012 - 233 | -63.00 | 15.00 | TPC21 - 234 | -77.00 | 12.00 | TPC22 - 235 | -80.00 | 15.00 | TPC23 - 236 | -76.00 | 22.00 | TPC24 - 237 | -80.00 | 24.00 | TPC25 - 238 | -63.50 | 16.50 | 42060 - 239 | -67.50 | 19.00 | CARCOOS01 - 240 | -65.50 | 19.00 | CARCOOS02 - 241 | -64.00 | 19.00 | CARCOOS03 - 242 | -64.40 | 17.30 | CARCOOS04 - 243 | -67.50 | 17.50 | CARCOOS05 - 244 | -80.06 | 19.70 | 42089 - 245 | -64.76 | 18.25 | 41052 - 246 | -65.00 | 18.26 | 41051 - 247 | -65.46 | 18.26 | 41056 - 248 | -81.24 | 24.53 | 42079 - 249 | -75.04 | 36.00 | 42086 - 250 | -81.97 | 24.41 | 42095 - 251 | -72.66 | 34.68 | 41001 - 252 | -75.36 | 32.32 | 41002 - 253 | -79.09 | 32.50 | 41004 - 254 | -78.47 | 28.95 | 41010 - 255 | -80.60 | 30.00 | 41012 - 256 | -77.74 | 33.44 | 41013 - 257 | -75.40 | 35.01 | 41025 - 258 | -77.28 | 34.48 | 41035 - 259 | -76.95 | 34.21 | 41036 - 260 | -65.01 | 20.99 | 41043 - 261 | -70.99 | 24.00 | 41046 - 262 | -71.49 | 27.47 | 41047 - 263 | -69.65 | 31.98 | 41048 - 264 | -74.84 | 36.61 | 44014 - 265 | -77.00 | 30.75 | CHS01 - 266 | -77.00 | 34.00 | ILM01 - 267 | -78.50 | 33.25 | ILM02 - 268 | -80.25 | 29.50 | JAX02 - 269 | -79.50 | 27.25 | MLB01 - 270 | -79.50 | 26.25 | MIA01 - 271 | -79.75 | 25.00 | MIA02 - 272 | -74.60 | 36.30 | OPCA05 - 273 | -73.80 | 35.60 | OPCA06 - 274 | -70.80 | 34.90 | OPCA07 - 275 | -76.00 | 33.80 | OPCA08 - 276 | -72.30 | 32.80 | OPCA09 - 277 | -77.36 | 33.99 | 41037 - 278 | -77.72 | 34.14 | 41038 - 279 | -63.00 | 27.50 | 41049 - 280 | -58.69 | 21.65 | 41044 - 281 | -77.30 | 34.48 | 41109 - 282 | -77.71 | 34.14 | 41110 - 283 | -65.16 | 18.48 | 41058 - 284 | -78.14 | 33.30 | 41027 - 285 | -79.34 | 32.52 | 41030 - 286 | -38.00 | 24.58 | 41061 - 287 | -75.10 | 35.78 | 41062 - 288 | -75.94 | 34.78 | 41063 - 289 | -76.95 | 34.21 | 41064 - 290 | -78.01 | 33.72 | 41108 - 291 | -76.95 | 34.21 | 41159 - 292 | -53.62 | 44.26 | 44138 - 293 | -66.58 | 41.11 | 44011 - 294 | -58.00 | 43.00 | 44141 - 295 | -64.02 | 42.50 | 44142 - 296 | -48.01 | 46.77 | WRB07 - 297 | -62.00 | 42.26 | 44137 - 298 | -57.08 | 44.26 | 44139 - 299 | -51.74 | 43.75 | 44140 - 300 | -64.01 | 42.50 | 44150 - 301 | -70.43 | 38.48 | 44004 - 302 | -69.16 | 43.19 | 44005 - 303 | -69.43 | 40.50 | 44008 - 304 | -74.70 | 38.46 | 44009 - 305 | -72.10 | 40.70 | 44017 - 306 | -69.29 | 41.26 | 44018 - 307 | -73.17 | 40.25 | 44025 - 308 | -70.25 | 42.50 | BOX01 - 309 | -67.50 | 44.00 | CAR01 - 310 | -71.01 | 41.38 | 44070 - 311 | -65.93 | 42.31 | 44024 - 312 | -67.31 | 44.27 | 44027 - 313 | -67.88 | 43.49 | 44037 - 314 | -66.55 | 43.62 | 44038 - 315 | -53.39 | 46.44 | 44251 - 316 | -57.35 | 47.28 | 44255 - 317 | -69.75 | 43.25 | GYX01 - 318 | -67.70 | 42.35 | OPCA01 - 319 | -72.00 | 39.30 | OPCA02 - 320 | -65.70 | 39.30 | OPCA03 - 321 | -70.10 | 37.30 | OPCA04 - 322 | -72.60 | 39.58 | 44066 - 323 | -75.49 | 36.87 | 44093 - 324 | -75.33 | 35.75 | 44095 - 325 | -71.13 | 40.97 | 44097 - 326 | -70.17 | 42.80 | 44098 - 327 | -70.65 | 42.35 | 44013 - 328 | -70.19 | 41.44 | 44020 - 329 | -70.57 | 42.52 | 44029 - 330 | -69.00 | 44.06 | 44033 - 331 | -68.11 | 44.11 | 44034 - 332 | -72.65 | 41.14 | 44039 - 333 | -72.07 | 41.26 | 44060 - 334 | -75.33 | 37.76 | 44089 - 335 | -70.33 | 41.84 | 44090 - 336 | -73.77 | 39.77 | 44091 - 337 | -70.63 | 42.94 | 44092 - 338 | -57.34 | 47.26 | 44235 - 339 | -48.75 | -32.00 | RIO_GRANDE - 340 | -46.25 | -28.00 | FLORIPA - 341 | -43.75 | -25.00 | SANTOS - 342 | -34.57 | -8.15 | 31052 - 343 | -43.09 | -23.03 | 31260 - 344 | -47.37 | -28.50 | 31374 - 345 | -44.93 | -25.28 | 31051 - 346 | -51.35 | -32.60 | 31053 - 347 | -42.19 | -22.99 | SIODOC - 348 | -44.27 | -23.42 | ILHAGRANDE - 349 | -43.46 | -23.16 | RECREIO - 350 | -43.12 | -23.11 | CAGARRAS - 351 | -49.81 | -31.52 | 31375 - 352 | -53.08 | 14.55 | 41040 - 353 | -46.00 | 14.53 | 41041 - 354 | -57.90 | 15.90 | 41100 - 355 | -56.20 | 14.60 | 41101 - 356 | -55.00 | 15.00 | TPC20 - 357 | -38.75 | -21.00 | CAMPOS - 358 | -36.25 | -13.00 | SALVADOR - 359 | -32.50 | -8.00 | RECIFE - 360 | -36.25 | -3.00 | FORTALEZA - 361 | -47.50 | 3.00 | AMAZON - 362 | -30.00 | 1.00 | PETER_PAUL - 363 | 3.50 | 6.00 | LAGOS - 364 | 5.00 | 5.00 | FORCADOS - 365 | 7.00 | 4.00 | NIGERDELTA - 366 | 8.00 | 4.00 | CALABAR - 367 | -50.95 | 14.75 | 41060 - 368 | -60.85 | 11.19 | 42087 - 369 | -60.52 | 11.30 | 42088 - 370 | -83.75 | 0.00 | TPC00 - 371 | -93.75 | 0.00 | TPC04 - 372 | 102.00 | -5.00 | P_ENGGANO - 373 | 107.00 | 0.00 | P_PENJAN - 374 | 110.00 | -5.00 | SEMARANG - 375 | 132.00 | 1.00 | P_IGI - 376 | 123.00 | 3.00 | BORNEO - 377 | 126.00 | 1.00 | P_GUREDA - 378 | -125.03 | 10.05 | 43010 - 379 | -144.67 | 13.73 | 52009 - 380 | -5.00 | 45.20 | 62001 - 381 | -20.00 | 41.60 | 62002 - 382 | -12.40 | 48.70 | 62029 - 383 | -7.90 | 51.40 | 62023 - 384 | -5.60 | 48.50 | 62052 - 385 | -13.30 | 51.00 | 62081 - 386 | -11.20 | 53.13 | 62090 - 387 | -10.55 | 51.22 | 62092 - 388 | -6.70 | 51.69 | 62094 - 389 | -15.92 | 53.06 | 62095 - 390 | -2.90 | 49.90 | 62103 - 391 | -12.36 | 54.54 | 62105 - 392 | -9.90 | 57.00 | 62106 - 393 | -6.10 | 50.10 | 62107 - 394 | -19.50 | 53.50 | 62108 - 395 | -8.50 | 47.50 | 62163 - 396 | -5.10 | 51.60 | 62303 - 397 | 2.00 | 51.40 | 62170 - 398 | -11.40 | 59.10 | 64045 - 399 | -4.50 | 60.70 | 64046 - 400 | -23.10 | 64.05 | TFGSK - 401 | -20.35 | 63.00 | TFSRT - 402 | 7.80 | 64.30 | LF3F - 403 | 1.10 | 55.30 | 62026 - 404 | 0.00 | 57.00 | 62109 - 405 | 0.40 | 58.10 | 62111 - 406 | 1.30 | 58.70 | 62112 - 407 | 1.40 | 57.70 | 62116 - 408 | 0.00 | 57.90 | 62117 - 409 | 2.00 | 57.00 | 62119 - 410 | 1.40 | 58.70 | 62128 - 411 | 2.00 | 56.40 | 62132 - 412 | 1.00 | 57.10 | 62133 - 413 | 2.10 | 53.00 | 62142 - 414 | 1.80 | 57.70 | 62143 - 415 | 1.70 | 53.40 | 62144 - 416 | 2.80 | 53.10 | 62145 - 417 | 1.80 | 57.00 | 62152 - 418 | 0.50 | 57.40 | 62162 - 419 | 0.50 | 57.20 | 62164 - 420 | 1.70 | 60.60 | 63055 - 421 | 1.60 | 59.50 | 63056 - 422 | 1.50 | 59.20 | 63057 - 423 | 1.10 | 61.20 | 63103 - 424 | 1.70 | 60.80 | 63108 - 425 | 1.50 | 59.50 | 63110 - 426 | 1.00 | 61.10 | 63112 - 427 | 1.70 | 61.00 | 63113 - 428 | 1.30 | 61.60 | 63115 - 429 | 2.30 | 61.20 | LF3J - 430 | 3.70 | 60.60 | LF4B - 431 | 2.20 | 59.60 | LF4H - 432 | 1.90 | 58.40 | LF4C - 433 | 3.20 | 56.50 | LF5U - 434 | 3.28 | 51.99 | EURO - 435 | 3.22 | 53.22 | K13 - 436 | -3.03 | 43.63 | 62024 - 437 | -15.82 | 28.18 | 13130 - 438 | -16.58 | 28.00 | 13131 - 439 | 0.90 | 57.70 | 62118 - 440 | 2.10 | 57.10 | 62146 - 441 | 1.50 | 59.50 | 63111 - 442 | 6.33 | 55.00 | BSH01 - 443 | 7.89 | 54.16 | BSH02 - 444 | 8.12 | 54.00 | BSH03 - 445 | 6.58 | 54.00 | BSH04 - 446 | 8.22 | 54.92 | BSH05 - 447 | -25.00 | 65.69 | TFBLK - 448 | -18.20 | 66.50 | TFGRS - 449 | 7.30 | 65.30 | LF3N - 450 | 8.10 | 66.00 | LF5T - 451 | 2.00 | 66.00 | LDWR - 452 | 21.10 | 71.60 | 3FYT - 453 | 15.50 | 73.50 | LFB1 - 454 | 30.00 | 74.00 | LFB2 - 455 | -9.26 | 68.48 | 64071 - 456 | -166.07 | 70.03 | 48012 - 457 | -169.45 | 65.01 | 48114 - 458 | -164.13 | 71.50 | 48213 - 459 | -165.25 | 70.87 | 48214 - 460 | -167.95 | 71.76 | 48216 - 461 | 72.49 | 17.02 | 23092 - 462 | 69.24 | 15.47 | 23097 - 463 | 72.51 | 10.65 | 23098 - 464 | 90.74 | 12.14 | 23099 - 465 | 87.56 | 18.35 | 23100 - 466 | 83.27 | 13.97 | 23101 - 467 | 87.50 | 15.00 | 23168 - 468 | 90.14 | 18.13 | 23169 - 469 | 72.66 | 8.33 | 23170 - 470 | 72.00 | 12.50 | 23172 - 471 | 78.57 | 8.21 | 23173 - 472 | 81.53 | 11.57 | 23174 - 473 | 93.00 | 6.00 | G_NICOBAR - 474 | 116.14 | -19.59 | 56002 - 475 | 115.40 | -32.11 | 56005 - 476 | 114.78 | -33.36 | 56006 - 477 | 114.94 | -21.41 | 56007 - 478 | 22.17 | -34.97 | AGULHAS_FA - 479 | 104.00 | 6.00 | MALAY01 - 480 | 105.00 | 3.00 | MALAY02 - 481 | 110.00 | 3.00 | MALAY03 - 482 | 113.00 | 5.00 | MALAY04 - 483 | 116.00 | 7.50 | MALAY05 - 484 | 117.00 | 7.50 | MALAY06 - 485 | 114.10 | -21.70 | 56012 - 486 | 136.20 | -36.10 | CADUCOU - 487 | 139.00 | -38.00 | SWROBE - 488 | 142.45 | -39.20 | WBAST1 - 489 | 141.50 | -40.00 | WBAST2 - 490 | 151.00 | -40.00 | EBAST - 491 | 146.50 | -40.50 | CBAST - 492 | 144.60 | -42.30 | CSORRELL - 493 | 144.50 | -40.10 | SEKING - 494 | 143.80 | -39.20 | NKING - 495 | 147.40 | -39.20 | EHOGAN - 496 | 147.00 | -44.00 | STHSEC - 497 | 149.50 | -41.50 | EBICHENO - 498 | 133.50 | -33.50 | WCAPYORK - 499 | 114.62 | -19.78 | JANSZ - 500 | 85.00 | 12.60 | 23167 - 501 | 70.00 | 11.02 | 23171 - 502 | 91.66 | 10.52 | 23451 - 503 | 89.04 | 10.97 | 23455 - 504 | 86.98 | 9.99 | 23456 - 505 | 70.10 | 5.16 | 23491 - 506 | 68.08 | 13.89 | 23492 - 507 | 66.98 | 11.12 | 23493 - 508 | 75.00 | 6.46 | 23494 - 509 | 68.97 | 7.13 | 23495 - 510 | 24.99 | -34.06 | JEFFREYBAY - 511 | 18.28 | -34.20 | CAPEPOINT - 512 | 22.17 | -34.97 | FAPLATFORM - 513 | -160.00 | 72.00 | Alaska_NS1 - 514 | -150.00 | 72.00 | Alaska_NS2 - 515 | -140.00 | 72.00 | Alaska_NS3 - 516 | -116.48 | 23.62 | EFT1 - 517 | -123.75 | 36.00 | DIABLO_01 - 518 | -122.00 | 35.75 | CDIP02 - 519 | -122.00 | 35.25 | CDIP03 - 520 | -122.00 | 34.75 | CDIP04 - 521 | -122.00 | 34.25 | CDIP05 - 522 | -122.00 | 33.75 | CDIP06 - 523 | -122.00 | 33.25 | CDIP07 - 524 | -122.00 | 32.75 | CDIP08 - 525 | -122.00 | 32.25 | CDIP09 - 526 | -122.00 | 31.75 | CDIP10 - 527 | -122.00 | 31.25 | CDIP11 - 528 | -122.00 | 30.75 | CDIP12 - 529 | -121.75 | 30.75 | CDIP13 - 530 | -121.25 | 30.75 | CDIP14 - 531 | -120.75 | 30.75 | CDIP15 - 532 | -120.25 | 30.75 | CDIP16 - 533 | -119.75 | 30.75 | CDIP17 - 534 | -119.25 | 30.75 | CDIP18 - 535 | -118.75 | 30.75 | CDIP19 - 536 | -118.25 | 30.75 | CDIP20 - 537 | -117.75 | 30.75 | CDIP21 - 538 | -117.25 | 30.75 | CDIP22 - 539 | -116.75 | 30.75 | CDIP23 - 540 | 90.00 | -15.00 | BKMG01 - 541 | 92.00 | -15.00 | BKMG02 - 542 | 94.00 | -15.00 | BKMG03 - 543 | 96.00 | -15.00 | BKMG04 - 544 | 98.00 | -15.00 | BKMG05 - 545 | 100.00 | -15.00 | BKMG06 - 546 | 102.00 | -15.00 | BKMG07 - 547 | 104.00 | -15.00 | BKMG08 - 548 | 106.00 | -15.00 | BKMG09 - 549 | 108.00 | -15.00 | BKMG10 - 550 | 110.00 | -15.00 | BKMG11 - 551 | 112.00 | -15.00 | BKMG12 - 552 | 114.00 | -15.00 | BKMG13 - 553 | 116.00 | -15.00 | BKMG14 - 554 | 118.00 | -15.00 | BKMG15 - 555 | 120.00 | -15.00 | BKMG16 - 556 | 122.00 | -15.00 | BKMG17 - 557 | 124.00 | -15.00 | BKMG18 - 558 | 124.50 | -15.00 | BKMG19 - 559 | 137.00 | -15.00 | BKMG20 - 560 | 139.00 | -15.00 | BKMG21 - 561 | 141.00 | -15.00 | BKMG22 - 562 | 145.00 | -13.50 | BKMG23 - 563 | 145.00 | -13.00 | BKMG24 - 564 | 145.00 | -11.00 | BKMG25 - 565 | 145.00 | -9.00 | BKMG26 - 566 | 145.00 | -8.00 | BKMG27 - 567 | 145.00 | -3.00 | BKMG28 - 568 | 145.00 | -1.00 | BKMG29 - 569 | 145.00 | 1.00 | BKMG30 - 570 | 145.00 | 3.00 | BKMG31 - 571 | 145.00 | 5.00 | BKMG32 - 572 | 145.00 | 7.00 | BKMG33 - 573 | 145.00 | 9.00 | BKMG34 - 574 | 145.00 | 11.00 | BKMG35 - 575 | 145.00 | 13.00 | BKMG36 - 576 | 145.00 | 15.00 | BKMG37 - 577 | 143.00 | 15.00 | BKMG38 - 578 | 141.00 | 15.00 | BKMG39 - 579 | 139.00 | 15.00 | BKMG40 - 580 | 137.00 | 15.00 | BKMG41 - 581 | 135.00 | 15.00 | BKMG42 - 582 | 133.00 | 15.00 | BKMG43 - 583 | 131.00 | 15.00 | BKMG44 - 584 | 129.00 | 15.00 | BKMG45 - 585 | 127.00 | 15.00 | BKMG46 - 586 | 125.00 | 15.00 | BKMG47 - 587 | 123.00 | 15.00 | BKMG48 - 588 | 122.50 | 15.00 | BKMG49 - 589 | 118.00 | 15.00 | BKMG51 - 590 | 116.00 | 15.00 | BKMG52 - 591 | 114.00 | 15.00 | BKMG53 - 592 | 112.00 | 15.00 | BKMG54 - 593 | 110.00 | 15.00 | BKMG55 - 594 | 97.50 | 15.00 | BKMG56 - 595 | 97.00 | 15.00 | BKMG57 - 596 | 95.00 | 15.00 | BKMG58 - 597 | 93.00 | 15.00 | BKMG59 - 598 | 91.00 | 15.00 | BKMG60 - 599 | 90.00 | 15.00 | BKMG61 - 600 | 90.00 | 13.00 | BKMG62 - 601 | 90.00 | 11.00 | BKMG63 - 602 | 90.00 | 9.00 | BKMG64 - 603 | 90.00 | 7.00 | BKMG65 - 604 | 90.00 | 5.00 | BKMG66 - 605 | 90.00 | 3.00 | BKMG67 - 606 | 90.00 | 1.00 | BKMG68 - 607 | 90.00 | -1.00 | BKMG69 - 608 | 90.00 | -3.00 | BKMG70 - 609 | 90.00 | -5.00 | BKMG71 - 610 | 90.00 | -7.00 | BKMG72 - 611 | 90.00 | -9.00 | BKMG73 - 612 | 90.00 | -11.00 | BKMG74 - 613 | 90.00 | -13.00 | BKMG75 - 614 | 134.90 | 28.90 | 21004 - 615 | 126.30 | 28.10 | 22001 - 616 | 126.33 | 33.00 | 22107 - 617 | 57.70 | -20.45 | MAUR01 - 618 | 57.75 | -20.10 | MAUR02 - 619 | 55.00 | -25.00 | V14003 - 620 | 60.00 | -25.00 | V14004 - 621 | 57.00 | -20.00 | V14005 - 622 | 60.00 | -20.00 | V14006 - 623 | 63.00 | -20.00 | V14007 - 624 | 64.00 | -19.00 | V14008 - 625 | 58.00 | -18.00 | V14009 - 626 | 60.00 | -16.00 | V14010 - 627 | 56.00 | -10.00 | V14011 - 628 | 57.00 | -11.00 | V14012 - 629 | 70.00 | -25.00 | V14013 - 630 | 80.00 | -25.00 | V14014 - 631 | 90.00 | -25.00 | V53015 - 632 | 70.00 | -15.00 | V14016 - 633 | 80.00 | -15.00 | V14017 - 634 | 90.00 | -15.00 | V53018 - 635 | 65.00 | -5.00 | V23019 - 636 | 75.00 | -5.00 | V23020 - 637 | 85.00 | -5.00 | V23021 - 638 | 95.00 | -5.00 | V23022 - 639 | 55.00 | -5.00 | V14023 - 640 | 10.93 | -17.33 | V14039 - 641 | 10.95 | -18.10 | V14040 - 642 | 11.68 | -19.00 | V14041 - 643 | 12.25 | -20.07 | V14042 - 644 | 12.65 | -20.97 | V14043 - 645 | 13.22 | -21.93 | V14044 - 646 | 13.42 | -22.62 | V14045 - 647 | 13.58 | -23.61 | V14046 - 648 | 13.66 | -24.38 | V14047 - 649 | 13.95 | -25.25 | V14048 - 650 | 14.03 | -26.08 | V14049 - 651 | 14.19 | -26.73 | V14050 - 652 | 14.54 | -27.59 | V14051 - 653 | 14.97 | -28.34 | V14052 - 654 | 15.68 | -29.08 | V14053 - 655 | 39.75 | -5.00 | V14054 - 656 | 39.75 | -5.58 | V14055 - 657 | 39.75 | -6.22 | V14056 - 658 | 39.80 | -6.75 | V14057 - 659 | 39.92 | -7.08 | V14058 - 660 | 39.75 | -7.30 | V14059 - 661 | 39.75 | -8.17 | V14060 - 662 | 40.25 | -8.83 | V14061 - 663 | 40.25 | -9.17 | V14062 - 664 | 40.25 | -9.92 | V14063 - 665 | 40.58 | -10.33 | V14064 - 666 | 42.00 | -11.00 | V14065 - 667 | 42.00 | -12.00 | V14066 - 668 | 42.00 | -13.00 | V14067 - 669 | 42.00 | -14.00 | V14068 - 670 | 42.00 | -15.00 | V14069 - 671 | 40.25 | -17.00 | V14070 - 672 | 39.00 | -17.83 | V14071 - 673 | 38.25 | -18.00 | V14072 - 674 | 37.25 | -19.00 | V14073 - 675 | 35.50 | -20.00 | V14074 - 676 | 36.00 | -21.00 | V14075 - 677 | 35.57 | -21.92 | V14076 - 678 | 35.58 | -22.00 | V14077 - 679 | 36.00 | -23.00 | V14079 - 680 | 36.83 | -24.00 | V14080 - 681 | 36.00 | -25.00 | V14081 - 682 | 35.00 | -25.50 | V14082 - 683 | 34.00 | -26.00 | V14083 - 684 | 34.00 | -27.00 | V14084 - 685 | 40.00 | -8.00 | KNY51 - 686 | 42.00 | -8.00 | KNY52 - 687 | 44.00 | -8.00 | KNY53 - 688 | 46.00 | -8.00 | KNY54 - 689 | 48.00 | -8.00 | KNY55 - 690 | 50.00 | -8.00 | KNY56 - 691 | 52.00 | -8.00 | KNY57 - 692 | 54.00 | -8.00 | KNY58 - 693 | 56.00 | -8.00 | KNY59 - 694 | 58.00 | -8.00 | KNY60 - 695 | 60.00 | -8.00 | KNY61 - 696 | 62.00 | -8.00 | KNY62 - 697 | 64.00 | -8.00 | KNY63 - 698 | 66.00 | -8.00 | KNY64 - 699 | 68.00 | -8.00 | KNY65 - 700 | 70.00 | -8.00 | KNY66 - 701 | 70.00 | -6.00 | KNY67 - 702 | 70.00 | -4.00 | KNY68 - 703 | 70.00 | -2.00 | KNY69 - 704 | 70.00 | 0.00 | KNY70 - 705 | 70.00 | 2.00 | KNY71 - 706 | 70.00 | 4.00 | KNY72 - 707 | 70.00 | 6.00 | KNY73 - 708 | 70.00 | 8.00 | KNY74 - 709 | 68.00 | 8.00 | KNY75 - 710 | 66.00 | 8.00 | KNY76 - 711 | 64.00 | 8.00 | KNY77 - 712 | 62.00 | 8.00 | KNY78 - 713 | 60.00 | 8.00 | KNY79 - 714 | 58.00 | 8.00 | KNY80 - 715 | 56.00 | 8.00 | KNY81 - 716 | 54.00 | 8.00 | KNY82 - 717 | 52.00 | 8.00 | KNY83 - 718 | 42.00 | -18.00 | MDG51 - 719 | 42.00 | -20.00 | MDG52 - 720 | 42.00 | -22.00 | MDG53 - 721 | 42.00 | -24.00 | MDG54 - 722 | 42.00 | -26.00 | MDG55 - 723 | 42.00 | -28.00 | MDG56 - 724 | 42.00 | -30.00 | MDG57 - 725 | 44.00 | -30.00 | MDG58 - 726 | 46.00 | -30.00 | MDG59 - 727 | 48.00 | -30.00 | MDG60 - 728 | 50.00 | -30.00 | MDG61 - 729 | 52.00 | -30.00 | MDG62 - 730 | 54.00 | -30.00 | MDG63 - 731 | 54.00 | -28.00 | MDG64 - 732 | 54.00 | -26.00 | MDG65 - 733 | 54.00 | -24.00 | MDG66 - 734 | 54.00 | -22.00 | MDG67 - 735 | 54.00 | -20.00 | MDG68 - 736 | 54.00 | -18.00 | MDG69 - 737 | 54.00 | -16.00 | MDG70 - 738 | 54.00 | -14.00 | MDG71 - 739 | 54.00 | -12.00 | MDG72 - 740 | 54.00 | -10.00 | MDG73 - 741 | 52.00 | -10.00 | MDG74 - 742 | 50.00 | -10.00 | MDG75 - 743 | 48.00 | -10.00 | MDG76 - 744 | 46.00 | -10.00 | MDG77 - 745 | 44.00 | -10.00 | MDG78 - 746 | 44.00 | -12.00 | MDG79 - 747 | 44.00 | -14.00 | MDG80 - 748 | 42.00 | -16.00 | MDG81 - 749 | 50.00 | -10.00 | SYC51 - 750 | 52.00 | -10.00 | SYC52 - 751 | 54.00 | -10.00 | SYC53 - 752 | 56.00 | -10.00 | SYC54 - 753 | 58.00 | -10.00 | SYC55 - 754 | 60.00 | -10.00 | SYC56 - 755 | 60.00 | -8.00 | SYC57 - 756 | 60.00 | -6.00 | SYC58 - 757 | 60.00 | -4.00 | SYC59 - 758 | 60.00 | -2.00 | SYC60 - 759 | 60.00 | 0.00 | SYC61 - 760 | 58.00 | 0.00 | SYC62 - 761 | 56.00 | 0.00 | SYC63 - 762 | 54.00 | 0.00 | SYC64 - 763 | 52.00 | 0.00 | SYC65 - 764 | 50.00 | 0.00 | SYC66 - 765 | 50.00 | -2.00 | SYC67 - 766 | 50.00 | -4.00 | SYC68 - 767 | 50.00 | -6.00 | SYC69 - 768 | 50.00 | -8.00 | SYC70 - 769 | 50.00 | -10.00 | SYC71 - 770 | -75.76 | 32.52 | ATX1 - 771 | -75.57 | 32.67 | ATX2 - 772 | -80.23 | 28.85 | CAPN - 773 | -80.25 | 30.92 | JAX - 774 | -83.75 | 28.50 | TPA - 775 | -84.20 | 29.25 | TLH - 776 | -87.50 | 29.80 | PEN - 777 | -64.97 | 42.05 | ECALE - 778 | -54.25 | 45.88 | ECALL - 779 | -9.25 | 51.23 | SHAN - 780 | -53.01 | 46.35 | STJO - 781 | -75.05 | 33.50 | CMFS - 782 | -143.64 | -21.94 | TAHI - 783 | -157.25 | 20.75 | HNL51 - 784 | -157.81 | 20.16 | HNL52 - 785 | -158.54 | 20.57 | HNL53 - 786 | -159.27 | 20.98 | HNL54 - 787 | -158.74 | 21.79 | HNL55 - 788 | -158.17 | 22.67 | HNL56 - 789 | -157.42 | 22.26 | HNL57 - 790 | -156.67 | 21.87 | HNL58 - 791 | -157.00 | 21.50 | HNL59 - 792 | -161.12 | 21.34 | HNL61 - 793 | -160.71 | 22.49 | HNL62 - 794 | -160.29 | 23.65 | HNL63 - 795 | -159.27 | 23.20 | HNL64 - 796 | -158.25 | 22.75 | HNL65 - 797 | -158.76 | 21.80 | HNL66 - 798 | -159.24 | 20.82 | HNL67 - 799 | -160.18 | 21.08 | HNL68 - 800 | -80.50 | 30.67 | CHS52 - 801 | -79.80 | 30.67 | CHS53 - 802 | -79.00 | 30.67 | CHS54 - 803 | -78.10 | 30.67 | CHS55 - 804 | -78.10 | 31.40 | CHS56 - 805 | -78.10 | 32.13 | CHS57 - 806 | -78.10 | 32.86 | CHS58 - 807 | -78.10 | 33.58 | CHS59 - 808 | -79.50 | 28.50 | JAX52 - 809 | -78.50 | 28.50 | JAX53 - 810 | -78.50 | 29.75 | JAX54 - 811 | -78.50 | 31.25 | JAX55 - 812 | -78.50 | 32.50 | JAX56 - 813 | -79.25 | 32.50 | JAX57 - 814 | -80.20 | 28.70 | JXFL51 - 815 | -79.70 | 28.70 | JXFL52 - 816 | -79.30 | 28.70 | JXFL53 - 817 | -79.30 | 29.20 | JXFL54 - 818 | -79.30 | 29.70 | JXFL55 - 819 | -79.30 | 30.20 | JXFL56 - 820 | -79.30 | 30.70 | JXFL57 - 821 | -79.30 | 31.20 | JXFL58 - 822 | -79.30 | 31.60 | JXFL59 - 823 | -79.30 | 32.00 | JXFL60 - 824 | -79.80 | 32.00 | JXFL61 - 825 | -80.20 | 32.00 | JXFL62 - 826 | -80.50 | 32.00 | JXFL63 - 827 | -72.40 | 40.60 | MNE51 - 828 | -72.40 | 40.20 | MNE52 - 829 | -71.60 | 40.20 | MNE53 - 830 | -70.80 | 40.20 | MNE54 - 831 | -70.00 | 40.20 | MNE55 - 832 | -69.20 | 40.20 | MNE56 - 833 | -68.40 | 40.20 | MNE57 - 834 | -67.60 | 40.20 | MNE58 - 835 | -66.80 | 40.20 | MNE59 - 836 | -66.80 | 41.00 | MNE60 - 837 | -66.80 | 41.80 | MNE61 - 838 | -66.80 | 42.60 | MNE62 - 839 | -66.80 | 43.40 | MNE63 - 840 | -66.80 | 44.20 | MNE64 - 841 | -68.00 | 17.00 | SJU51 - 842 | -68.00 | 17.50 | SJU52 - 843 | -68.00 | 18.00 | SJU53 - 844 | -68.00 | 18.50 | SJU54 - 845 | -68.00 | 19.00 | SJU55 - 846 | -68.00 | 19.50 | SJU56 - 847 | -67.50 | 19.50 | SJU57 - 848 | -67.00 | 19.50 | SJU58 - 849 | -66.50 | 19.50 | SJU59 - 850 | -66.00 | 19.50 | SJU60 - 851 | -65.50 | 19.50 | SJU61 - 852 | -65.00 | 19.50 | SJU62 - 853 | -64.50 | 19.50 | SJU63 - 854 | -64.00 | 19.50 | SJU64 - 855 | -64.00 | 19.00 | SJU65 - 856 | -64.00 | 18.50 | SJU66 - 857 | -64.00 | 18.00 | SJU67 - 858 | -64.00 | 17.50 | SJU68 - 859 | -64.00 | 17.00 | SJU69 - 860 | -64.50 | 17.00 | SJU70 - 861 | -65.00 | 17.00 | SJU71 - 862 | -65.50 | 17.00 | SJU72 - 863 | -66.00 | 17.00 | SJU73 - 864 | -66.50 | 17.00 | SJU74 - 865 | -67.00 | 17.00 | SJU75 - 866 | -67.50 | 17.00 | SJU76 - 867 | -68.00 | 17.00 | SJU77 - 868 | -65.50 | 31.50 | BER51 - 869 | -65.25 | 31.50 | BER52 - 870 | -65.00 | 31.50 | BER53 - 871 | -64.75 | 31.50 | BER54 - 872 | -64.50 | 31.50 | BER55 - 873 | -64.25 | 31.50 | BER56 - 874 | -64.00 | 31.50 | BER57 - 875 | -64.00 | 31.75 | BER58 - 876 | -64.00 | 32.00 | BER59 - 877 | -64.00 | 32.25 | BER60 - 878 | -64.00 | 32.50 | BER61 - 879 | -64.00 | 32.75 | BER62 - 880 | -64.00 | 33.00 | BER63 - 881 | -64.25 | 33.00 | BER64 - 882 | -64.50 | 33.00 | BER65 - 883 | -64.75 | 33.00 | BER66 - 884 | -64.00 | 33.00 | BER67 - 885 | -65.25 | 33.00 | BER68 - 886 | -65.50 | 33.00 | BER69 - 887 | -65.50 | 32.75 | BER70 - 888 | -65.50 | 32.50 | BER71 - 889 | -65.50 | 32.25 | BER72 - 890 | -65.50 | 32.00 | BER73 - 891 | -65.50 | 31.75 | BER74 - 892 | -97.00 | 23.00 | SRH51 - 893 | -96.00 | 23.00 | SRH52 - 894 | -95.00 | 23.00 | SRH53 - 895 | -94.00 | 23.00 | SRH54 - 896 | -93.00 | 23.00 | SRH55 - 897 | -92.00 | 23.00 | SRH56 - 898 | -91.00 | 23.00 | SRH57 - 899 | -90.00 | 23.00 | SRH58 - 900 | -89.00 | 23.00 | SRH59 - 901 | -88.00 | 23.00 | SRH60 - 902 | -87.00 | 23.00 | SRH61 - 903 | -86.00 | 23.00 | SRH62 - 904 | -85.00 | 23.00 | SRH63 - 905 | -79.00 | 23.00 | SRH64 - 906 | -78.00 | 23.00 | SRH65 - 907 | -78.00 | 28.00 | SRH66 - 908 | -78.00 | 29.00 | SRH67 - 909 | -78.00 | 30.00 | SRH68 - 910 | -78.00 | 31.00 | SRH69 - 911 | -78.00 | 32.00 | SRH70 - 912 | -78.00 | 33.00 | SRH71 - 913 | -96.41 | 25.26 | BRO52 - 914 | -95.91 | 25.26 | BRO53 - 915 | -95.41 | 25.26 | BRO54 - 916 | -95.41 | 25.76 | BRO55 - 917 | -95.41 | 26.26 | BRO56 - 918 | -95.41 | 26.76 | BRO57 - 919 | -95.41 | 27.26 | BRO58 - 920 | -95.41 | 27.76 | BRO59 - 921 | -95.41 | 28.26 | BRO60 - 922 | -96.87 | 27.00 | HGX51 - 923 | -96.37 | 27.00 | HGX52 - 924 | -95.87 | 27.00 | HGX53 - 925 | -95.37 | 27.00 | HGX54 - 926 | -94.87 | 27.00 | HGX55 - 927 | -94.37 | 27.00 | HGX56 - 928 | -93.87 | 27.00 | HGX57 - 929 | -93.37 | 27.00 | HGX58 - 930 | -93.37 | 27.50 | HGX59 - 931 | -93.37 | 28.00 | HGX60 - 932 | -93.37 | 28.50 | HGX61 - 933 | -93.37 | 29.00 | HGX62 - 934 | -96.73 | 25.75 | CRP52 - 935 | -96.48 | 25.75 | CRP53 - 936 | -96.16 | 25.76 | CRP54 - 937 | -95.77 | 25.73 | CRP55 - 938 | -95.39 | 25.73 | CRP56 - 939 | -95.15 | 25.75 | CRP57 - 940 | -95.14 | 25.98 | CRP58 - 941 | -95.13 | 26.35 | CRP59 - 942 | -95.13 | 26.68 | CRP60 - 943 | -95.11 | 27.00 | CRP61 - 944 | -95.10 | 27.35 | CRP62 - 945 | -95.09 | 27.76 | CRP63 - 946 | -95.09 | 28.10 | CRP64 - 947 | -95.08 | 28.36 | CRP65 - 948 | -95.07 | 28.64 | CRP66 - 949 | -95.06 | 28.86 | CRP67 - 950 | -96.50 | 26.00 | CCTX52 - 951 | -96.00 | 26.00 | CCTX53 - 952 | -95.50 | 26.00 | CCTX54 - 953 | -95.00 | 26.00 | CCTX55 - 954 | -95.00 | 26.50 | CCTX56 - 955 | -95.00 | 27.00 | CCTX57 - 956 | -95.00 | 27.50 | CCTX58 - 957 | -95.00 | 28.00 | CCTX59 - 958 | -95.00 | 28.50 | CCTX60 - 959 | -95.00 | 28.80 | CCTX61 - 960 | -95.03 | 28.55 | LCH51 - 961 | -95.03 | 28.23 | LCH52 - 962 | -95.03 | 27.79 | LCH53 - 963 | -95.03 | 27.41 | LCH54 - 964 | -94.45 | 27.41 | LCH55 - 965 | -93.53 | 27.41 | LCH56 - 966 | -92.66 | 27.41 | LCH57 - 967 | -91.69 | 27.41 | LCH58 - 968 | -90.43 | 27.41 | LCH59 - 969 | -90.43 | 27.74 | LCH60 - 970 | -90.43 | 28.16 | LCH61 - 971 | -90.43 | 28.52 | LCH62 - 972 | -90.43 | 28.68 | LCH63 - 973 | -91.80 | 28.90 | LIX51 - 974 | -91.80 | 28.50 | LIX52 - 975 | -91.80 | 28.00 | LIX53 - 976 | -91.80 | 27.50 | LIX54 - 977 | -91.30 | 27.50 | LIX55 - 978 | -90.80 | 27.50 | LIX56 - 979 | -90.30 | 27.50 | LIX57 - 980 | -89.80 | 27.50 | LIX58 - 981 | -89.30 | 27.50 | LIX59 - 982 | -88.80 | 27.50 | LIX60 - 983 | -88.30 | 27.50 | LIX61 - 984 | -87.80 | 27.50 | LIX62 - 985 | -87.40 | 27.50 | LIX63 - 986 | -87.40 | 28.00 | LIX64 - 987 | -87.40 | 28.50 | LIX65 - 988 | -87.40 | 29.00 | LIX66 - 989 | -87.40 | 29.50 | LIX67 - 990 | -87.40 | 30.00 | LIX68 - 991 | -89.10 | 28.80 | MOB51 - 992 | -89.10 | 28.50 | MOB52 - 993 | -88.40 | 28.50 | MOB53 - 994 | -87.90 | 28.50 | MOB54 - 995 | -87.40 | 28.50 | MOB55 - 996 | -86.90 | 28.50 | MOB56 - 997 | -86.40 | 28.50 | MOB57 - 998 | -85.90 | 28.50 | MOB58 - 999 | -85.90 | 29.00 | MOB59 - 1000 | -85.90 | 29.50 | MOB60 - 1001 | -87.35 | 28.35 | PCB51 - 1002 | -87.35 | 28.65 | PCB52 - 1003 | -87.35 | 28.95 | PCB53 - 1004 | -87.35 | 29.25 | PCB54 - 1005 | -87.35 | 29.55 | PCB55 - 1006 | -87.35 | 29.85 | PCB56 - 1007 | -87.05 | 28.35 | PCB58 - 1008 | -86.75 | 28.35 | PCB59 - 1009 | -86.45 | 28.35 | PCB60 - 1010 | -86.15 | 28.35 | PCB61 - 1011 | -85.85 | 28.35 | PCB62 - 1012 | -85.55 | 28.35 | PCB63 - 1013 | -85.25 | 28.35 | PCB64 - 1014 | -84.95 | 28.35 | PCB65 - 1015 | -84.65 | 28.35 | PCB66 - 1016 | -84.35 | 28.35 | PCB67 - 1017 | -84.05 | 28.35 | PCB68 - 1018 | -83.75 | 28.35 | PCB69 - 1019 | -83.45 | 28.35 | PCB70 - 1020 | -84.80 | 29.45 | TBW51 - 1021 | -84.80 | 28.95 | TBW52 - 1022 | -84.80 | 28.45 | TBW53 - 1023 | -84.80 | 27.95 | TBW54 - 1024 | -84.80 | 27.45 | TBW55 - 1025 | -84.80 | 26.95 | TBW56 - 1026 | -84.80 | 26.45 | TBW57 - 1027 | -84.80 | 25.95 | TBW58 - 1028 | -84.80 | 25.45 | TBW59 - 1029 | -84.30 | 25.45 | TBW60 - 1030 | -83.80 | 25.45 | TBW61 - 1031 | -83.30 | 25.45 | TBW62 - 1032 | -82.80 | 25.45 | TBW63 - 1033 | -82.30 | 25.45 | TBW64 - 1034 | -83.54 | 24.10 | MIA51 - 1035 | -83.54 | 24.82 | MIA52 - 1036 | -83.54 | 25.54 | MIA53 - 1037 | -83.54 | 26.26 | MIA54 - 1038 | -83.54 | 26.98 | MIA55 - 1039 | -83.54 | 27.70 | MIA56 - 1040 | -83.14 | 27.70 | MIA57 - 1041 | -80.01 | 27.70 | MIA58 - 1042 | -79.61 | 27.70 | MIA59 - 1043 | -79.21 | 27.70 | MIA60 - 1044 | -78.81 | 27.70 | MIA61 - 1045 | -78.41 | 27.70 | MIA62 - 1046 | -78.41 | 27.40 | MIA63 - 1047 | -78.41 | 26.50 | MIA64 - 1048 | -78.41 | 26.20 | MIA65 - 1049 | -78.41 | 25.90 | MIA66 - 1050 | -78.41 | 24.10 | MIA67 - 1051 | -78.92 | 24.10 | MIA68 - 1052 | -79.44 | 24.10 | MIA69 - 1053 | -79.95 | 24.10 | MIA70 - 1054 | -80.46 | 24.10 | MIA71 - 1055 | -80.97 | 24.10 | MIA72 - 1056 | -81.49 | 24.10 | MIA73 - 1057 | -82.00 | 24.10 | MIA74 - 1058 | -82.51 | 24.10 | MIA75 - 1059 | -83.03 | 24.10 | MIA76 - 1060 | -81.00 | 30.00 | MLB51 - 1061 | -80.60 | 30.00 | MLB52 - 1062 | -80.20 | 30.00 | MLB53 - 1063 | -79.80 | 30.00 | MLB54 - 1064 | -79.40 | 30.00 | MLB55 - 1065 | -79.00 | 30.00 | MLB56 - 1066 | -78.51 | 30.00 | MLB57 - 1067 | -78.51 | 29.70 | MLB58 - 1068 | -78.51 | 29.40 | MLB59 - 1069 | -78.51 | 29.10 | MLB60 - 1070 | -78.51 | 28.80 | MLB61 - 1071 | -78.51 | 28.50 | MLB62 - 1072 | -78.51 | 28.20 | MLB63 - 1073 | -78.51 | 27.90 | MLB64 - 1074 | -78.51 | 27.60 | MLB65 - 1075 | -78.51 | 27.30 | MLB66 - 1076 | -79.90 | 26.50 | MLB67 - 1077 | -79.70 | 26.50 | MLB68 - 1078 | -79.50 | 26.50 | MLB69 - 1079 | -79.30 | 26.50 | MLB70 - 1080 | -79.10 | 26.50 | MLB71 - 1081 | -78.90 | 26.50 | MLB72 - 1082 | -83.50 | 26.00 | KEY51 - 1083 | -83.50 | 25.75 | KEY52 - 1084 | -83.50 | 25.50 | KEY53 - 1085 | -83.50 | 25.25 | KEY54 - 1086 | -83.50 | 25.00 | KEY55 - 1087 | -83.50 | 24.75 | KEY56 - 1088 | -83.50 | 24.50 | KEY57 - 1089 | -83.50 | 24.25 | KEY58 - 1090 | -83.50 | 24.00 | KEY59 - 1091 | -83.50 | 23.75 | KEY60 - 1092 | -83.50 | 23.50 | KEY61 - 1093 | -83.50 | 23.25 | KEY62 - 1094 | -83.50 | 23.00 | KEY63 - 1095 | -79.20 | 23.00 | KEY67 - 1096 | -79.00 | 23.00 | KEY68 - 1097 | -79.00 | 23.20 | KEY69 - 1098 | -79.00 | 23.40 | KEY70 - 1099 | -79.00 | 26.00 | KEY71 - 1100 | -79.20 | 26.00 | KEY72 - 1101 | -79.40 | 26.00 | KEY73 - 1102 | -79.60 | 26.00 | KEY74 - 1103 | -79.80 | 26.00 | KEY75 - 1104 | -82.45 | 26.00 | KEY78 - 1105 | -82.70 | 26.00 | KEY79 - 1106 | -82.95 | 26.00 | KEY80 - 1107 | -83.20 | 26.00 | KEY81 - 1108 | -69.20 | 43.30 | NW-CAR51 - 1109 | -68.70 | 43.30 | NW-CAR52 - 1110 | -68.20 | 43.30 | NW-CAR53 - 1111 | -67.70 | 43.30 | NW-CAR54 - 1112 | -67.20 | 43.30 | NW-CAR55 - 1113 | -66.70 | 43.30 | NW-CAR56 - 1114 | -66.30 | 43.30 | NW-CAR57 - 1115 | -66.30 | 43.55 | NW-CAR58 - 1116 | -66.30 | 43.80 | NW-CAR59 - 1117 | -66.30 | 44.05 | NW-CAR60 - 1118 | -66.30 | 44.30 | NW-CAR61 - 1119 | -66.30 | 44.55 | NW-CAR62 - 1120 | -66.30 | 44.80 | NW-CAR63 - 1121 | -66.30 | 45.05 | NW-CAR64 - 1122 | -69.20 | 43.80 | NW-CAR65 - 1123 | -69.20 | 43.55 | NW-CAR66 - 1124 | -69.20 | 43.30 | NW-CAR67 - 1125 | -72.50 | 40.45 | NW-BOX51 - 1126 | -72.00 | 40.45 | NW-BOX52 - 1127 | -71.50 | 40.45 | NW-BOX53 - 1128 | -71.00 | 40.45 | NW-BOX54 - 1129 | -70.50 | 40.45 | NW-BOX55 - 1130 | -70.00 | 40.45 | NW-BOX56 - 1131 | -69.50 | 40.45 | NW-BOX57 - 1132 | -69.00 | 40.45 | NW-BOX58 - 1133 | -68.85 | 40.45 | NW-BOX59 - 1134 | -68.85 | 40.95 | NW-BOX60 - 1135 | -68.85 | 41.45 | NW-BOX61 - 1136 | -68.85 | 41.95 | NW-BOX62 - 1137 | -68.85 | 42.45 | NW-BOX63 - 1138 | -68.85 | 42.95 | NW-BOX64 - 1139 | -68.85 | 43.25 | NW-BOX65 - 1140 | -69.35 | 43.25 | NW-BOX66 - 1141 | -69.85 | 43.25 | NW-BOX67 - 1142 | -72.50 | 40.70 | NW-BOX70 - 1143 | -72.50 | 41.20 | NW-BOX71 - 1144 | -71.55 | 39.75 | NW-NYC51 - 1145 | -72.05 | 39.75 | NW-NYC52 - 1146 | -72.55 | 39.75 | NW-NYC53 - 1147 | -73.05 | 39.75 | NW-NYC54 - 1148 | -73.55 | 39.75 | NW-NYC55 - 1149 | -74.05 | 39.75 | NW-NYC56 - 1150 | -71.05 | 39.75 | NW-NYC57 - 1151 | -71.05 | 40.25 | NW-NYC58 - 1152 | -71.05 | 40.75 | NW-NYC59 - 1153 | -71.05 | 41.25 | NW-NYC60 - 1154 | -71.05 | 41.38 | NW-NYC61 - 1155 | -75.10 | 38.20 | NW-PHI51 - 1156 | -74.75 | 38.20 | NW-PHI52 - 1157 | -74.25 | 38.20 | NW-PHI53 - 1158 | -73.75 | 38.20 | NW-PHI54 - 1159 | -73.35 | 38.20 | NW-PHI55 - 1160 | -73.35 | 38.60 | NW-PHI56 - 1161 | -73.35 | 39.10 | NW-PHI57 - 1162 | -73.35 | 39.60 | NW-PHI58 - 1163 | -73.35 | 40.10 | NW-PHI59 - 1164 | -75.40 | 36.75 | NW-LWX51 - 1165 | -75.40 | 37.00 | NW-LWX54 - 1166 | -75.40 | 37.25 | NW-LWX55 - 1167 | -75.40 | 37.50 | NW-LWX56 - 1168 | -75.40 | 37.75 | NW-LWX57 - 1169 | -79.40 | 32.50 | NW-ILM53 - 1170 | -78.90 | 32.50 | NW-ILM54 - 1171 | -78.40 | 32.50 | NW-ILM55 - 1172 | -77.90 | 32.50 | NW-ILM56 - 1173 | -77.40 | 32.50 | NW-ILM57 - 1174 | -76.90 | 32.50 | NW-ILM58 - 1175 | -76.50 | 32.50 | NW-ILM59 - 1176 | -76.50 | 33.00 | NW-ILM60 - 1177 | -76.50 | 33.50 | NW-ILM61 - 1178 | -76.50 | 34.00 | NW-ILM62 - 1179 | -76.50 | 34.50 | NW-ILM63 - 1180 | -70.80 | 42.40 | NW-GYX51 - 1181 | -70.36 | 42.40 | NW-GYX52 - 1182 | -69.86 | 42.40 | NW-GYX53 - 1183 | -69.36 | 42.40 | NW-GYX54 - 1184 | -68.86 | 42.40 | NW-GYX55 - 1185 | -68.36 | 42.40 | NW-GYX56 - 1186 | -67.86 | 42.40 | NW-GYX57 - 1187 | -67.85 | 42.40 | NW-GYX58 - 1188 | -67.85 | 42.90 | NW-GYX59 - 1189 | -67.85 | 43.40 | NW-GYX60 - 1190 | -67.85 | 43.90 | NW-GYX61 - 1191 | -77.91 | 33.85 | NW-MHX51 - 1192 | -77.41 | 33.85 | NW-MHX52 - 1193 | -76.91 | 33.85 | NW-MHX53 - 1194 | -76.41 | 33.85 | NW-MHX54 - 1195 | -75.91 | 33.85 | NW-MHX55 - 1196 | -75.41 | 33.85 | NW-MHX56 - 1197 | -74.91 | 33.85 | NW-MHX57 - 1198 | -74.75 | 33.85 | NW-MHX58 - 1199 | -74.75 | 34.35 | NW-MHX59 - 1200 | -74.75 | 34.85 | NW-MHX60 - 1201 | -74.75 | 35.35 | NW-MHX61 - 1202 | -74.75 | 35.85 | NW-MHX62 - 1203 | -74.75 | 36.35 | NW-MHX63 - 1204 | -74.75 | 36.60 | NW-MHX64 - 1205 | -75.25 | 36.60 | NW-MHX65 - 1206 | -74.25 | 39.50 | NW-AKQ51 - 1207 | -74.25 | 39.00 | NW-AKQ52 - 1208 | -74.25 | 38.50 | NW-AKQ53 - 1209 | -74.25 | 38.00 | NW-AKQ54 - 1210 | -74.25 | 37.50 | NW-AKQ55 - 1211 | -74.25 | 37.00 | NW-AKQ56 - 1212 | -74.25 | 36.50 | NW-AKQ57 - 1213 | -74.25 | 36.00 | NW-AKQ58 - 1214 | -74.25 | 35.80 | NW-AKQ59 - 1215 | -74.50 | 35.80 | NW-AKQ60 - 1216 | -75.00 | 35.80 | NW-AKQ61 - 1217 | -75.50 | 35.80 | NW-AKQ62 - 1218 | -119.00 | 32.08 | NW-SGX51 - 1219 | -119.00 | 32.33 | NW-SGX52 - 1220 | -119.00 | 32.58 | NW-SGX53 - 1221 | -119.00 | 32.83 | NW-SGX54 - 1222 | -119.00 | 33.08 | NW-SGX55 - 1223 | -119.00 | 33.33 | NW-SGX56 - 1224 | -118.00 | 32.08 | NW-SGX66 - 1225 | -118.25 | 32.08 | NW-SGX67 - 1226 | -118.50 | 32.08 | NW-SGX68 - 1227 | -118.75 | 32.08 | NW-SGX69 - 1228 | -123.36 | 32.60 | NW-LOX51 - 1229 | -123.36 | 33.10 | NW-LOX52 - 1230 | -123.36 | 33.60 | NW-LOX53 - 1231 | -123.36 | 34.10 | NW-LOX54 - 1232 | -123.36 | 34.60 | NW-LOX55 - 1233 | -123.36 | 35.10 | NW-LOX56 - 1234 | -123.36 | 35.60 | NW-LOX57 - 1235 | -123.36 | 36.10 | NW-LOX58 - 1236 | -123.36 | 36.10 | NW-LOX59 - 1237 | -122.86 | 36.10 | NW-LOX60 - 1238 | -122.36 | 36.10 | NW-LOX61 - 1239 | -123.36 | 32.60 | NW-LOX63 - 1240 | -122.86 | 32.60 | NW-LOX64 - 1241 | -122.36 | 32.60 | NW-LOX65 - 1242 | -121.86 | 32.60 | NW-LOX66 - 1243 | -121.36 | 32.60 | NW-LOX67 - 1244 | -120.86 | 32.60 | NW-LOX68 - 1245 | -120.36 | 32.60 | NW-LOX69 - 1246 | -119.86 | 32.60 | NW-LOX70 - 1247 | -119.36 | 32.60 | NW-LOX71 - 1248 | -118.86 | 32.60 | NW-LOX72 - 1249 | -118.36 | 32.60 | NW-LOX73 - 1250 | -125.80 | 39.40 | NW-MTR51 - 1251 | -125.80 | 38.90 | NW-MTR52 - 1252 | -125.80 | 38.40 | NW-MTR53 - 1253 | -125.80 | 37.90 | NW-MTR54 - 1254 | -125.80 | 37.40 | NW-MTR55 - 1255 | -125.80 | 36.90 | NW-MTR56 - 1256 | -125.80 | 36.40 | NW-MTR57 - 1257 | -125.80 | 35.90 | NW-MTR58 - 1258 | -125.80 | 35.40 | NW-MTR59 - 1259 | -125.80 | 35.00 | NW-MTR60 - 1260 | -125.30 | 35.00 | NW-MTR61 - 1261 | -124.80 | 35.00 | NW-MTR62 - 1262 | -124.30 | 35.00 | NW-MTR63 - 1263 | -123.80 | 35.00 | NW-MTR64 - 1264 | -123.30 | 35.00 | NW-MTR65 - 1265 | -122.80 | 35.00 | NW-MTR66 - 1266 | -122.30 | 35.00 | NW-MTR67 - 1267 | -121.80 | 35.00 | NW-MTR68 - 1268 | -125.80 | 39.40 | NW-MTR71 - 1269 | -125.30 | 39.40 | NW-MTR72 - 1270 | -124.80 | 39.40 | NW-MTR73 - 1271 | -124.30 | 39.40 | NW-MTR74 - 1272 | -127.30 | 41.00 | NW-MFR51 - 1273 | -127.30 | 41.50 | NW-MFR52 - 1274 | -127.30 | 42.00 | NW-MFR53 - 1275 | -127.30 | 42.50 | NW-MFR54 - 1276 | -127.30 | 43.00 | NW-MFR55 - 1277 | -127.30 | 43.50 | NW-MFR56 - 1278 | -127.30 | 44.00 | NW-MFR57 - 1279 | -127.30 | 44.45 | NW-MFR58 - 1280 | -126.80 | 44.45 | NW-MFR59 - 1281 | -126.30 | 44.45 | NW-MFR60 - 1282 | -125.80 | 44.45 | NW-MFR61 - 1283 | -125.30 | 44.45 | NW-MFR62 - 1284 | -124.80 | 44.45 | NW-MFR63 - 1285 | -124.80 | 41.00 | NW-MFR66 - 1286 | -125.30 | 41.00 | NW-MFR67 - 1287 | -125.80 | 41.00 | NW-MFR68 - 1288 | -126.30 | 41.00 | NW-MFR69 - 1289 | -126.80 | 41.00 | NW-MFR70 - 1290 | -127.30 | 41.00 | NW-MFR71 - 1291 | -126.28 | 43.50 | NW-PQR51 - 1292 | -126.28 | 44.00 | NW-PQR52 - 1293 | -126.28 | 44.50 | NW-PQR53 - 1294 | -126.28 | 45.00 | NW-PQR54 - 1295 | -126.28 | 45.50 | NW-PQR55 - 1296 | -126.28 | 46.00 | NW-PQR56 - 1297 | -126.28 | 46.50 | NW-PQR57 - 1298 | -126.28 | 47.00 | NW-PQR58 - 1299 | -126.28 | 47.15 | NW-PQR59 - 1300 | -125.78 | 47.15 | NW-PQR60 - 1301 | -125.28 | 47.15 | NW-PQR61 - 1302 | -125.30 | 43.50 | NW-PQR66 - 1303 | -125.80 | 43.50 | NW-PQR67 - 1304 | -127.00 | 46.10 | NW-SEW51 - 1305 | -127.00 | 46.60 | NW-SEW52 - 1306 | -127.00 | 47.10 | NW-SEW53 - 1307 | -127.00 | 47.60 | NW-SEW54 - 1308 | -127.00 | 48.10 | NW-SEW55 - 1309 | -127.00 | 48.60 | NW-SEW56 - 1310 | -127.00 | 49.10 | NW-SEW57 - 1311 | -127.00 | 49.42 | NW-SEW58 - 1312 | -124.91 | 46.10 | NW-SEW62 - 1313 | -125.41 | 46.10 | NW-SEW63 - 1314 | -125.91 | 46.10 | NW-SEW64 - 1315 | -126.41 | 46.10 | NW-SEW65 - 1316 | -126.91 | 46.10 | NW-SEW66 - 1317 | -126.27 | 38.40 | NW-EKA51 - 1318 | -126.27 | 38.90 | NW-EKA52 - 1319 | -126.27 | 39.40 | NW-EKA53 - 1320 | -126.27 | 39.90 | NW-EKA54 - 1321 | -126.27 | 40.40 | NW-EKA55 - 1322 | -126.27 | 40.90 | NW-EKA56 - 1323 | -126.27 | 41.40 | NW-EKA57 - 1324 | -126.27 | 41.90 | NW-EKA58 - 1325 | -126.27 | 42.20 | NW-EKA59 - 1326 | -125.77 | 42.20 | NW-EKA60 - 1327 | -125.27 | 42.20 | NW-EKA61 - 1328 | -124.70 | 38.40 | NW-EKA66 - 1329 | -125.20 | 38.40 | NW-EKA67 - 1330 | -125.70 | 38.40 | NW-EKA68 - 1331 | -126.20 | 38.40 | NW-EKA69 - 1332 | -126.27 | 38.40 | NW-EKA70 - 1333 | -161.50 | 18.07 | NW-HFO51 - 1334 | -161.50 | 18.57 | NW-HFO52 - 1335 | -161.50 | 19.07 | NW-HFO53 - 1336 | -161.50 | 19.57 | NW-HFO54 - 1337 | -161.50 | 20.07 | NW-HFO55 - 1338 | -161.50 | 20.57 | NW-HFO56 - 1339 | -161.50 | 21.07 | NW-HFO57 - 1340 | -161.50 | 21.57 | NW-HFO58 - 1341 | -161.50 | 22.07 | NW-HFO59 - 1342 | -161.50 | 22.57 | NW-HFO60 - 1343 | -161.50 | 23.09 | NW-HFO61 - 1344 | -161.50 | 23.09 | NW-HFO62 - 1345 | -161.00 | 23.09 | NW-HFO63 - 1346 | -160.50 | 23.09 | NW-HFO64 - 1347 | -160.00 | 23.09 | NW-HFO65 - 1348 | -159.50 | 23.09 | NW-HFO66 - 1349 | -159.00 | 23.09 | NW-HFO67 - 1350 | -158.50 | 23.09 | NW-HFO68 - 1351 | -158.00 | 23.09 | NW-HFO69 - 1352 | -157.50 | 23.09 | NW-HFO70 - 1353 | -157.00 | 23.09 | NW-HFO71 - 1354 | -156.50 | 23.09 | NW-HFO72 - 1355 | -156.00 | 23.09 | NW-HFO73 - 1356 | -155.50 | 23.09 | NW-HFO74 - 1357 | -155.00 | 23.09 | NW-HFO75 - 1358 | -154.50 | 23.09 | NW-HFO76 - 1359 | -153.90 | 23.09 | NW-HFO77 - 1360 | -153.90 | 23.00 | NW-HFO78 - 1361 | -153.90 | 22.50 | NW-HFO79 - 1362 | -153.90 | 22.00 | NW-HFO80 - 1363 | -153.90 | 21.50 | NW-HFO81 - 1364 | -153.90 | 21.00 | NW-HFO82 - 1365 | -153.90 | 20.50 | NW-HFO83 - 1366 | -153.90 | 20.00 | NW-HFO84 - 1367 | -153.90 | 19.50 | NW-HFO85 - 1368 | -153.90 | 19.00 | NW-HFO86 - 1369 | -153.90 | 18.50 | NW-HFO87 - 1370 | -153.90 | 18.07 | NW-HFO88 - 1371 | -154.00 | 18.07 | NW-HFO89 - 1372 | -154.50 | 18.07 | NW-HFO90 - 1373 | -155.00 | 18.07 | NW-HFO91 - 1374 | -155.50 | 18.07 | NW-HFO92 - 1375 | -156.00 | 18.07 | NW-HFO93 - 1376 | -156.50 | 18.07 | NW-HFO94 - 1377 | -157.00 | 18.07 | NW-HFO95 - 1378 | -157.50 | 18.07 | NW-HFO96 - 1379 | -158.00 | 18.07 | NW-HFO97 - 1380 | -158.50 | 18.07 | NW-HFO98 - 1381 | -159.00 | 18.07 | NW-HFO99 - 1382 | -159.50 | 18.07 | NW-HFO100 - 1383 | -160.00 | 18.07 | NW-HFO101 - 1384 | -160.50 | 18.07 | NW-HFO102 - 1385 | -161.00 | 18.07 | NW-HFO103 - 1386 | -144.50 | 53.90 | NW-AJK51 - 1387 | -144.50 | 54.00 | NW-AJK52 - 1388 | -144.50 | 54.50 | NW-AJK53 - 1389 | -144.50 | 55.00 | NW-AJK54 - 1390 | -144.50 | 55.50 | NW-AJK55 - 1391 | -144.50 | 56.00 | NW-AJK56 - 1392 | -144.50 | 56.50 | NW-AJK57 - 1393 | -144.50 | 57.00 | NW-AJK58 - 1394 | -144.50 | 57.50 | NW-AJK59 - 1395 | -144.50 | 58.00 | NW-AJK60 - 1396 | -144.50 | 58.50 | NW-AJK61 - 1397 | -144.50 | 59.00 | NW-AJK62 - 1398 | -144.50 | 59.50 | NW-AJK63 - 1399 | -131.50 | 53.90 | NW-AJK66 - 1400 | -133.50 | 53.90 | NW-AJK67 - 1401 | -134.00 | 53.90 | NW-AJK68 - 1402 | -134.50 | 53.90 | NW-AJK69 - 1403 | -135.00 | 53.90 | NW-AJK70 - 1404 | -135.50 | 53.90 | NW-AJK71 - 1405 | -136.00 | 53.90 | NW-AJK72 - 1406 | -136.50 | 53.90 | NW-AJK73 - 1407 | -137.00 | 53.90 | NW-AJK74 - 1408 | -137.50 | 53.90 | NW-AJK75 - 1409 | -138.00 | 53.90 | NW-AJK76 - 1410 | -138.50 | 53.90 | NW-AJK77 - 1411 | -139.00 | 53.90 | NW-AJK78 - 1412 | -139.50 | 53.90 | NW-AJK79 - 1413 | -140.00 | 53.90 | NW-AJK80 - 1414 | -140.50 | 53.90 | NW-AJK81 - 1415 | -141.00 | 53.90 | NW-AJK82 - 1416 | -141.50 | 53.90 | NW-AJK83 - 1417 | -142.00 | 53.90 | NW-AJK84 - 1418 | -142.50 | 53.90 | NW-AJK85 - 1419 | -143.00 | 53.90 | NW-AJK86 - 1420 | -143.50 | 53.90 | NW-AJK87 - 1421 | -144.00 | 53.90 | NW-AJK88 - 1422 | -144.50 | 53.90 | NW-AJK89 - 1423 | -158.00 | 55.00 | NW-AER51 - 1424 | -158.00 | 55.50 | NW-AER52 - 1425 | -158.00 | 56.00 | NW-AER53 - 1426 | -158.00 | 58.00 | NW-AER55 - 1427 | -142.85 | 59.35 | NW-AER57 - 1428 | -142.85 | 58.85 | NW-AER58 - 1429 | -142.85 | 58.35 | NW-AER59 - 1430 | -142.85 | 57.85 | NW-AER60 - 1431 | -142.85 | 57.35 | NW-AER61 - 1432 | -142.85 | 56.85 | NW-AER62 - 1433 | -142.85 | 56.35 | NW-AER63 - 1434 | -142.85 | 55.85 | NW-AER64 - 1435 | -142.85 | 55.35 | NW-AER65 - 1436 | -142.85 | 55.00 | NW-AER66 - 1437 | -143.35 | 55.00 | NW-AER67 - 1438 | -143.85 | 55.00 | NW-AER68 - 1439 | -144.35 | 55.00 | NW-AER69 - 1440 | -144.85 | 55.00 | NW-AER70 - 1441 | -145.35 | 55.00 | NW-AER71 - 1442 | -145.85 | 55.00 | NW-AER72 - 1443 | -146.35 | 55.00 | NW-AER73 - 1444 | -146.85 | 55.00 | NW-AER74 - 1445 | -147.35 | 55.00 | NW-AER75 - 1446 | -147.85 | 55.00 | NW-AER76 - 1447 | -148.35 | 55.00 | NW-AER77 - 1448 | -148.85 | 55.00 | NW-AER78 - 1449 | -149.35 | 55.00 | NW-AER79 - 1450 | -149.85 | 55.00 | NW-AER80 - 1451 | -150.35 | 55.00 | NW-AER81 - 1452 | -150.85 | 55.00 | NW-AER82 - 1453 | -151.35 | 55.00 | NW-AER83 - 1454 | -151.85 | 55.00 | NW-AER84 - 1455 | -152.35 | 55.00 | NW-AER85 - 1456 | -152.85 | 55.00 | NW-AER86 - 1457 | -153.35 | 55.00 | NW-AER87 - 1458 | -153.85 | 55.00 | NW-AER88 - 1459 | -154.35 | 55.00 | NW-AER89 - 1460 | -154.85 | 55.00 | NW-AER90 - 1461 | -155.35 | 55.00 | NW-AER91 - 1462 | -155.85 | 55.00 | NW-AER92 - 1463 | -156.35 | 55.00 | NW-AER93 - 1464 | -156.85 | 55.00 | NW-AER94 - 1465 | -157.35 | 55.00 | NW-AER95 - 1466 | -157.85 | 55.00 | NW-AER96 - 1467 | -158.00 | 55.00 | NW-AER97 - 1468 | -177.00 | 51.00 | NW-ALU51 - 1469 | -176.50 | 51.00 | NW-ALU52 - 1470 | -176.00 | 51.00 | NW-ALU53 - 1471 | -175.50 | 51.00 | NW-ALU54 - 1472 | -175.00 | 51.00 | NW-ALU55 - 1473 | -174.50 | 51.00 | NW-ALU56 - 1474 | -174.00 | 51.00 | NW-ALU57 - 1475 | -173.50 | 51.00 | NW-ALU58 - 1476 | -173.00 | 51.00 | NW-ALU59 - 1477 | -172.50 | 51.00 | NW-ALU60 - 1478 | -172.00 | 51.00 | NW-ALU61 - 1479 | -171.50 | 51.00 | NW-ALU62 - 1480 | -171.00 | 51.00 | NW-ALU63 - 1481 | -170.50 | 51.00 | NW-ALU64 - 1482 | -170.00 | 51.00 | NW-ALU65 - 1483 | -169.50 | 51.00 | NW-ALU66 - 1484 | -169.00 | 51.00 | NW-ALU67 - 1485 | -168.50 | 51.00 | NW-ALU68 - 1486 | -168.00 | 51.00 | NW-ALU69 - 1487 | -167.50 | 51.00 | NW-ALU70 - 1488 | -167.00 | 51.00 | NW-ALU71 - 1489 | -166.50 | 51.00 | NW-ALU72 - 1490 | -166.00 | 51.00 | NW-ALU73 - 1491 | -165.50 | 51.00 | NW-ALU74 - 1492 | -165.00 | 51.00 | NW-ALU75 - 1493 | -164.50 | 51.00 | NW-ALU76 - 1494 | -164.00 | 51.00 | NW-ALU77 - 1495 | -163.50 | 51.00 | NW-ALU78 - 1496 | -163.00 | 51.00 | NW-ALU79 - 1497 | -162.50 | 51.00 | NW-ALU80 - 1498 | -162.00 | 51.00 | NW-ALU81 - 1499 | -161.50 | 51.00 | NW-ALU82 - 1500 | -161.00 | 51.00 | NW-ALU83 - 1501 | -160.50 | 51.00 | NW-ALU84 - 1502 | -160.00 | 51.00 | NW-ALU85 - 1503 | -159.50 | 51.00 | NW-ALU86 - 1504 | -159.00 | 51.00 | NW-ALU87 - 1505 | -158.50 | 51.00 | NW-ALU88 - 1506 | -158.00 | 51.00 | NW-ALU89 - 1507 | -157.50 | 51.00 | NW-ALU90 - 1508 | -157.00 | 51.00 | NW-ALU91 - 1509 | -156.50 | 51.00 | NW-ALU92 - 1510 | -156.00 | 51.00 | NW-ALU93 - 1511 | -155.50 | 51.00 | NW-ALU94 - 1512 | -155.00 | 51.00 | NW-ALU95 - 1513 | -154.50 | 51.00 | NW-ALU96 - 1514 | -154.00 | 51.00 | NW-ALU97 - 1515 | -153.50 | 51.00 | NW-ALU98 - 1516 | -153.00 | 51.00 | NW-ALU99 - 1517 | -152.40 | 51.00 | NW-ALU100 - 1518 | -177.00 | 51.50 | NW-ALU101 - 1519 | -177.00 | 52.00 | NW-ALU102 - 1520 | -177.00 | 52.50 | NW-ALU103 - 1521 | -177.00 | 53.00 | NW-ALU104 - 1522 | -177.00 | 53.50 | NW-ALU105 - 1523 | -177.00 | 54.00 | NW-ALU106 - 1524 | -177.00 | 54.50 | NW-ALU107 - 1525 | -177.00 | 55.00 | NW-ALU108 - 1526 | -177.00 | 55.50 | NW-ALU109 - 1527 | -177.00 | 56.00 | NW-ALU110 - 1528 | -177.00 | 56.50 | NW-ALU111 - 1529 | -177.00 | 57.00 | NW-ALU112 - 1530 | -177.00 | 57.50 | NW-ALU113 - 1531 | -177.00 | 58.00 | NW-ALU114 - 1532 | -177.00 | 58.50 | NW-ALU115 - 1533 | -177.00 | 59.00 | NW-ALU116 - 1534 | -177.00 | 59.50 | NW-ALU117 - 1535 | -177.00 | 60.00 | NW-ALU118 - 1536 | -177.00 | 60.50 | NW-ALU119 - 1537 | -177.00 | 61.00 | NW-ALU120 - 1538 | -177.00 | 61.50 | NW-ALU121 - 1539 | -177.00 | 62.00 | NW-ALU122 - 1540 | -176.50 | 62.00 | NW-ALU123 - 1541 | -176.00 | 62.00 | NW-ALU124 - 1542 | -175.50 | 62.00 | NW-ALU125 - 1543 | -175.00 | 62.00 | NW-ALU126 - 1544 | -174.50 | 62.00 | NW-ALU127 - 1545 | -174.00 | 62.00 | NW-ALU128 - 1546 | -173.50 | 62.00 | NW-ALU129 - 1547 | -173.00 | 62.00 | NW-ALU130 - 1548 | -172.50 | 62.00 | NW-ALU131 - 1549 | -172.00 | 62.00 | NW-ALU132 - 1550 | -171.50 | 62.00 | NW-ALU133 - 1551 | -171.00 | 62.00 | NW-ALU134 - 1552 | -170.50 | 62.00 | NW-ALU135 - 1553 | -170.00 | 62.00 | NW-ALU136 - 1554 | -169.50 | 62.00 | NW-ALU137 - 1555 | -169.00 | 62.00 | NW-ALU138 - 1556 | -168.50 | 62.00 | NW-ALU139 - 1557 | -168.00 | 62.00 | NW-ALU140 - 1558 | -167.50 | 62.00 | NW-ALU141 - 1559 | -167.00 | 62.00 | NW-ALU142 - 1560 | -152.40 | 59.20 | NW-ALU147 - 1561 | -152.40 | 58.70 | NW-ALU148 - 1562 | -152.40 | 57.20 | NW-ALU149 - 1563 | -152.40 | 56.70 | NW-ALU150 - 1564 | -152.40 | 56.20 | NW-ALU151 - 1565 | -152.40 | 55.70 | NW-ALU152 - 1566 | -152.40 | 55.20 | NW-ALU153 - 1567 | -152.40 | 54.70 | NW-ALU154 - 1568 | -152.40 | 54.20 | NW-ALU155 - 1569 | -152.40 | 53.70 | NW-ALU156 - 1570 | -152.40 | 53.20 | NW-ALU157 - 1571 | -152.40 | 52.70 | NW-ALU158 - 1572 | -152.40 | 52.20 | NW-ALU159 - 1573 | -152.40 | 51.70 | NW-ALU160 - 1574 | -152.40 | 51.20 | NW-ALU161 - 1575 | -152.40 | 51.00 | NW-ALU162 - 1576 | -177.00 | 53.15 | NW-ALU163 - 1577 | -176.50 | 53.15 | NW-ALU164 - 1578 | -176.00 | 53.15 | NW-ALU165 - 1579 | -175.50 | 53.15 | NW-ALU166 - 1580 | -175.00 | 53.15 | NW-ALU167 - 1581 | -174.50 | 53.15 | NW-ALU168 - 1582 | -174.00 | 53.15 | NW-ALU169 - 1583 | -173.50 | 53.15 | NW-ALU170 - 1584 | -173.00 | 53.15 | NW-ALU171 - 1585 | -172.50 | 53.15 | NW-ALU172 - 1586 | -172.00 | 53.15 | NW-ALU173 - 1587 | -171.50 | 53.15 | NW-ALU174 - 1588 | -171.00 | 53.15 | NW-ALU175 - 1589 | -170.50 | 53.15 | NW-ALU176 - 1590 | -170.00 | 53.15 | NW-ALU177 - 1591 | -169.50 | 53.15 | NW-ALU178 - 1592 | -169.00 | 53.15 | NW-ALU179 - 1593 | -168.18 | 53.15 | NW-ALU180 - 1594 | -168.00 | 53.15 | NW-ALU181 - 1595 | -167.50 | 53.15 | NW-ALU182 - 1596 | -167.00 | 53.15 | NW-ALU183 - 1597 | -166.50 | 53.15 | NW-ALU184 - 1598 | -166.00 | 53.15 | NW-ALU185 - 1599 | -165.50 | 53.15 | NW-ALU186 - 1600 | -165.00 | 53.15 | NW-ALU187 - 1601 | -164.50 | 53.15 | NW-ALU188 - 1602 | -164.00 | 53.15 | NW-ALU189 - 1603 | -163.50 | 53.15 | NW-ALU190 - 1604 | -163.00 | 53.15 | NW-ALU191 - 1605 | -162.50 | 53.15 | NW-ALU192 - 1606 | -162.00 | 53.15 | NW-ALU193 - 1607 | -161.50 | 53.15 | NW-ALU194 - 1608 | -161.00 | 53.15 | NW-ALU195 - 1609 | -160.50 | 53.15 | NW-ALU196 - 1610 | -160.00 | 53.15 | NW-ALU197 - 1611 | -159.50 | 53.15 | NW-ALU198 - 1612 | -159.00 | 53.15 | NW-ALU199 - 1613 | -158.50 | 53.15 | NW-ALU200 - 1614 | -158.00 | 53.15 | NW-ALU201 - 1615 | -157.50 | 53.15 | NW-ALU202 - 1616 | -157.00 | 53.15 | NW-ALU203 - 1617 | -156.50 | 53.15 | NW-ALU204 - 1618 | -156.00 | 53.15 | NW-ALU205 - 1619 | -155.50 | 53.15 | NW-ALU206 - 1620 | -155.00 | 53.15 | NW-ALU207 - 1621 | -154.50 | 53.15 | NW-ALU208 - 1622 | -154.00 | 53.15 | NW-ALU209 - 1623 | -153.50 | 53.15 | NW-ALU210 - 1624 | -153.00 | 53.15 | NW-ALU211 - 1625 | -152.40 | 53.15 | NW-ALU212 - 1626 | -176.50 | 73.00 | NW-AFG51 - 1627 | -176.00 | 73.00 | NW-AFG52 - 1628 | -175.50 | 73.00 | NW-AFG53 - 1629 | -175.00 | 73.00 | NW-AFG54 - 1630 | -174.50 | 73.00 | NW-AFG55 - 1631 | -174.00 | 73.00 | NW-AFG56 - 1632 | -173.50 | 73.00 | NW-AFG57 - 1633 | -173.00 | 73.00 | NW-AFG58 - 1634 | -172.50 | 73.00 | NW-AFG59 - 1635 | -172.00 | 73.00 | NW-AFG60 - 1636 | -171.50 | 73.00 | NW-AFG61 - 1637 | -171.00 | 73.00 | NW-AFG62 - 1638 | -170.50 | 73.00 | NW-AFG63 - 1639 | -170.00 | 73.00 | NW-AFG64 - 1640 | -169.50 | 73.00 | NW-AFG65 - 1641 | -169.00 | 73.00 | NW-AFG66 - 1642 | -168.50 | 73.00 | NW-AFG67 - 1643 | -168.00 | 73.00 | NW-AFG68 - 1644 | -167.50 | 73.00 | NW-AFG69 - 1645 | -167.00 | 73.00 | NW-AFG70 - 1646 | -166.50 | 73.00 | NW-AFG71 - 1647 | -166.00 | 73.00 | NW-AFG72 - 1648 | -165.50 | 73.00 | NW-AFG73 - 1649 | -165.00 | 73.00 | NW-AFG74 - 1650 | -164.50 | 73.00 | NW-AFG75 - 1651 | -164.00 | 73.00 | NW-AFG76 - 1652 | -163.50 | 73.00 | NW-AFG77 - 1653 | -163.00 | 73.00 | NW-AFG78 - 1654 | -162.50 | 73.00 | NW-AFG79 - 1655 | -162.00 | 73.00 | NW-AFG80 - 1656 | -161.50 | 73.00 | NW-AFG81 - 1657 | -161.00 | 73.00 | NW-AFG82 - 1658 | -160.50 | 73.00 | NW-AFG83 - 1659 | -160.00 | 73.00 | NW-AFG84 - 1660 | -159.50 | 73.00 | NW-AFG85 - 1661 | -159.00 | 73.00 | NW-AFG86 - 1662 | -158.50 | 73.00 | NW-AFG87 - 1663 | -158.00 | 73.00 | NW-AFG88 - 1664 | -157.50 | 73.00 | NW-AFG89 - 1665 | -157.00 | 73.00 | NW-AFG90 - 1666 | -156.50 | 73.00 | NW-AFG91 - 1667 | -156.00 | 73.00 | NW-AFG92 - 1668 | -155.50 | 73.00 | NW-AFG93 - 1669 | -155.00 | 73.00 | NW-AFG94 - 1670 | -154.50 | 73.00 | NW-AFG95 - 1671 | -154.00 | 73.00 | NW-AFG96 - 1672 | -153.50 | 73.00 | NW-AFG97 - 1673 | -153.00 | 73.00 | NW-AFG98 - 1674 | -152.50 | 73.00 | NW-AFG99 - 1675 | -152.00 | 73.00 | NW-AFG100 - 1676 | -151.50 | 73.00 | NW-AFG101 - 1677 | -151.00 | 73.00 | NW-AFG102 - 1678 | -150.50 | 73.00 | NW-AFG103 - 1679 | -150.00 | 73.00 | NW-AFG104 - 1680 | -149.50 | 73.00 | NW-AFG105 - 1681 | -149.00 | 73.00 | NW-AFG106 - 1682 | -148.50 | 73.00 | NW-AFG107 - 1683 | -148.00 | 73.00 | NW-AFG108 - 1684 | -147.50 | 73.00 | NW-AFG109 - 1685 | -147.00 | 73.00 | NW-AFG110 - 1686 | -146.50 | 73.00 | NW-AFG111 - 1687 | -146.00 | 73.00 | NW-AFG112 - 1688 | -145.50 | 73.00 | NW-AFG113 - 1689 | -145.00 | 73.00 | NW-AFG114 - 1690 | -144.50 | 73.00 | NW-AFG115 - 1691 | -144.00 | 73.00 | NW-AFG116 - 1692 | -143.50 | 73.00 | NW-AFG117 - 1693 | -143.00 | 73.00 | NW-AFG118 - 1694 | -142.50 | 73.00 | NW-AFG119 - 1695 | -142.00 | 73.00 | NW-AFG120 - 1696 | -141.50 | 73.00 | NW-AFG121 - 1697 | -141.00 | 73.00 | NW-AFG122 - 1698 | -140.50 | 73.00 | NW-AFG123 - 1699 | -140.00 | 73.00 | NW-AFG124 - 1700 | -139.50 | 73.00 | NW-AFG125 - 1701 | -139.00 | 73.00 | NW-AFG126 - 1702 | -139.00 | 72.50 | NW-AFG127 - 1703 | -139.00 | 72.00 | NW-AFG128 - 1704 | -139.00 | 71.50 | NW-AFG129 - 1705 | -139.00 | 71.00 | NW-AFG130 - 1706 | -139.00 | 70.50 | NW-AFG131 - 1707 | -139.00 | 70.00 | NW-AFG132 - 1708 | -176.50 | 72.50 | NW-AFG134 - 1709 | -176.50 | 72.00 | NW-AFG135 - 1710 | -176.50 | 71.50 | NW-AFG136 - 1711 | -176.50 | 71.00 | NW-AFG137 - 1712 | -176.50 | 70.50 | NW-AFG138 - 1713 | -176.50 | 70.00 | NW-AFG139 - 1714 | -176.50 | 69.50 | NW-AFG140 - 1715 | -176.50 | 69.00 | NW-AFG141 - 1716 | -176.50 | 68.50 | NW-AFG142 - 1717 | -176.50 | 65.00 | NW-AFG143 - 1718 | -176.50 | 64.50 | NW-AFG144 - 1719 | -176.50 | 64.00 | NW-AFG145 - 1720 | -176.50 | 63.50 | NW-AFG146 - 1721 | -176.50 | 63.00 | NW-AFG147 - 1722 | -176.50 | 62.50 | NW-AFG148 - 1723 | -176.50 | 62.00 | NW-AFG149 - 1724 | -176.50 | 61.50 | NW-AFG150 - 1725 | -176.50 | 61.00 | NW-AFG151 - 1726 | -176.00 | 61.00 | NW-AFG152 - 1727 | -175.50 | 61.00 | NW-AFG153 - 1728 | -175.00 | 61.00 | NW-AFG154 - 1729 | -174.50 | 61.00 | NW-AFG155 - 1730 | -174.00 | 61.00 | NW-AFG156 - 1731 | -173.50 | 61.00 | NW-AFG157 - 1732 | -173.00 | 61.00 | NW-AFG158 - 1733 | -172.50 | 61.00 | NW-AFG159 - 1734 | -172.00 | 61.00 | NW-AFG160 - 1735 | -171.50 | 61.00 | NW-AFG161 - 1736 | -171.00 | 61.00 | NW-AFG162 - 1737 | -170.50 | 61.00 | NW-AFG163 - 1738 | -170.00 | 61.00 | NW-AFG164 - 1739 | -169.50 | 61.00 | NW-AFG165 - 1740 | -169.00 | 61.00 | NW-AFG166 - 1741 | -168.50 | 61.00 | NW-AFG167 - 1742 | -168.00 | 61.00 | NW-AFG168 - 1743 | -167.50 | 61.00 | NW-AFG169 - 1744 | -167.00 | 61.00 | NW-AFG170 - 1745 | 143.67 | 12.34 | NW-GUM51 - 1746 | 143.67 | 12.84 | NW-GUM52 - 1747 | 143.67 | 13.34 | NW-GUM53 - 1748 | 143.67 | 13.84 | NW-GUM54 - 1749 | 143.67 | 14.34 | NW-GUM55 - 1750 | 143.67 | 14.84 | NW-GUM56 - 1751 | 143.67 | 15.34 | NW-GUM57 - 1752 | 143.67 | 15.84 | NW-GUM58 - 1753 | 143.67 | 16.34 | NW-GUM59 - 1754 | 143.67 | 16.81 | NW-GUM60 - 1755 | 144.17 | 16.81 | NW-GUM61 - 1756 | 144.67 | 16.81 | NW-GUM62 - 1757 | 145.17 | 16.81 | NW-GUM63 - 1758 | 145.67 | 16.81 | NW-GUM64 - 1759 | 146.17 | 16.81 | NW-GUM65 - 1760 | 146.67 | 16.81 | NW-GUM66 - 1761 | 147.17 | 16.81 | NW-GUM67 - 1762 | 147.67 | 16.81 | NW-GUM68 - 1763 | 148.17 | 16.81 | NW-GUM69 - 1764 | 148.29 | 16.81 | NW-GUM70 - 1765 | 148.29 | 16.34 | NW-GUM71 - 1766 | 148.29 | 15.84 | NW-GUM72 - 1767 | 148.29 | 15.34 | NW-GUM73 - 1768 | 148.29 | 14.84 | NW-GUM74 - 1769 | 148.29 | 14.34 | NW-GUM75 - 1770 | 148.29 | 13.84 | NW-GUM76 - 1771 | 148.29 | 13.34 | NW-GUM77 - 1772 | 148.29 | 12.84 | NW-GUM78 - 1773 | 148.29 | 12.34 | NW-GUM79 - 1774 | 147.79 | 12.34 | NW-GUM80 - 1775 | 147.29 | 12.34 | NW-GUM81 - 1776 | 146.79 | 12.34 | NW-GUM82 - 1777 | 146.29 | 12.34 | NW-GUM83 - 1778 | 145.79 | 12.34 | NW-GUM84 - 1779 | 145.29 | 12.34 | NW-GUM85 - 1780 | 144.79 | 12.34 | NW-GUM86 - 1781 | 144.29 | 12.34 | NW-GUM87 - 1782 | 143.79 | 12.34 | NW-GUM88 - 1783 | 143.67 | 16.84 | NW-GUM89 - 1784 | 143.67 | 17.34 | NW-GUM90 - 1785 | 143.67 | 17.84 | NW-GUM91 - 1786 | 143.67 | 18.34 | NW-GUM92 - 1787 | 143.67 | 18.84 | NW-GUM93 - 1788 | 143.67 | 19.34 | NW-GUM94 - 1789 | 143.67 | 19.84 | NW-GUM95 - 1790 | 143.67 | 20.00 | NW-GUM96 - 1791 | 144.17 | 20.00 | NW-GUM97 - 1792 | 144.67 | 20.00 | NW-GUM98 - 1793 | 145.17 | 20.00 | NW-GUM99 - 1794 | 145.67 | 20.00 | NW-GUM100 - 1795 | 146.17 | 20.00 | NW-GUM101 - 1796 | 146.67 | 20.00 | NW-GUM102 - 1797 | 147.17 | 20.00 | NW-GUM103 - 1798 | 147.67 | 20.00 | NW-GUM104 - 1799 | 148.17 | 20.00 | NW-GUM105 - 1800 | 148.29 | 20.00 | NW-GUM106 - 1801 | 148.29 | 19.84 | NW-GUM107 - 1802 | 148.29 | 19.34 | NW-GUM108 - 1803 | 148.29 | 18.84 | NW-GUM109 - 1804 | 148.29 | 18.34 | NW-GUM110 - 1805 | 148.29 | 17.84 | NW-GUM111 - 1806 | 148.29 | 17.34 | NW-GUM112 - 1807 | 148.29 | 16.84 | NW-GUM113 - 1808 | -172.68 | -15.69 | NW-PGG50 - 1809 | -172.68 | -15.19 | NW-PGG51 - 1810 | -172.68 | -14.69 | NW-PGG52 - 1811 | -172.68 | -14.19 | NW-PGG53 - 1812 | -172.68 | -13.69 | NW-PGG54 - 1813 | -172.68 | -13.19 | NW-PGG55 - 1814 | -172.68 | -12.69 | NW-PGG56 - 1815 | -172.68 | -12.19 | NW-PGG57 - 1816 | -172.68 | -11.69 | NW-PGG58 - 1817 | -172.68 | -11.19 | NW-PGG59 - 1818 | -172.68 | -10.69 | NW-PGG60 - 1819 | -172.68 | -10.19 | NW-PGG61 - 1820 | -172.68 | -9.70 | NW-PGG62 - 1821 | -172.18 | -9.70 | NW-PGG63 - 1822 | -171.68 | -9.70 | NW-PGG64 - 1823 | -171.18 | -9.70 | NW-PGG65 - 1824 | -170.68 | -9.70 | NW-PGG66 - 1825 | -170.18 | -9.70 | NW-PGG67 - 1826 | -169.68 | -9.70 | NW-PGG68 - 1827 | -169.18 | -9.70 | NW-PGG69 - 1828 | -168.68 | -9.70 | NW-PGG70 - 1829 | -168.18 | -9.70 | NW-PGG71 - 1830 | -167.68 | -9.70 | NW-PGG72 - 1831 | -167.00 | -9.70 | NW-PGG73 - 1832 | -167.00 | -10.19 | NW-PGG74 - 1833 | -167.00 | -10.69 | NW-PGG75 - 1834 | -167.00 | -11.19 | NW-PGG76 - 1835 | -167.00 | -11.69 | NW-PGG77 - 1836 | -167.00 | -12.19 | NW-PGG78 - 1837 | -167.00 | -12.69 | NW-PGG79 - 1838 | -167.00 | -13.19 | NW-PGG80 - 1839 | -167.00 | -13.69 | NW-PGG81 - 1840 | -167.00 | -14.19 | NW-PGG82 - 1841 | -167.00 | -14.69 | NW-PGG83 - 1842 | -167.00 | -15.19 | NW-PGG84 - 1843 | -167.00 | -15.69 | NW-PGG85 - 1844 | -167.68 | -15.69 | NW-PGG86 - 1845 | -168.18 | -15.69 | NW-PGG87 - 1846 | -168.68 | -15.69 | NW-PGG88 - 1847 | -169.18 | -15.69 | NW-PGG89 - 1848 | -169.68 | -15.69 | NW-PGG90 - 1849 | -170.18 | -15.69 | NW-PGG91 - 1850 | -170.68 | -15.69 | NW-PGG92 - 1851 | -171.18 | -15.69 | NW-PGG93 - 1852 | -171.68 | -15.69 | NW-PGG94 - 1853 | -172.18 | -15.69 | NW-PGG95 - 1854 | -80.00 | 27.50 | NW-OPC51 - 1855 | -79.50 | 27.50 | NW-OPC52 - 1856 | -79.00 | 27.50 | NW-OPC53 - 1857 | -78.50 | 27.50 | NW-OPC54 - 1858 | -78.00 | 27.50 | NW-OPC55 - 1859 | -77.50 | 27.50 | NW-OPC56 - 1860 | -77.00 | 27.50 | NW-OPC57 - 1861 | -76.50 | 27.50 | NW-OPC58 - 1862 | -76.00 | 27.50 | NW-OPC59 - 1863 | -75.50 | 27.50 | NW-OPC60 - 1864 | -75.00 | 27.50 | NW-OPC61 - 1865 | -74.50 | 27.50 | NW-OPC62 - 1866 | -74.00 | 27.50 | NW-OPC63 - 1867 | -73.50 | 27.50 | NW-OPC64 - 1868 | -73.00 | 27.50 | NW-OPC65 - 1869 | -72.50 | 27.50 | NW-OPC66 - 1870 | -72.00 | 27.50 | NW-OPC67 - 1871 | -71.50 | 27.50 | NW-OPC68 - 1872 | -71.00 | 27.50 | NW-OPC69 - 1873 | -70.50 | 27.50 | NW-OPC70 - 1874 | -70.00 | 27.50 | NW-OPC71 - 1875 | -69.50 | 27.50 | NW-OPC72 - 1876 | -69.00 | 27.50 | NW-OPC73 - 1877 | -68.50 | 27.50 | NW-OPC74 - 1878 | -68.00 | 27.50 | NW-OPC75 - 1879 | -67.50 | 27.50 | NW-OPC76 - 1880 | -67.00 | 27.50 | NW-OPC77 - 1881 | -66.50 | 27.50 | NW-OPC78 - 1882 | -66.00 | 27.50 | NW-OPC79 - 1883 | -65.50 | 27.50 | NW-OPC80 - 1884 | -65.00 | 27.50 | NW-OPC81 - 1885 | -64.50 | 27.50 | NW-OPC82 - 1886 | -64.20 | 27.50 | NW-OPC83 - 1887 | -64.20 | 28.00 | NW-OPC84 - 1888 | -64.20 | 28.50 | NW-OPC85 - 1889 | -64.20 | 29.00 | NW-OPC86 - 1890 | -64.20 | 29.50 | NW-OPC87 - 1891 | -64.20 | 30.00 | NW-OPC88 - 1892 | -64.20 | 30.50 | NW-OPC89 - 1893 | -64.20 | 31.00 | NW-OPC90 - 1894 | -64.20 | 31.50 | NW-OPC91 - 1895 | -64.20 | 32.00 | NW-OPC92 - 1896 | -64.20 | 32.50 | NW-OPC93 - 1897 | -64.20 | 33.00 | NW-OPC94 - 1898 | -64.20 | 33.50 | NW-OPC95 - 1899 | -64.20 | 34.00 | NW-OPC96 - 1900 | -64.20 | 34.50 | NW-OPC97 - 1901 | -64.20 | 35.00 | NW-OPC98 - 1902 | -64.20 | 35.50 | NW-OPC99 - 1903 | -64.20 | 36.00 | NW-OPC100 - 1904 | -64.20 | 36.50 | NW-OPC101 - 1905 | -64.20 | 37.00 | NW-OPC102 - 1906 | -64.20 | 37.50 | NW-OPC103 - 1907 | -64.20 | 38.00 | NW-OPC104 - 1908 | -64.20 | 38.50 | NW-OPC105 - 1909 | -64.20 | 39.00 | NW-OPC106 - 1910 | -64.20 | 39.50 | NW-OPC107 - 1911 | -64.20 | 40.00 | NW-OPC108 - 1912 | -64.20 | 40.50 | NW-OPC109 - 1913 | -64.20 | 41.00 | NW-OPC110 - 1914 | -64.20 | 41.50 | NW-OPC111 - 1915 | -64.20 | 42.00 | NW-OPC112 - 1916 | -64.20 | 42.50 | NW-OPC113 - 1917 | -64.20 | 43.00 | NW-OPC114 - 1918 | -64.20 | 43.50 | NW-OPC115 - 1919 | -64.20 | 44.00 | NW-OPC116 - 1920 | -66.00 | 44.80 | NW-OPC117 - 1921 | -66.50 | 44.80 | NW-OPC118 - 1922 | 40.00 | -5.00 | HWRFa-50 - 1923 | 42.00 | -5.00 | HWRFa-51 - 1924 | 44.00 | -5.00 | HWRFa-52 - 1925 | 46.00 | -5.00 | HWRFa-53 - 1926 | 48.00 | -5.00 | HWRFa-54 - 1927 | 50.00 | -5.00 | HWRFa-55 - 1928 | 52.00 | -5.00 | HWRFa-56 - 1929 | 54.00 | -5.00 | HWRFa-57 - 1930 | 56.00 | -5.00 | HWRFa-58 - 1931 | 58.00 | -5.00 | HWRFa-59 - 1932 | 60.00 | -5.00 | HWRFa-60 - 1933 | 62.00 | -5.00 | HWRFa-61 - 1934 | 64.00 | -5.00 | HWRFa-62 - 1935 | 66.00 | -5.00 | HWRFa-63 - 1936 | 68.00 | -5.00 | HWRFa-64 - 1937 | 70.00 | -5.00 | HWRFa-65 - 1938 | 72.00 | -5.00 | HWRFa-66 - 1939 | 74.00 | -5.00 | HWRFa-67 - 1940 | 76.00 | -5.00 | HWRFa-68 - 1941 | 78.00 | -5.00 | HWRFa-69 - 1942 | 80.00 | -5.00 | HWRFa-70 - 1943 | 82.00 | -5.00 | HWRFa-71 - 1944 | 84.00 | -5.00 | HWRFa-72 - 1945 | 86.00 | -5.00 | HWRFa-73 - 1946 | 88.00 | -5.00 | HWRFa-74 - 1947 | 90.00 | -5.00 | HWRFa-75 - 1948 | 92.00 | -5.00 | HWRFa-76 - 1949 | 94.00 | -5.00 | HWRFa-77 - 1950 | 96.00 | -5.00 | HWRFa-78 - 1951 | 98.00 | -5.00 | HWRFa-79 - 1952 | 100.00 | -5.00 | HWRFa-80 - 1953 | 102.00 | -5.00 | HWRFa-81 - 1954 | 105.00 | -1.00 | HWRFa-82 - 1955 | 105.00 | 1.00 | HWRFa-83 - 1956 | 105.00 | 3.00 | HWRFa-84 - 1957 | 105.00 | 5.00 | HWRFa-85 - 1958 | 105.00 | 7.00 | HWRFa-86 - 1959 | 125.35 | 1.50 | HWRFc-50 - 1960 | 126.00 | 1.50 | HWRFc-51 - 1961 | 129.00 | 1.50 | HWRFc-54 - 1962 | 130.00 | 1.50 | HWRFc-55 - 1963 | 131.00 | 1.50 | HWRFc-56 - 1964 | 132.00 | 1.50 | HWRFc-57 - 1965 | 133.00 | 1.50 | HWRFc-58 - 1966 | 134.00 | 1.50 | HWRFc-59 - 1967 | 135.00 | 1.50 | HWRFc-60 - 1968 | 136.00 | 1.50 | HWRFc-61 - 1969 | 137.00 | 1.50 | HWRFc-62 - 1970 | 138.00 | 1.50 | HWRFc-63 - 1971 | 139.00 | 1.50 | HWRFc-64 - 1972 | 140.00 | 1.50 | HWRFc-65 - 1973 | 141.00 | 1.50 | HWRFc-66 - 1974 | 142.00 | 1.50 | HWRFc-67 - 1975 | 143.00 | 1.50 | HWRFc-68 - 1976 | 144.00 | 1.50 | HWRFc-69 - 1977 | 145.00 | 1.50 | HWRFc-70 - 1978 | 146.00 | 1.50 | HWRFc-71 - 1979 | 147.00 | 1.50 | HWRFc-72 - 1980 | 148.00 | 1.50 | HWRFc-73 - 1981 | 149.00 | 1.50 | HWRFc-74 - 1982 | 150.00 | 1.50 | HWRFc-75 - 1983 | 151.00 | 1.50 | HWRFc-76 - 1984 | 152.00 | 1.50 | HWRFc-77 - 1985 | 153.00 | 1.50 | HWRFc-78 - 1986 | 154.00 | 1.50 | HWRFc-79 - 1987 | 155.00 | 1.50 | HWRFc-80 - 1988 | 156.00 | 1.50 | HWRFc-81 - 1989 | 157.00 | 1.50 | HWRFc-82 - 1990 | 158.00 | 1.50 | HWRFc-83 - 1991 | 159.00 | 1.50 | HWRFc-84 - 1992 | 160.00 | 1.50 | HWRFc-85 - 1993 | 161.00 | 1.50 | HWRFc-86 - 1994 | 162.00 | 1.50 | HWRFc-87 - 1995 | 163.00 | 1.50 | HWRFc-88 - 1996 | 164.00 | 1.50 | HWRFc-89 - 1997 | 165.00 | 1.50 | HWRFc-90 - 1998 | 166.00 | 1.50 | HWRFc-91 - 1999 | 167.00 | 1.50 | HWRFc-92 - 2000 | 168.00 | 1.50 | HWRFc-93 - 2001 | 169.00 | 1.50 | HWRFc-94 - 2002 | 170.00 | 1.50 | HWRFc-95 - 2003 | 171.00 | 1.50 | HWRFc-96 - 2004 | 172.00 | 1.50 | HWRFc-97 - 2005 | 173.00 | 1.50 | HWRFc-98 - 2006 | 174.00 | 1.50 | HWRFc-99 - 2007 | 175.00 | 1.50 | HWRFc-100 - 2008 | 176.00 | 1.50 | HWRFc-101 - 2009 | 177.00 | 1.50 | HWRFc-102 - 2010 | 178.00 | 1.50 | HWRFc-103 - 2011 | 179.00 | 1.50 | HWRFc-104 - 2012 | 180.00 | 1.50 | HWRFc-105 - 2013 | 181.00 | 1.50 | HWRFc-106 - 2014 | 182.00 | 1.50 | HWRFc-107 - 2015 | 183.00 | 1.50 | HWRFc-108 - 2016 | 184.00 | 1.50 | HWRFc-109 - 2017 | 185.00 | 1.50 | HWRFc-110 - 2018 | 186.00 | 1.50 | HWRFc-111 - 2019 | 187.00 | 1.50 | HWRFc-112 - 2020 | 188.00 | 1.50 | HWRFc-113 - 2021 | 189.00 | 1.50 | HWRFc-114 - 2022 | 190.00 | 1.50 | HWRFc-115 - 2023 | 191.00 | 1.50 | HWRFc-116 - 2024 | 192.00 | 1.50 | HWRFc-117 - 2025 | 193.00 | 1.50 | HWRFc-118 - 2026 | 194.00 | 1.50 | HWRFc-119 - 2027 | 195.00 | 1.50 | HWRFc-120 - 2028 | 196.00 | 1.50 | HWRFc-121 - 2029 | 197.00 | 1.50 | HWRFc-122 - 2030 | 198.00 | 1.50 | HWRFc-123 - 2031 | 199.00 | 1.50 | HWRFc-124 - 2032 | 200.00 | 1.50 | HWRFc-125 - 2033 | 201.00 | 1.50 | HWRFc-126 - 2034 | 202.00 | 1.50 | HWRFc-127 - 2035 | 203.00 | 1.50 | HWRFc-128 - 2036 | 204.00 | 1.50 | HWRFc-129 - 2037 | 205.00 | 1.50 | HWRFc-130 - 2038 | 206.00 | 1.50 | HWRFc-131 - 2039 | 207.00 | 1.50 | HWRFc-132 - 2040 | 208.00 | 1.50 | HWRFc-133 - 2041 | 209.00 | 1.50 | HWRFc-134 - 2042 | 210.00 | 1.50 | HWRFc-135 - 2043 | 211.00 | 1.50 | HWRFc-136 - 2044 | 212.00 | 1.50 | HWRFc-137 - 2045 | 213.00 | 1.50 | HWRFc-138 - 2046 | 214.00 | 1.50 | HWRFc-139 - 2047 | 215.00 | 1.50 | HWRFc-140 - 2048 | 216.00 | 1.50 | HWRFc-141 - 2049 | 217.00 | 1.50 | HWRFc-142 - 2050 | 218.00 | 1.50 | HWRFc-143 - 2051 | 219.00 | 1.50 | HWRFc-144 - 2052 | 220.00 | 1.50 | HWRFc-145 - 2053 | 220.00 | 2.50 | HWRFc-146 - 2054 | 220.00 | 3.50 | HWRFc-147 - 2055 | 220.00 | 4.50 | HWRFc-148 - 2056 | 220.00 | 5.50 | HWRFc-149 - 2057 | 220.00 | 6.50 | HWRFc-150 - 2058 | 220.00 | 7.50 | HWRFc-151 - 2059 | 220.00 | 8.50 | HWRFc-152 - 2060 | 220.00 | 9.50 | HWRFc-153 - 2061 | 220.00 | 10.50 | HWRFc-154 - 2062 | 220.00 | 11.50 | HWRFc-155 - 2063 | 220.00 | 12.50 | HWRFc-156 - 2064 | 220.00 | 13.50 | HWRFc-157 - 2065 | 220.00 | 14.50 | HWRFc-158 - 2066 | 220.00 | 15.50 | HWRFc-159 - 2067 | 220.00 | 16.50 | HWRFc-160 - 2068 | 220.00 | 17.50 | HWRFc-161 - 2069 | 220.00 | 18.50 | HWRFc-162 - 2070 | 220.00 | 19.50 | HWRFc-163 - 2071 | 220.00 | 20.50 | HWRFc-164 - 2072 | 220.00 | 21.50 | HWRFc-165 - 2073 | 220.00 | 22.50 | HWRFc-166 - 2074 | 220.00 | 23.50 | HWRFc-167 - 2075 | 220.00 | 24.50 | HWRFc-168 - 2076 | 220.00 | 25.50 | HWRFc-169 - 2077 | 220.00 | 26.50 | HWRFc-170 - 2078 | 220.00 | 27.50 | HWRFc-171 - 2079 | 220.00 | 28.50 | HWRFc-172 - 2080 | 220.00 | 29.50 | HWRFc-173 - 2081 | 220.00 | 30.50 | HWRFc-174 - 2082 | 220.00 | 31.50 | HWRFc-175 - 2083 | 220.00 | 32.50 | HWRFc-176 - 2084 | 220.00 | 33.50 | HWRFc-177 - 2085 | 220.00 | 34.50 | HWRFc-178 - 2086 | 220.00 | 35.50 | HWRFc-179 - 2087 | 220.00 | 36.50 | HWRFc-180 - 2088 | 220.00 | 37.50 | HWRFc-181 - 2089 | 220.00 | 38.50 | HWRFc-182 - 2090 | 220.00 | 39.50 | HWRFc-183 - 2091 | 220.00 | 40.50 | HWRFc-184 - 2092 | 220.00 | 41.50 | HWRFc-185 - 2093 | 219.00 | 41.50 | HWRFc-186 - 2094 | 218.00 | 41.50 | HWRFc-187 - 2095 | 217.00 | 41.50 | HWRFc-188 - 2096 | 216.00 | 41.50 | HWRFc-189 - 2097 | 215.00 | 41.50 | HWRFc-190 - 2098 | 214.00 | 41.50 | HWRFc-191 - 2099 | 213.00 | 41.50 | HWRFc-192 - 2100 | 212.00 | 41.50 | HWRFc-193 - 2101 | 211.00 | 41.50 | HWRFc-194 - 2102 | 210.00 | 41.50 | HWRFc-195 - 2103 | 209.00 | 41.50 | HWRFc-196 - 2104 | 208.00 | 41.50 | HWRFc-197 - 2105 | 207.00 | 41.50 | HWRFc-198 - 2106 | 206.00 | 41.50 | HWRFc-199 - 2107 | 205.00 | 41.50 | HWRFc-200 - 2108 | 204.00 | 41.50 | HWRFc-201 - 2109 | 203.00 | 41.50 | HWRFc-202 - 2110 | 202.00 | 41.50 | HWRFc-203 - 2111 | 201.00 | 41.50 | HWRFc-204 - 2112 | 200.00 | 41.50 | HWRFc-205 - 2113 | 199.00 | 41.50 | HWRFc-206 - 2114 | 198.00 | 41.50 | HWRFc-207 - 2115 | 197.00 | 41.50 | HWRFc-208 - 2116 | 196.00 | 41.50 | HWRFc-209 - 2117 | 195.00 | 41.50 | HWRFc-210 - 2118 | 194.00 | 41.50 | HWRFc-211 - 2119 | 193.00 | 41.50 | HWRFc-212 - 2120 | 192.00 | 41.50 | HWRFc-213 - 2121 | 191.00 | 41.50 | HWRFc-214 - 2122 | 190.00 | 41.50 | HWRFc-215 - 2123 | 189.00 | 41.50 | HWRFc-216 - 2124 | 188.00 | 41.50 | HWRFc-217 - 2125 | 187.00 | 41.50 | HWRFc-218 - 2126 | 186.00 | 41.50 | HWRFc-219 - 2127 | 185.00 | 41.50 | HWRFc-220 - 2128 | 184.00 | 41.50 | HWRFc-221 - 2129 | 183.00 | 41.50 | HWRFc-222 - 2130 | 182.00 | 41.50 | HWRFc-223 - 2131 | 181.00 | 41.50 | HWRFc-224 - 2132 | 180.00 | 41.50 | HWRFc-225 - 2133 | 179.00 | 41.50 | HWRFc-226 - 2134 | 178.00 | 41.50 | HWRFc-227 - 2135 | 177.00 | 41.50 | HWRFc-228 - 2136 | 176.00 | 41.50 | HWRFc-229 - 2137 | 175.00 | 41.50 | HWRFc-230 - 2138 | 174.00 | 41.50 | HWRFc-231 - 2139 | 173.00 | 41.50 | HWRFc-232 - 2140 | 172.00 | 41.50 | HWRFc-233 - 2141 | 171.00 | 41.50 | HWRFc-234 - 2142 | 170.00 | 41.50 | HWRFc-235 - 2143 | 169.00 | 41.50 | HWRFc-236 - 2144 | 168.00 | 41.50 | HWRFc-237 - 2145 | 167.00 | 41.50 | HWRFc-238 - 2146 | 166.00 | 41.50 | HWRFc-239 - 2147 | 165.00 | 41.50 | HWRFc-240 - 2148 | 164.00 | 41.50 | HWRFc-241 - 2149 | 163.00 | 41.50 | HWRFc-242 - 2150 | 162.00 | 41.50 | HWRFc-243 - 2151 | 161.00 | 41.50 | HWRFc-244 - 2152 | 160.00 | 41.50 | HWRFc-245 - 2153 | 159.00 | 41.50 | HWRFc-246 - 2154 | 158.00 | 41.50 | HWRFc-247 - 2155 | 157.00 | 41.50 | HWRFc-248 - 2156 | 156.00 | 41.50 | HWRFc-249 - 2157 | 155.00 | 41.50 | HWRFc-250 - 2158 | 154.00 | 41.50 | HWRFc-251 - 2159 | 153.00 | 41.50 | HWRFc-252 - 2160 | 152.00 | 41.50 | HWRFc-253 - 2161 | 151.00 | 41.50 | HWRFc-254 - 2162 | 150.00 | 41.50 | HWRFc-255 - 2163 | 149.00 | 41.50 | HWRFc-256 - 2164 | 148.00 | 41.50 | HWRFc-257 - 2165 | 147.00 | 41.50 | HWRFc-258 - 2166 | 146.00 | 41.50 | HWRFc-259 - 2167 | 145.00 | 41.50 | HWRFc-260 - 2168 | 144.00 | 41.50 | HWRFc-261 - 2169 | 143.00 | 41.50 | HWRFc-262 - 2170 | 142.00 | 41.50 | HWRFc-263 - 2171 | 141.00 | 41.50 | HWRFc-264 - 2172 | 140.00 | 41.50 | HWRFc-265 - 2173 | 139.00 | 41.50 | HWRFc-266 - 2174 | 138.00 | 41.50 | HWRFc-267 - 2175 | 137.00 | 41.50 | HWRFc-268 - 2176 | 136.00 | 41.50 | HWRFc-269 - 2177 | 135.00 | 41.50 | HWRFc-270 - 2178 | 134.00 | 41.50 | HWRFc-271 - 2179 | 133.00 | 41.50 | HWRFc-272 - 2180 | 132.00 | 41.50 | HWRFc-273 - 2181 | 125.00 | 37.50 | HWRFc-276 - 2182 | 125.00 | 36.50 | HWRFc-277 - 2183 | 125.00 | 35.50 | HWRFc-278 - 2184 | 125.00 | 34.50 | HWRFc-279 - 2185 | 125.00 | 33.50 | HWRFc-280 - 2186 | 125.00 | 32.50 | HWRFc-281 - 2187 | 125.00 | 31.50 | HWRFc-282 - 2188 | 125.00 | 30.50 | HWRFc-283 - 2189 | 125.00 | 29.50 | HWRFc-284 - 2190 | 125.00 | 28.50 | HWRFc-285 - 2191 | 125.00 | 27.50 | HWRFc-286 - 2192 | 125.00 | 26.50 | HWRFc-287 - 2193 | 125.00 | 25.50 | HWRFc-288 - 2194 | 125.00 | 24.50 | HWRFc-289 - 2195 | 125.00 | 23.50 | HWRFc-290 - 2196 | 125.00 | 22.50 | HWRFc-291 - 2197 | 125.00 | 21.50 | HWRFc-292 - 2198 | 125.00 | 20.50 | HWRFc-293 - 2199 | 125.00 | 19.50 | HWRFc-294 - 2200 | 125.00 | 18.50 | HWRFc-295 - 2201 | 125.00 | 17.50 | HWRFc-296 - 2202 | 125.00 | 16.50 | HWRFc-297 - 2203 | 125.00 | 15.50 | HWRFc-298 - 2204 | 125.00 | 14.50 | HWRFc-299 - 2205 | 125.00 | 13.50 | HWRFc-300 - 2206 | 125.00 | 13.00 | HWRFc-301 - 2207 | 125.00 | 12.60 | HWRFc-302 - 2208 | 125.00 | 10.00 | HWRFc-303 - 2209 | 125.00 | 9.50 | HWRFc-304 - 2210 | 125.00 | 5.50 | HWRFc-307 - 2211 | 125.00 | 5.00 | HWRFc-308 - 2212 | 125.00 | 4.50 | HWRFc-309 - 2213 | 125.00 | 3.50 | HWRFc-310 - 2214 | 125.00 | 2.50 | HWRFc-311 - 2215 | 180.00 | 1.50 | HWRFe-50 - 2216 | 180.50 | 1.50 | HWRFe-51 - 2217 | 181.00 | 1.50 | HWRFe-52 - 2218 | 181.50 | 1.50 | HWRFe-53 - 2219 | 182.00 | 1.50 | HWRFe-54 - 2220 | 182.50 | 1.50 | HWRFe-55 - 2221 | 183.00 | 1.50 | HWRFe-56 - 2222 | 183.50 | 1.50 | HWRFe-57 - 2223 | 184.00 | 1.50 | HWRFe-58 - 2224 | 184.50 | 1.50 | HWRFe-59 - 2225 | 185.00 | 1.50 | HWRFe-60 - 2226 | 185.50 | 1.50 | HWRFe-61 - 2227 | 186.00 | 1.50 | HWRFe-62 - 2228 | 186.50 | 1.50 | HWRFe-63 - 2229 | 187.00 | 1.50 | HWRFe-64 - 2230 | 187.50 | 1.50 | HWRFe-65 - 2231 | 188.00 | 1.50 | HWRFe-66 - 2232 | 188.50 | 1.50 | HWRFe-67 - 2233 | 189.00 | 1.50 | HWRFe-68 - 2234 | 189.50 | 1.50 | HWRFe-69 - 2235 | 190.00 | 1.50 | HWRFe-70 - 2236 | 190.50 | 1.50 | HWRFe-71 - 2237 | 191.00 | 1.50 | HWRFe-72 - 2238 | 191.50 | 1.50 | HWRFe-73 - 2239 | 192.00 | 1.50 | HWRFe-74 - 2240 | 192.50 | 1.50 | HWRFe-75 - 2241 | 193.00 | 1.50 | HWRFe-76 - 2242 | 193.50 | 1.50 | HWRFe-77 - 2243 | 194.00 | 1.50 | HWRFe-78 - 2244 | 194.50 | 1.50 | HWRFe-79 - 2245 | 195.00 | 1.50 | HWRFe-80 - 2246 | 195.50 | 1.50 | HWRFe-81 - 2247 | 196.00 | 1.50 | HWRFe-82 - 2248 | 196.50 | 1.50 | HWRFe-83 - 2249 | 197.00 | 1.50 | HWRFe-84 - 2250 | 197.50 | 1.50 | HWRFe-85 - 2251 | 198.00 | 1.50 | HWRFe-86 - 2252 | 198.50 | 1.50 | HWRFe-87 - 2253 | 199.00 | 1.50 | HWRFe-88 - 2254 | 199.50 | 1.50 | HWRFe-89 - 2255 | 200.00 | 1.50 | HWRFe-90 - 2256 | 200.50 | 1.50 | HWRFe-91 - 2257 | 201.00 | 1.50 | HWRFe-92 - 2258 | 201.50 | 1.50 | HWRFe-93 - 2259 | 202.00 | 1.50 | HWRFe-94 - 2260 | 202.50 | 1.50 | HWRFe-95 - 2261 | 203.00 | 1.50 | HWRFe-96 - 2262 | 203.50 | 1.50 | HWRFe-97 - 2263 | 204.00 | 1.50 | HWRFe-98 - 2264 | 204.50 | 1.50 | HWRFe-99 - 2265 | 205.00 | 1.50 | HWRFe-100 - 2266 | 205.50 | 1.50 | HWRFe-101 - 2267 | 206.00 | 1.50 | HWRFe-102 - 2268 | 206.50 | 1.50 | HWRFe-103 - 2269 | 207.00 | 1.50 | HWRFe-104 - 2270 | 207.50 | 1.50 | HWRFe-105 - 2271 | 208.00 | 1.50 | HWRFe-106 - 2272 | 208.50 | 1.50 | HWRFe-107 - 2273 | 209.00 | 1.50 | HWRFe-108 - 2274 | 209.50 | 1.50 | HWRFe-109 - 2275 | 210.00 | 1.50 | HWRFe-110 - 2276 | 210.50 | 1.50 | HWRFe-111 - 2277 | 211.00 | 1.50 | HWRFe-112 - 2278 | 211.50 | 1.50 | HWRFe-113 - 2279 | 212.00 | 1.50 | HWRFe-114 - 2280 | 212.50 | 1.50 | HWRFe-115 - 2281 | 213.00 | 1.50 | HWRFe-116 - 2282 | 213.50 | 1.50 | HWRFe-117 - 2283 | 214.00 | 1.50 | HWRFe-118 - 2284 | 214.50 | 1.50 | HWRFe-119 - 2285 | 215.00 | 1.50 | HWRFe-120 - 2286 | 215.50 | 1.50 | HWRFe-121 - 2287 | 216.00 | 1.50 | HWRFe-122 - 2288 | 216.50 | 1.50 | HWRFe-123 - 2289 | 217.00 | 1.50 | HWRFe-124 - 2290 | 217.50 | 1.50 | HWRFe-125 - 2291 | 218.00 | 1.50 | HWRFe-126 - 2292 | 218.50 | 1.50 | HWRFe-127 - 2293 | 219.00 | 1.50 | HWRFe-128 - 2294 | 219.50 | 1.50 | HWRFe-129 - 2295 | 220.00 | 1.50 | HWRFe-130 - 2296 | 220.50 | 1.50 | HWRFe-131 - 2297 | 221.00 | 1.50 | HWRFe-132 - 2298 | 221.50 | 1.50 | HWRFe-133 - 2299 | 222.00 | 1.50 | HWRFe-134 - 2300 | 222.50 | 1.50 | HWRFe-135 - 2301 | 223.00 | 1.50 | HWRFe-136 - 2302 | 223.50 | 1.50 | HWRFe-137 - 2303 | 224.00 | 1.50 | HWRFe-138 - 2304 | 224.50 | 1.50 | HWRFe-139 - 2305 | 225.00 | 1.50 | HWRFe-140 - 2306 | 225.50 | 1.50 | HWRFe-141 - 2307 | 226.00 | 1.50 | HWRFe-142 - 2308 | 226.50 | 1.50 | HWRFe-143 - 2309 | 227.00 | 1.50 | HWRFe-144 - 2310 | 227.50 | 1.50 | HWRFe-145 - 2311 | 228.00 | 1.50 | HWRFe-146 - 2312 | 228.50 | 1.50 | HWRFe-147 - 2313 | 229.00 | 1.50 | HWRFe-148 - 2314 | 229.50 | 1.50 | HWRFe-149 - 2315 | 230.00 | 1.50 | HWRFe-150 - 2316 | 230.50 | 1.50 | HWRFe-151 - 2317 | 231.00 | 1.50 | HWRFe-152 - 2318 | 231.50 | 1.50 | HWRFe-153 - 2319 | 232.00 | 1.50 | HWRFe-154 - 2320 | 232.50 | 1.50 | HWRFe-155 - 2321 | 233.00 | 1.50 | HWRFe-156 - 2322 | 233.50 | 1.50 | HWRFe-157 - 2323 | 234.00 | 1.50 | HWRFe-158 - 2324 | 234.50 | 1.50 | HWRFe-159 - 2325 | 235.00 | 1.50 | HWRFe-160 - 2326 | 235.50 | 1.50 | HWRFe-161 - 2327 | 236.00 | 1.50 | HWRFe-162 - 2328 | 236.50 | 1.50 | HWRFe-163 - 2329 | 237.00 | 1.50 | HWRFe-164 - 2330 | 237.50 | 1.50 | HWRFe-165 - 2331 | 238.00 | 1.50 | HWRFe-166 - 2332 | 238.50 | 1.50 | HWRFe-167 - 2333 | 239.00 | 1.50 | HWRFe-168 - 2334 | 239.50 | 1.50 | HWRFe-169 - 2335 | 240.00 | 1.50 | HWRFe-170 - 2336 | 240.50 | 1.50 | HWRFe-171 - 2337 | 241.00 | 1.50 | HWRFe-172 - 2338 | 241.50 | 1.50 | HWRFe-173 - 2339 | 242.00 | 1.50 | HWRFe-174 - 2340 | 242.50 | 1.50 | HWRFe-175 - 2341 | 243.00 | 1.50 | HWRFe-176 - 2342 | 243.50 | 1.50 | HWRFe-177 - 2343 | 244.00 | 1.50 | HWRFe-178 - 2344 | 244.50 | 1.50 | HWRFe-179 - 2345 | 245.00 | 1.50 | HWRFe-180 - 2346 | 245.50 | 1.50 | HWRFe-181 - 2347 | 246.00 | 1.50 | HWRFe-182 - 2348 | 246.50 | 1.50 | HWRFe-183 - 2349 | 247.00 | 1.50 | HWRFe-184 - 2350 | 247.50 | 1.50 | HWRFe-185 - 2351 | 248.00 | 1.50 | HWRFe-186 - 2352 | 248.50 | 1.50 | HWRFe-187 - 2353 | 249.00 | 1.50 | HWRFe-188 - 2354 | 249.50 | 1.50 | HWRFe-189 - 2355 | 250.00 | 1.50 | HWRFe-190 - 2356 | 250.50 | 1.50 | HWRFe-191 - 2357 | 251.00 | 1.50 | HWRFe-192 - 2358 | 251.50 | 1.50 | HWRFe-193 - 2359 | 252.00 | 1.50 | HWRFe-194 - 2360 | 252.50 | 1.50 | HWRFe-195 - 2361 | 253.00 | 1.50 | HWRFe-196 - 2362 | 253.50 | 1.50 | HWRFe-197 - 2363 | 254.00 | 1.50 | HWRFe-198 - 2364 | 254.50 | 1.50 | HWRFe-199 - 2365 | 255.00 | 1.50 | HWRFe-200 - 2366 | 255.50 | 1.50 | HWRFe-201 - 2367 | 256.00 | 1.50 | HWRFe-202 - 2368 | 256.50 | 1.50 | HWRFe-203 - 2369 | 257.00 | 1.50 | HWRFe-204 - 2370 | 257.50 | 1.50 | HWRFe-205 - 2371 | 258.00 | 1.50 | HWRFe-206 - 2372 | 258.50 | 1.50 | HWRFe-207 - 2373 | 259.00 | 1.50 | HWRFe-208 - 2374 | 259.50 | 1.50 | HWRFe-209 - 2375 | 260.00 | 1.50 | HWRFe-210 - 2376 | 260.50 | 1.50 | HWRFe-211 - 2377 | 261.00 | 1.50 | HWRFe-212 - 2378 | 261.50 | 1.50 | HWRFe-213 - 2379 | 262.00 | 1.50 | HWRFe-214 - 2380 | 262.50 | 1.50 | HWRFe-215 - 2381 | 263.00 | 1.50 | HWRFe-216 - 2382 | 263.50 | 1.50 | HWRFe-217 - 2383 | 264.00 | 1.50 | HWRFe-218 - 2384 | 264.50 | 1.50 | HWRFe-219 - 2385 | 265.00 | 1.50 | HWRFe-220 - 2386 | 265.50 | 1.50 | HWRFe-221 - 2387 | 266.00 | 1.50 | HWRFe-222 - 2388 | 266.50 | 1.50 | HWRFe-223 - 2389 | 267.00 | 1.50 | HWRFe-224 - 2390 | 267.50 | 1.50 | HWRFe-225 - 2391 | 268.00 | 1.50 | HWRFe-226 - 2392 | 268.50 | 1.50 | HWRFe-227 - 2393 | 269.00 | 1.50 | HWRFe-228 - 2394 | 269.50 | 1.50 | HWRFe-229 - 2395 | 270.00 | 1.50 | HWRFe-230 - 2396 | 270.50 | 1.50 | HWRFe-231 - 2397 | 271.00 | 1.50 | HWRFe-232 - 2398 | 271.50 | 1.50 | HWRFe-233 - 2399 | 272.00 | 1.50 | HWRFe-234 - 2400 | 272.50 | 1.50 | HWRFe-235 - 2401 | 273.00 | 1.50 | HWRFe-236 - 2402 | 273.50 | 1.50 | HWRFe-237 - 2403 | 274.00 | 1.50 | HWRFe-238 - 2404 | 274.50 | 1.50 | HWRFe-239 - 2405 | 275.00 | 1.50 | HWRFe-240 - 2406 | 275.50 | 1.50 | HWRFe-241 - 2407 | 276.00 | 1.50 | HWRFe-242 - 2408 | 276.50 | 1.50 | HWRFe-243 - 2409 | 277.00 | 1.50 | HWRFe-244 - 2410 | 277.50 | 1.50 | HWRFe-245 - 2411 | 278.00 | 1.50 | HWRFe-246 - 2412 | 234.00 | 41.50 | HWRFe-247 - 2413 | 233.50 | 41.50 | HWRFe-248 - 2414 | 233.00 | 41.50 | HWRFe-249 - 2415 | 232.50 | 41.50 | HWRFe-250 - 2416 | 232.00 | 41.50 | HWRFe-251 - 2417 | 231.50 | 41.50 | HWRFe-252 - 2418 | 231.00 | 41.50 | HWRFe-253 - 2419 | 230.50 | 41.50 | HWRFe-254 - 2420 | 230.00 | 41.50 | HWRFe-255 - 2421 | 229.50 | 41.50 | HWRFe-256 - 2422 | 229.00 | 41.50 | HWRFe-257 - 2423 | 228.50 | 41.50 | HWRFe-258 - 2424 | 228.00 | 41.50 | HWRFe-259 - 2425 | 227.50 | 41.50 | HWRFe-260 - 2426 | 227.00 | 41.50 | HWRFe-261 - 2427 | 226.50 | 41.50 | HWRFe-262 - 2428 | 226.00 | 41.50 | HWRFe-263 - 2429 | 225.50 | 41.50 | HWRFe-264 - 2430 | 225.00 | 41.50 | HWRFe-265 - 2431 | 224.50 | 41.50 | HWRFe-266 - 2432 | 224.00 | 41.50 | HWRFe-267 - 2433 | 223.50 | 41.50 | HWRFe-268 - 2434 | 223.00 | 41.50 | HWRFe-269 - 2435 | 222.50 | 41.50 | HWRFe-270 - 2436 | 222.00 | 41.50 | HWRFe-271 - 2437 | 221.50 | 41.50 | HWRFe-272 - 2438 | 221.00 | 41.50 | HWRFe-273 - 2439 | 220.50 | 41.50 | HWRFe-274 - 2440 | 220.00 | 41.50 | HWRFe-275 - 2441 | 219.50 | 41.50 | HWRFe-276 - 2442 | 219.00 | 41.50 | HWRFe-277 - 2443 | 218.50 | 41.50 | HWRFe-278 - 2444 | 218.00 | 41.50 | HWRFe-279 - 2445 | 217.50 | 41.50 | HWRFe-280 - 2446 | 217.00 | 41.50 | HWRFe-281 - 2447 | 216.50 | 41.50 | HWRFe-282 - 2448 | 216.00 | 41.50 | HWRFe-283 - 2449 | 215.50 | 41.50 | HWRFe-284 - 2450 | 215.00 | 41.50 | HWRFe-285 - 2451 | 214.50 | 41.50 | HWRFe-286 - 2452 | 214.00 | 41.50 | HWRFe-287 - 2453 | 213.50 | 41.50 | HWRFe-288 - 2454 | 213.00 | 41.50 | HWRFe-289 - 2455 | 212.50 | 41.50 | HWRFe-290 - 2456 | 212.00 | 41.50 | HWRFe-291 - 2457 | 211.50 | 41.50 | HWRFe-292 - 2458 | 211.00 | 41.50 | HWRFe-293 - 2459 | 210.50 | 41.50 | HWRFe-294 - 2460 | 210.00 | 41.50 | HWRFe-295 - 2461 | 209.50 | 41.50 | HWRFe-296 - 2462 | 209.00 | 41.50 | HWRFe-297 - 2463 | 208.50 | 41.50 | HWRFe-298 - 2464 | 208.00 | 41.50 | HWRFe-299 - 2465 | 207.50 | 41.50 | HWRFe-300 - 2466 | 207.00 | 41.50 | HWRFe-301 - 2467 | 206.50 | 41.50 | HWRFe-302 - 2468 | 206.00 | 41.50 | HWRFe-303 - 2469 | 205.50 | 41.50 | HWRFe-304 - 2470 | 205.00 | 41.50 | HWRFe-305 - 2471 | 204.50 | 41.50 | HWRFe-306 - 2472 | 204.00 | 41.50 | HWRFe-307 - 2473 | 203.50 | 41.50 | HWRFe-308 - 2474 | 203.00 | 41.50 | HWRFe-309 - 2475 | 202.50 | 41.50 | HWRFe-310 - 2476 | 202.00 | 41.50 | HWRFe-311 - 2477 | 201.50 | 41.50 | HWRFe-312 - 2478 | 201.00 | 41.50 | HWRFe-313 - 2479 | 200.50 | 41.50 | HWRFe-314 - 2480 | 200.00 | 41.50 | HWRFe-315 - 2481 | 199.50 | 41.50 | HWRFe-316 - 2482 | 199.00 | 41.50 | HWRFe-317 - 2483 | 198.50 | 41.50 | HWRFe-318 - 2484 | 198.00 | 41.50 | HWRFe-319 - 2485 | 197.50 | 41.50 | HWRFe-320 - 2486 | 197.00 | 41.50 | HWRFe-321 - 2487 | 196.50 | 41.50 | HWRFe-322 - 2488 | 196.00 | 41.50 | HWRFe-323 - 2489 | 195.50 | 41.50 | HWRFe-324 - 2490 | 195.00 | 41.50 | HWRFe-325 - 2491 | 194.50 | 41.50 | HWRFe-326 - 2492 | 194.00 | 41.50 | HWRFe-327 - 2493 | 193.50 | 41.50 | HWRFe-328 - 2494 | 193.00 | 41.50 | HWRFe-329 - 2495 | 192.50 | 41.50 | HWRFe-330 - 2496 | 192.00 | 41.50 | HWRFe-331 - 2497 | 191.50 | 41.50 | HWRFe-332 - 2498 | 191.00 | 41.50 | HWRFe-333 - 2499 | 190.50 | 41.50 | HWRFe-334 - 2500 | 190.00 | 41.50 | HWRFe-335 - 2501 | 189.50 | 41.50 | HWRFe-336 - 2502 | 189.00 | 41.50 | HWRFe-337 - 2503 | 188.50 | 41.50 | HWRFe-338 - 2504 | 188.00 | 41.50 | HWRFe-339 - 2505 | 187.50 | 41.50 | HWRFe-340 - 2506 | 187.00 | 41.50 | HWRFe-341 - 2507 | 186.50 | 41.50 | HWRFe-342 - 2508 | 186.00 | 41.50 | HWRFe-343 - 2509 | 185.50 | 41.50 | HWRFe-344 - 2510 | 185.00 | 41.50 | HWRFe-345 - 2511 | 184.50 | 41.50 | HWRFe-346 - 2512 | 184.00 | 41.50 | HWRFe-347 - 2513 | 183.50 | 41.50 | HWRFe-348 - 2514 | 183.00 | 41.50 | HWRFe-349 - 2515 | 182.50 | 41.50 | HWRFe-350 - 2516 | 182.00 | 41.50 | HWRFe-351 - 2517 | 181.50 | 41.50 | HWRFe-352 - 2518 | 181.00 | 41.50 | HWRFe-353 - 2519 | 180.50 | 41.50 | HWRFe-354 - 2520 | 180.00 | 41.50 | HWRFe-355 - 2521 | 180.00 | 41.00 | HWRFe-356 - 2522 | 180.00 | 40.50 | HWRFe-357 - 2523 | 180.00 | 40.00 | HWRFe-358 - 2524 | 180.00 | 39.50 | HWRFe-359 - 2525 | 180.00 | 39.00 | HWRFe-360 - 2526 | 180.00 | 38.50 | HWRFe-361 - 2527 | 180.00 | 38.00 | HWRFe-362 - 2528 | 180.00 | 37.50 | HWRFe-363 - 2529 | 180.00 | 37.00 | HWRFe-364 - 2530 | 180.00 | 36.50 | HWRFe-365 - 2531 | 180.00 | 36.00 | HWRFe-366 - 2532 | 180.00 | 35.50 | HWRFe-367 - 2533 | 180.00 | 35.00 | HWRFe-368 - 2534 | 180.00 | 34.50 | HWRFe-369 - 2535 | 180.00 | 34.00 | HWRFe-370 - 2536 | 180.00 | 33.50 | HWRFe-371 - 2537 | 180.00 | 33.00 | HWRFe-372 - 2538 | 180.00 | 32.50 | HWRFe-373 - 2539 | 180.00 | 32.00 | HWRFe-374 - 2540 | 180.00 | 31.50 | HWRFe-375 - 2541 | 180.00 | 31.00 | HWRFe-376 - 2542 | 180.00 | 30.50 | HWRFe-377 - 2543 | 180.00 | 30.00 | HWRFe-378 - 2544 | 180.00 | 29.50 | HWRFe-379 - 2545 | 180.00 | 29.00 | HWRFe-380 - 2546 | 180.00 | 28.50 | HWRFe-381 - 2547 | 180.00 | 28.00 | HWRFe-382 - 2548 | 180.00 | 27.50 | HWRFe-383 - 2549 | 180.00 | 27.00 | HWRFe-384 - 2550 | 180.00 | 26.50 | HWRFe-385 - 2551 | 180.00 | 26.00 | HWRFe-386 - 2552 | 180.00 | 25.50 | HWRFe-387 - 2553 | 180.00 | 25.00 | HWRFe-388 - 2554 | 180.00 | 24.50 | HWRFe-389 - 2555 | 180.00 | 24.00 | HWRFe-390 - 2556 | 180.00 | 23.50 | HWRFe-391 - 2557 | 180.00 | 23.00 | HWRFe-392 - 2558 | 180.00 | 22.50 | HWRFe-393 - 2559 | 180.00 | 22.00 | HWRFe-394 - 2560 | 180.00 | 21.50 | HWRFe-395 - 2561 | 180.00 | 21.00 | HWRFe-396 - 2562 | 180.00 | 20.50 | HWRFe-397 - 2563 | 180.00 | 20.00 | HWRFe-398 - 2564 | 180.00 | 19.50 | HWRFe-399 - 2565 | 180.00 | 19.00 | HWRFe-400 - 2566 | 180.00 | 18.50 | HWRFe-401 - 2567 | 180.00 | 18.00 | HWRFe-402 - 2568 | 180.00 | 17.50 | HWRFe-403 - 2569 | 180.00 | 17.00 | HWRFe-404 - 2570 | 180.00 | 16.50 | HWRFe-405 - 2571 | 180.00 | 16.00 | HWRFe-406 - 2572 | 180.00 | 15.50 | HWRFe-407 - 2573 | 180.00 | 15.00 | HWRFe-408 - 2574 | 180.00 | 14.50 | HWRFe-409 - 2575 | 180.00 | 14.00 | HWRFe-410 - 2576 | 180.00 | 13.50 | HWRFe-411 - 2577 | 180.00 | 13.00 | HWRFe-412 - 2578 | 180.00 | 12.50 | HWRFe-413 - 2579 | 180.00 | 12.00 | HWRFe-414 - 2580 | 180.00 | 11.50 | HWRFe-415 - 2581 | 180.00 | 11.00 | HWRFe-416 - 2582 | 180.00 | 10.50 | HWRFe-417 - 2583 | 180.00 | 10.00 | HWRFe-418 - 2584 | 180.00 | 9.50 | HWRFe-419 - 2585 | 180.00 | 9.00 | HWRFe-420 - 2586 | 180.00 | 8.50 | HWRFe-421 - 2587 | 180.00 | 8.00 | HWRFe-422 - 2588 | 180.00 | 7.50 | HWRFe-423 - 2589 | 180.00 | 7.00 | HWRFe-424 - 2590 | 180.00 | 6.50 | HWRFe-425 - 2591 | 180.00 | 6.00 | HWRFe-426 - 2592 | 180.00 | 5.50 | HWRFe-427 - 2593 | 180.00 | 5.00 | HWRFe-428 - 2594 | 180.00 | 4.50 | HWRFe-429 - 2595 | 180.00 | 4.00 | HWRFe-430 - 2596 | 180.00 | 3.50 | HWRFe-431 - 2597 | 180.00 | 3.00 | HWRFe-432 - 2598 | 180.00 | 2.50 | HWRFe-433 - 2599 | 180.00 | 2.00 | HWRFe-434 - 2600 | 312.00 | 1.50 | HWRFl-50 - 2601 | 312.50 | 1.50 | HWRFl-51 - 2602 | 313.00 | 1.50 | HWRFl-52 - 2603 | 313.50 | 1.50 | HWRFl-53 - 2604 | 314.00 | 1.50 | HWRFl-54 - 2605 | 314.50 | 1.50 | HWRFl-55 - 2606 | 315.00 | 1.50 | HWRFl-56 - 2607 | 315.50 | 1.50 | HWRFl-57 - 2608 | 316.00 | 1.50 | HWRFl-58 - 2609 | 316.50 | 1.50 | HWRFl-59 - 2610 | 317.00 | 1.50 | HWRFl-60 - 2611 | 317.50 | 1.50 | HWRFl-61 - 2612 | 318.00 | 1.50 | HWRFl-62 - 2613 | 318.50 | 1.50 | HWRFl-63 - 2614 | 319.00 | 1.50 | HWRFl-64 - 2615 | 319.50 | 1.50 | HWRFl-65 - 2616 | 320.00 | 1.50 | HWRFl-66 - 2617 | 320.50 | 1.50 | HWRFl-67 - 2618 | 321.00 | 1.50 | HWRFl-68 - 2619 | 321.50 | 1.50 | HWRFl-69 - 2620 | 322.00 | 1.50 | HWRFl-70 - 2621 | 322.50 | 1.50 | HWRFl-71 - 2622 | 323.00 | 1.50 | HWRFl-72 - 2623 | 323.50 | 1.50 | HWRFl-73 - 2624 | 324.00 | 1.50 | HWRFl-74 - 2625 | 324.50 | 1.50 | HWRFl-75 - 2626 | 325.00 | 1.50 | HWRFl-76 - 2627 | 325.50 | 1.50 | HWRFl-77 - 2628 | 326.00 | 1.50 | HWRFl-78 - 2629 | 326.50 | 1.50 | HWRFl-79 - 2630 | 327.00 | 1.50 | HWRFl-80 - 2631 | 327.50 | 1.50 | HWRFl-81 - 2632 | 328.00 | 1.50 | HWRFl-82 - 2633 | 328.50 | 1.50 | HWRFl-83 - 2634 | 329.00 | 1.50 | HWRFl-84 - 2635 | 329.50 | 1.50 | HWRFl-85 - 2636 | 330.00 | 1.50 | HWRFl-86 - 2637 | 330.50 | 1.50 | HWRFl-87 - 2638 | 331.00 | 1.50 | HWRFl-88 - 2639 | 331.50 | 1.50 | HWRFl-89 - 2640 | 332.00 | 1.50 | HWRFl-90 - 2641 | 332.50 | 1.50 | HWRFl-91 - 2642 | 333.00 | 1.50 | HWRFl-92 - 2643 | 333.50 | 1.50 | HWRFl-93 - 2644 | 334.00 | 1.50 | HWRFl-94 - 2645 | 334.50 | 1.50 | HWRFl-95 - 2646 | 335.00 | 1.50 | HWRFl-96 - 2647 | 335.50 | 1.50 | HWRFl-97 - 2648 | 336.00 | 1.50 | HWRFl-98 - 2649 | 336.50 | 1.50 | HWRFl-99 - 2650 | 337.00 | 1.50 | HWRFl-100 - 2651 | 337.50 | 1.50 | HWRFl-101 - 2652 | 338.00 | 1.50 | HWRFl-102 - 2653 | 338.50 | 1.50 | HWRFl-103 - 2654 | 339.00 | 1.50 | HWRFl-104 - 2655 | 339.50 | 1.50 | HWRFl-105 - 2656 | 340.00 | 1.50 | HWRFl-106 - 2657 | 340.50 | 1.50 | HWRFl-107 - 2658 | 341.00 | 1.50 | HWRFl-108 - 2659 | 341.50 | 1.50 | HWRFl-109 - 2660 | 342.00 | 1.50 | HWRFl-110 - 2661 | 342.50 | 1.50 | HWRFl-111 - 2662 | 343.00 | 1.50 | HWRFl-112 - 2663 | 343.50 | 1.50 | HWRFl-113 - 2664 | 344.00 | 1.50 | HWRFl-114 - 2665 | 344.50 | 1.50 | HWRFl-115 - 2666 | 345.00 | 1.50 | HWRFl-116 - 2667 | 345.50 | 1.50 | HWRFl-117 - 2668 | 346.00 | 1.50 | HWRFl-118 - 2669 | 346.50 | 1.50 | HWRFl-119 - 2670 | 347.00 | 1.50 | HWRFl-120 - 2671 | 347.50 | 1.50 | HWRFl-121 - 2672 | 348.00 | 1.50 | HWRFl-122 - 2673 | 348.50 | 1.50 | HWRFl-123 - 2674 | 349.00 | 1.50 | HWRFl-124 - 2675 | 349.50 | 1.50 | HWRFl-125 - 2676 | 350.00 | 1.50 | HWRFl-126 - 2677 | 350.50 | 1.50 | HWRFl-127 - 2678 | 351.00 | 1.50 | HWRFl-128 - 2679 | 351.50 | 1.50 | HWRFl-129 - 2680 | 352.00 | 1.50 | HWRFl-130 - 2681 | 352.00 | 2.00 | HWRFl-131 - 2682 | 352.00 | 2.50 | HWRFl-132 - 2683 | 352.00 | 3.00 | HWRFl-133 - 2684 | 352.00 | 3.50 | HWRFl-134 - 2685 | 352.00 | 45.50 | HWRFl-135 - 2686 | 351.50 | 45.50 | HWRFl-136 - 2687 | 351.00 | 45.50 | HWRFl-137 - 2688 | 350.50 | 45.50 | HWRFl-138 - 2689 | 350.00 | 45.50 | HWRFl-139 - 2690 | 349.50 | 45.50 | HWRFl-140 - 2691 | 349.00 | 45.50 | HWRFl-141 - 2692 | 348.50 | 45.50 | HWRFl-142 - 2693 | 348.00 | 45.50 | HWRFl-143 - 2694 | 347.50 | 45.50 | HWRFl-144 - 2695 | 347.00 | 45.50 | HWRFl-145 - 2696 | 346.50 | 45.50 | HWRFl-146 - 2697 | 346.00 | 45.50 | HWRFl-147 - 2698 | 345.50 | 45.50 | HWRFl-148 - 2699 | 345.00 | 45.50 | HWRFl-149 - 2700 | 344.50 | 45.50 | HWRFl-150 - 2701 | 344.00 | 45.50 | HWRFl-151 - 2702 | 343.50 | 45.50 | HWRFl-152 - 2703 | 343.00 | 45.50 | HWRFl-153 - 2704 | 342.50 | 45.50 | HWRFl-154 - 2705 | 342.00 | 45.50 | HWRFl-155 - 2706 | 341.50 | 45.50 | HWRFl-156 - 2707 | 341.00 | 45.50 | HWRFl-157 - 2708 | 340.50 | 45.50 | HWRFl-158 - 2709 | 340.00 | 45.50 | HWRFl-159 - 2710 | 339.50 | 45.50 | HWRFl-160 - 2711 | 339.00 | 45.50 | HWRFl-161 - 2712 | 338.50 | 45.50 | HWRFl-162 - 2713 | 338.00 | 45.50 | HWRFl-163 - 2714 | 337.50 | 45.50 | HWRFl-164 - 2715 | 337.00 | 45.50 | HWRFl-165 - 2716 | 336.50 | 45.50 | HWRFl-166 - 2717 | 336.00 | 45.50 | HWRFl-167 - 2718 | 335.50 | 45.50 | HWRFl-168 - 2719 | 335.00 | 45.50 | HWRFl-169 - 2720 | 334.50 | 45.50 | HWRFl-170 - 2721 | 334.00 | 45.50 | HWRFl-171 - 2722 | 333.50 | 45.50 | HWRFl-172 - 2723 | 333.00 | 45.50 | HWRFl-173 - 2724 | 332.50 | 45.50 | HWRFl-174 - 2725 | 332.00 | 45.50 | HWRFl-175 - 2726 | 331.50 | 45.50 | HWRFl-176 - 2727 | 331.00 | 45.50 | HWRFl-177 - 2728 | 330.50 | 45.50 | HWRFl-178 - 2729 | 330.00 | 45.50 | HWRFl-179 - 2730 | 329.50 | 45.50 | HWRFl-180 - 2731 | 329.00 | 45.50 | HWRFl-181 - 2732 | 328.50 | 45.50 | HWRFl-182 - 2733 | 328.00 | 45.50 | HWRFl-183 - 2734 | 327.50 | 45.50 | HWRFl-184 - 2735 | 327.00 | 45.50 | HWRFl-185 - 2736 | 326.50 | 45.50 | HWRFl-186 - 2737 | 326.00 | 45.50 | HWRFl-187 - 2738 | 325.50 | 45.50 | HWRFl-188 - 2739 | 325.00 | 45.50 | HWRFl-189 - 2740 | 324.50 | 45.50 | HWRFl-190 - 2741 | 324.00 | 45.50 | HWRFl-191 - 2742 | 323.50 | 45.50 | HWRFl-192 - 2743 | 323.00 | 45.50 | HWRFl-193 - 2744 | 322.50 | 45.50 | HWRFl-194 - 2745 | 322.00 | 45.50 | HWRFl-195 - 2746 | 321.50 | 45.50 | HWRFl-196 - 2747 | 321.00 | 45.50 | HWRFl-197 - 2748 | 320.50 | 45.50 | HWRFl-198 - 2749 | 320.00 | 45.50 | HWRFl-199 - 2750 | 319.50 | 45.50 | HWRFl-200 - 2751 | 319.00 | 45.50 | HWRFl-201 - 2752 | 318.50 | 45.50 | HWRFl-202 - 2753 | 318.00 | 45.50 | HWRFl-203 - 2754 | 317.50 | 45.50 | HWRFl-204 - 2755 | 317.00 | 45.50 | HWRFl-205 - 2756 | 316.50 | 45.50 | HWRFl-206 - 2757 | 316.00 | 45.50 | HWRFl-207 - 2758 | 315.50 | 45.50 | HWRFl-208 - 2759 | 315.00 | 45.50 | HWRFl-209 - 2760 | 314.50 | 45.50 | HWRFl-210 - 2761 | 314.00 | 45.50 | HWRFl-211 - 2762 | 313.50 | 45.50 | HWRFl-212 - 2763 | 313.00 | 45.50 | HWRFl-213 - 2764 | 312.50 | 45.50 | HWRFl-214 - 2765 | 312.00 | 45.50 | HWRFl-215 - 2766 | 311.50 | 45.50 | HWRFl-216 - 2767 | 311.00 | 45.50 | HWRFl-217 - 2768 | 310.50 | 45.50 | HWRFl-218 - 2769 | 310.00 | 45.50 | HWRFl-219 - 2770 | 309.50 | 45.50 | HWRFl-220 - 2771 | 309.00 | 45.50 | HWRFl-221 - 2772 | 308.50 | 45.50 | HWRFl-222 - 2773 | 308.00 | 45.50 | HWRFl-223 - 2774 | 307.50 | 45.50 | HWRFl-224 - 2775 | 307.00 | 45.50 | HWRFl-225 - 2776 | 306.50 | 45.50 | HWRFl-226 - 2777 | 306.00 | 45.50 | HWRFl-227 - 2778 | 305.50 | 45.50 | HWRFl-228 - 2779 | 305.00 | 45.50 | HWRFl-229 - 2780 | 304.50 | 45.50 | HWRFl-230 - 2781 | 304.00 | 45.50 | HWRFl-231 - 2782 | 303.50 | 45.50 | HWRFl-232 - 2783 | 303.00 | 45.50 | HWRFl-233 - 2784 | 302.50 | 45.50 | HWRFl-234 - 2785 | 302.00 | 45.50 | HWRFl-235 - 2786 | 301.50 | 45.50 | HWRFl-236 - 2787 | 301.00 | 45.50 | HWRFl-237 - 2788 | 300.50 | 45.50 | HWRFl-238 - 2789 | 300.00 | 45.50 | HWRFl-239 - 2790 | 125.00 | -42.50 | HWRFp-50 - 2791 | 127.00 | -42.50 | HWRFp-51 - 2792 | 129.00 | -42.50 | HWRFp-52 - 2793 | 131.00 | -42.50 | HWRFp-53 - 2794 | 133.00 | -42.50 | HWRFp-54 - 2795 | 135.00 | -42.50 | HWRFp-55 - 2796 | 137.00 | -42.50 | HWRFp-56 - 2797 | 139.00 | -42.50 | HWRFp-57 - 2798 | 141.00 | -42.50 | HWRFp-58 - 2799 | 143.00 | -42.50 | HWRFp-59 - 2800 | 145.00 | -42.50 | HWRFp-60 - 2801 | 149.00 | -42.50 | HWRFp-62 - 2802 | 151.00 | -42.50 | HWRFp-63 - 2803 | 153.00 | -42.50 | HWRFp-64 - 2804 | 155.00 | -42.50 | HWRFp-65 - 2805 | 157.00 | -42.50 | HWRFp-66 - 2806 | 159.00 | -42.50 | HWRFp-67 - 2807 | 161.00 | -42.50 | HWRFp-68 - 2808 | 163.00 | -42.50 | HWRFp-69 - 2809 | 165.00 | -42.50 | HWRFp-70 - 2810 | 167.00 | -42.50 | HWRFp-71 - 2811 | 169.00 | -42.50 | HWRFp-72 - 2812 | 175.00 | -42.50 | HWRFp-73 - 2813 | 177.00 | -42.50 | HWRFp-74 - 2814 | 179.00 | -42.50 | HWRFp-75 - 2815 | 181.00 | -42.50 | HWRFp-76 - 2816 | 183.00 | -42.50 | HWRFp-77 - 2817 | 185.00 | -42.50 | HWRFp-78 - 2818 | 187.00 | -42.50 | HWRFp-79 - 2819 | 189.00 | -42.50 | HWRFp-80 - 2820 | 191.00 | -42.50 | HWRFp-81 - 2821 | 193.00 | -42.50 | HWRFp-82 - 2822 | 195.00 | -42.50 | HWRFp-83 - 2823 | 197.00 | -42.50 | HWRFp-84 - 2824 | 199.00 | -42.50 | HWRFp-85 - 2825 | 201.00 | -42.50 | HWRFp-86 - 2826 | 203.00 | -42.50 | HWRFp-87 - 2827 | 205.00 | -42.50 | HWRFp-88 - 2828 | 207.00 | -42.50 | HWRFp-89 - 2829 | 209.00 | -42.50 | HWRFp-90 - 2830 | 211.00 | -42.50 | HWRFp-91 - 2831 | 213.00 | -42.50 | HWRFp-92 - 2832 | 215.00 | -42.50 | HWRFp-93 - 2833 | 217.00 | -42.50 | HWRFp-94 - 2834 | 219.00 | -42.50 | HWRFp-95 - 2835 | 221.00 | -42.50 | HWRFp-96 - 2836 | 223.00 | -42.50 | HWRFp-97 - 2837 | 225.00 | -42.50 | HWRFp-98 - 2838 | 225.00 | -40.50 | HWRFp-99 - 2839 | 225.00 | -38.50 | HWRFp-100 - 2840 | 225.00 | -36.50 | HWRFp-101 - 2841 | 225.00 | -34.50 | HWRFp-102 - 2842 | 225.00 | -32.50 | HWRFp-103 - 2843 | 225.00 | -30.50 | HWRFp-104 - 2844 | 225.00 | -28.50 | HWRFp-105 - 2845 | 225.00 | -26.50 | HWRFp-106 - 2846 | 225.00 | -24.50 | HWRFp-107 - 2847 | 225.00 | -22.50 | HWRFp-108 - 2848 | 225.00 | -20.50 | HWRFp-109 - 2849 | 225.00 | -18.50 | HWRFp-110 - 2850 | 225.00 | -16.50 | HWRFp-111 - 2851 | 225.00 | -14.50 | HWRFp-112 - 2852 | 225.00 | -12.50 | HWRFp-113 - 2853 | 225.00 | -10.50 | HWRFp-114 - 2854 | 225.00 | -8.50 | HWRFp-115 - 2855 | 225.00 | -6.50 | HWRFp-116 - 2856 | 225.00 | -4.50 | HWRFp-117 - 2857 | 225.00 | -2.50 | HWRFp-118 - 2858 | 223.00 | -2.50 | HWRFp-119 - 2859 | 221.00 | -2.50 | HWRFp-120 - 2860 | 219.00 | -2.50 | HWRFp-121 - 2861 | 217.00 | -2.50 | HWRFp-122 - 2862 | 215.00 | -2.50 | HWRFp-123 - 2863 | 213.00 | -2.50 | HWRFp-124 - 2864 | 211.00 | -2.50 | HWRFp-125 - 2865 | 209.00 | -2.50 | HWRFp-126 - 2866 | 207.00 | -2.50 | HWRFp-127 - 2867 | 205.00 | -2.50 | HWRFp-128 - 2868 | 203.00 | -2.50 | HWRFp-129 - 2869 | 201.00 | -2.50 | HWRFp-130 - 2870 | 199.00 | -2.50 | HWRFp-131 - 2871 | 197.00 | -2.50 | HWRFp-132 - 2872 | 195.00 | -2.50 | HWRFp-133 - 2873 | 193.00 | -2.50 | HWRFp-134 - 2874 | 191.00 | -2.50 | HWRFp-135 - 2875 | 189.00 | -2.50 | HWRFp-136 - 2876 | 187.00 | -2.50 | HWRFp-137 - 2877 | 185.00 | -2.50 | HWRFp-138 - 2878 | 183.00 | -2.50 | HWRFp-139 - 2879 | 181.00 | -2.50 | HWRFp-140 - 2880 | 179.00 | -2.50 | HWRFp-141 - 2881 | 177.00 | -2.50 | HWRFp-142 - 2882 | 175.00 | -2.50 | HWRFp-143 - 2883 | 173.00 | -2.50 | HWRFp-144 - 2884 | 171.00 | -2.50 | HWRFp-145 - 2885 | 169.00 | -2.50 | HWRFp-146 - 2886 | 167.00 | -2.50 | HWRFp-147 - 2887 | 165.00 | -2.50 | HWRFp-148 - 2888 | 163.00 | -2.50 | HWRFp-149 - 2889 | 161.00 | -2.50 | HWRFp-150 - 2890 | 159.00 | -2.50 | HWRFp-151 - 2891 | 157.00 | -2.50 | HWRFp-152 - 2892 | 155.00 | -2.50 | HWRFp-153 - 2893 | 153.00 | -2.50 | HWRFp-154 - 2894 | 151.00 | -2.50 | HWRFp-155 - 2895 | 149.00 | -2.50 | HWRFp-156 - 2896 | 147.00 | -2.50 | HWRFp-157 - 2897 | 145.00 | -2.50 | HWRFp-158 - 2898 | 131.00 | -2.50 | HWRFp-159 - 2899 | 127.00 | -2.50 | HWRFp-160 - 2900 | 125.00 | -2.50 | HWRFp-161 - 2901 | 125.00 | -4.50 | HWRFp-162 - 2902 | 125.00 | -6.50 | HWRFp-163 - 2903 | 125.00 | -10.50 | HWRFp-164 - 2904 | 125.00 | -12.50 | HWRFp-165 - 2905 | 125.00 | -34.50 | HWRFp-166 - 2906 | 125.00 | -36.50 | HWRFp-167 - 2907 | 125.00 | -38.50 | HWRFp-168 - 2908 | 125.00 | -40.50 | HWRFp-169 - 2909 | 31.00 | -42.50 | HWRFs-50 - 2910 | 33.00 | -42.50 | HWRFs-51 - 2911 | 35.00 | -42.50 | HWRFs-52 - 2912 | 37.00 | -42.50 | HWRFs-53 - 2913 | 39.00 | -42.50 | HWRFs-54 - 2914 | 41.00 | -42.50 | HWRFs-55 - 2915 | 43.00 | -42.50 | HWRFs-56 - 2916 | 45.00 | -42.50 | HWRFs-57 - 2917 | 47.00 | -42.50 | HWRFs-58 - 2918 | 49.00 | -42.50 | HWRFs-59 - 2919 | 51.00 | -42.50 | HWRFs-60 - 2920 | 53.00 | -42.50 | HWRFs-61 - 2921 | 55.00 | -42.50 | HWRFs-62 - 2922 | 57.00 | -42.50 | HWRFs-63 - 2923 | 59.00 | -42.50 | HWRFs-64 - 2924 | 61.00 | -42.50 | HWRFs-65 - 2925 | 63.00 | -42.50 | HWRFs-66 - 2926 | 65.00 | -42.50 | HWRFs-67 - 2927 | 67.00 | -42.50 | HWRFs-68 - 2928 | 69.00 | -42.50 | HWRFs-69 - 2929 | 71.00 | -42.50 | HWRFs-70 - 2930 | 73.00 | -42.50 | HWRFs-71 - 2931 | 75.00 | -42.50 | HWRFs-72 - 2932 | 77.00 | -42.50 | HWRFs-73 - 2933 | 79.00 | -42.50 | HWRFs-74 - 2934 | 81.00 | -42.50 | HWRFs-75 - 2935 | 83.00 | -42.50 | HWRFs-76 - 2936 | 85.00 | -42.50 | HWRFs-77 - 2937 | 87.00 | -42.50 | HWRFs-78 - 2938 | 89.00 | -42.50 | HWRFs-79 - 2939 | 91.00 | -42.50 | HWRFs-80 - 2940 | 93.00 | -42.50 | HWRFs-81 - 2941 | 95.00 | -42.50 | HWRFs-82 - 2942 | 97.00 | -42.50 | HWRFs-83 - 2943 | 99.00 | -42.50 | HWRFs-84 - 2944 | 101.00 | -42.50 | HWRFs-85 - 2945 | 103.00 | -42.50 | HWRFs-86 - 2946 | 105.00 | -42.50 | HWRFs-87 - 2947 | 107.00 | -42.50 | HWRFs-88 - 2948 | 109.00 | -42.50 | HWRFs-89 - 2949 | 111.00 | -42.50 | HWRFs-90 - 2950 | 113.00 | -42.50 | HWRFs-91 - 2951 | 115.00 | -42.50 | HWRFs-92 - 2952 | 117.00 | -42.50 | HWRFs-93 - 2953 | 119.00 | -42.50 | HWRFs-94 - 2954 | 121.00 | -42.50 | HWRFs-95 - 2955 | 123.00 | -42.50 | HWRFs-96 - 2956 | 125.00 | -42.50 | HWRFs-97 - 2957 | 127.00 | -42.50 | HWRFs-98 - 2958 | 128.00 | -42.50 | HWRFs-99 - 2959 | 128.00 | -40.50 | HWRFs-100 - 2960 | 128.00 | -38.50 | HWRFs-101 - 2961 | 128.00 | -36.50 | HWRFs-102 - 2962 | 128.00 | -34.50 | HWRFs-103 - 2963 | 128.00 | -12.50 | HWRFs-104 - 2964 | 128.00 | -10.50 | HWRFs-105 - 2965 | 128.00 | -8.50 | HWRFs-106 - 2966 | 128.00 | -6.50 | HWRFs-107 - 2967 | 128.00 | -4.50 | HWRFs-108 - 2968 | 127.00 | -2.50 | HWRFs-110 - 2969 | 125.00 | -2.50 | HWRFs-111 - 2970 | 123.00 | -2.50 | HWRFs-112 - 2971 | 117.00 | -2.50 | HWRFs-113 - 2972 | 109.00 | -2.50 | HWRFs-114 - 2973 | 99.00 | -2.50 | HWRFs-117 - 2974 | 97.00 | -2.50 | HWRFs-118 - 2975 | 95.00 | -2.50 | HWRFs-119 - 2976 | 93.00 | -2.50 | HWRFs-120 - 2977 | 91.00 | -2.50 | HWRFs-121 - 2978 | 89.00 | -2.50 | HWRFs-122 - 2979 | 87.00 | -2.50 | HWRFs-123 - 2980 | 85.00 | -2.50 | HWRFs-124 - 2981 | 83.00 | -2.50 | HWRFs-125 - 2982 | 81.00 | -2.50 | HWRFs-126 - 2983 | 79.00 | -2.50 | HWRFs-127 - 2984 | 77.00 | -2.50 | HWRFs-128 - 2985 | 75.00 | -2.50 | HWRFs-129 - 2986 | 73.00 | -2.50 | HWRFs-130 - 2987 | 71.00 | -2.50 | HWRFs-131 - 2988 | 69.00 | -2.50 | HWRFs-132 - 2989 | 67.00 | -2.50 | HWRFs-133 - 2990 | 65.00 | -2.50 | HWRFs-134 - 2991 | 63.00 | -2.50 | HWRFs-135 - 2992 | 61.00 | -2.50 | HWRFs-136 - 2993 | 59.00 | -2.50 | HWRFs-137 - 2994 | 57.00 | -2.50 | HWRFs-138 - 2995 | 55.00 | -2.50 | HWRFs-139 - 2996 | 53.00 | -2.50 | HWRFs-140 - 2997 | 51.00 | -2.50 | HWRFs-141 - 2998 | 49.00 | -2.50 | HWRFs-142 - 2999 | 47.00 | -2.50 | HWRFs-143 - 3000 | 45.00 | -2.50 | HWRFs-144 - 3001 | 43.00 | -2.50 | HWRFs-145 - 3002 | 31.00 | -32.50 | HWRFs-146 - 3003 | 31.00 | -34.50 | HWRFs-147 - 3004 | 31.00 | -36.50 | HWRFs-148 - 3005 | 31.00 | -38.50 | HWRFs-149 - 3006 | 31.00 | -40.50 | HWRFs-150 - 3007 | 105.00 | 1.50 | HWRFw-50 - 3008 | 107.00 | 1.50 | HWRFw-51 - 3009 | 123.00 | 1.50 | HWRFw-52 - 3010 | 131.00 | 1.50 | HWRFw-54 - 3011 | 133.00 | 1.50 | HWRFw-55 - 3012 | 135.00 | 1.50 | HWRFw-56 - 3013 | 137.00 | 1.50 | HWRFw-57 - 3014 | 139.00 | 1.50 | HWRFw-58 - 3015 | 141.00 | 1.50 | HWRFw-59 - 3016 | 143.00 | 1.50 | HWRFw-60 - 3017 | 145.00 | 1.50 | HWRFw-61 - 3018 | 147.00 | 1.50 | HWRFw-62 - 3019 | 149.00 | 1.50 | HWRFw-63 - 3020 | 151.00 | 1.50 | HWRFw-64 - 3021 | 153.00 | 1.50 | HWRFw-65 - 3022 | 155.00 | 1.50 | HWRFw-66 - 3023 | 157.00 | 1.50 | HWRFw-67 - 3024 | 159.00 | 1.50 | HWRFw-68 - 3025 | 161.00 | 1.50 | HWRFw-69 - 3026 | 163.00 | 1.50 | HWRFw-70 - 3027 | 165.00 | 1.50 | HWRFw-71 - 3028 | 167.00 | 1.50 | HWRFw-72 - 3029 | 169.00 | 1.50 | HWRFw-73 - 3030 | 171.00 | 1.50 | HWRFw-74 - 3031 | 173.00 | 1.50 | HWRFw-75 - 3032 | 175.00 | 1.50 | HWRFw-76 - 3033 | 177.00 | 1.50 | HWRFw-77 - 3034 | 179.00 | 1.50 | HWRFw-78 - 3035 | 180.00 | 1.50 | HWRFw-79 - 3036 | 180.00 | 3.50 | HWRFw-80 - 3037 | 180.00 | 5.50 | HWRFw-81 - 3038 | 180.00 | 7.50 | HWRFw-82 - 3039 | 180.00 | 9.50 | HWRFw-83 - 3040 | 180.00 | 11.50 | HWRFw-84 - 3041 | 180.00 | 13.50 | HWRFw-85 - 3042 | 180.00 | 15.50 | HWRFw-86 - 3043 | 180.00 | 17.50 | HWRFw-87 - 3044 | 180.00 | 19.50 | HWRFw-88 - 3045 | 180.00 | 21.50 | HWRFw-89 - 3046 | 180.00 | 23.50 | HWRFw-90 - 3047 | 180.00 | 25.50 | HWRFw-91 - 3048 | 180.00 | 27.50 | HWRFw-92 - 3049 | 180.00 | 29.50 | HWRFw-93 - 3050 | 180.00 | 31.50 | HWRFw-94 - 3051 | 180.00 | 33.50 | HWRFw-95 - 3052 | 180.00 | 35.50 | HWRFw-96 - 3053 | 180.00 | 37.50 | HWRFw-97 - 3054 | 180.00 | 39.50 | HWRFw-98 - 3055 | 180.00 | 41.50 | HWRFw-99 - 3056 | 179.00 | 41.50 | HWRFw-100 - 3057 | 177.00 | 41.50 | HWRFw-101 - 3058 | 175.00 | 41.50 | HWRFw-102 - 3059 | 173.00 | 41.50 | HWRFw-103 - 3060 | 171.00 | 41.50 | HWRFw-104 - 3061 | 169.00 | 41.50 | HWRFw-105 - 3062 | 167.00 | 41.50 | HWRFw-106 - 3063 | 165.00 | 41.50 | HWRFw-107 - 3064 | 163.00 | 41.50 | HWRFw-108 - 3065 | 161.00 | 41.50 | HWRFw-109 - 3066 | 159.00 | 41.50 | HWRFw-110 - 3067 | 157.00 | 41.50 | HWRFw-111 - 3068 | 155.00 | 41.50 | HWRFw-112 - 3069 | 153.00 | 41.50 | HWRFw-113 - 3070 | 151.00 | 41.50 | HWRFw-114 - 3071 | 149.00 | 41.50 | HWRFw-115 - 3072 | 147.00 | 41.50 | HWRFw-116 - 3073 | 145.00 | 41.50 | HWRFw-117 - 3074 | 143.00 | 41.50 | HWRFw-118 - 3075 | 139.00 | 41.50 | HWRFw-119 - 3076 | 137.00 | 41.50 | HWRFw-120 - 3077 | 135.00 | 41.50 | HWRFw-121 - 3078 | 133.00 | 41.50 | HWRFw-122 - 3079 | 131.00 | 41.50 | HWRFw-123 - 3080 | -98.00 | 3.00 | RW-NH2-51 - 3081 | -98.00 | 3.50 | RW-NH2-52 - 3082 | -98.00 | 4.00 | RW-NH2-53 - 3083 | -98.00 | 4.50 | RW-NH2-54 - 3084 | -98.00 | 5.00 | RW-NH2-55 - 3085 | -98.00 | 5.50 | RW-NH2-56 - 3086 | -98.00 | 6.00 | RW-NH2-57 - 3087 | -98.00 | 6.50 | RW-NH2-58 - 3088 | -98.00 | 7.00 | RW-NH2-59 - 3089 | -98.00 | 7.50 | RW-NH2-60 - 3090 | -98.00 | 8.00 | RW-NH2-61 - 3091 | -98.00 | 8.50 | RW-NH2-62 - 3092 | -98.00 | 9.00 | RW-NH2-63 - 3093 | -98.00 | 9.50 | RW-NH2-64 - 3094 | -98.00 | 10.00 | RW-NH2-65 - 3095 | -98.00 | 10.50 | RW-NH2-66 - 3096 | -98.00 | 11.00 | RW-NH2-67 - 3097 | -98.00 | 11.50 | RW-NH2-68 - 3098 | -98.00 | 12.00 | RW-NH2-69 - 3099 | -98.00 | 12.50 | RW-NH2-70 - 3100 | -98.00 | 13.00 | RW-NH2-71 - 3101 | -98.00 | 13.50 | RW-NH2-72 - 3102 | -98.00 | 14.00 | RW-NH2-73 - 3103 | -98.00 | 14.50 | RW-NH2-74 - 3104 | -98.00 | 15.00 | RW-NH2-75 - 3105 | -80.50 | 32.00 | RW-NH2-79 - 3106 | -80.00 | 32.00 | RW-NH2-80 - 3107 | -79.50 | 32.00 | RW-NH2-81 - 3108 | -79.00 | 32.00 | RW-NH2-82 - 3109 | -78.50 | 32.00 | RW-NH2-83 - 3110 | -78.00 | 32.00 | RW-NH2-84 - 3111 | -77.50 | 32.00 | RW-NH2-85 - 3112 | -77.00 | 32.00 | RW-NH2-86 - 3113 | -76.50 | 32.00 | RW-NH2-87 - 3114 | -76.00 | 32.00 | RW-NH2-88 - 3115 | -75.50 | 32.00 | RW-NH2-89 - 3116 | -75.00 | 32.00 | RW-NH2-90 - 3117 | -74.50 | 32.00 | RW-NH2-91 - 3118 | -74.00 | 32.00 | RW-NH2-92 - 3119 | -73.50 | 32.00 | RW-NH2-93 - 3120 | -73.00 | 32.00 | RW-NH2-94 - 3121 | -72.50 | 32.00 | RW-NH2-95 - 3122 | -72.00 | 32.00 | RW-NH2-96 - 3123 | -71.50 | 32.00 | RW-NH2-97 - 3124 | -71.00 | 32.00 | RW-NH2-98 - 3125 | -70.50 | 32.00 | RW-NH2-99 - 3126 | -70.00 | 32.00 | RW-NH2-100 - 3127 | -69.50 | 32.00 | RW-NH2-101 - 3128 | -69.00 | 32.00 | RW-NH2-102 - 3129 | -68.50 | 32.00 | RW-NH2-103 - 3130 | -68.00 | 32.00 | RW-NH2-104 - 3131 | -67.50 | 32.00 | RW-NH2-105 - 3132 | -67.00 | 32.00 | RW-NH2-106 - 3133 | -66.50 | 32.00 | RW-NH2-107 - 3134 | -66.00 | 32.00 | RW-NH2-108 - 3135 | -65.50 | 32.00 | RW-NH2-109 - 3136 | -65.00 | 32.00 | RW-NH2-110 - 3137 | -64.50 | 32.00 | RW-NH2-111 - 3138 | -64.00 | 32.00 | RW-NH2-112 - 3139 | -63.50 | 32.00 | RW-NH2-113 - 3140 | -63.00 | 32.00 | RW-NH2-114 - 3141 | -62.50 | 32.00 | RW-NH2-115 - 3142 | -62.00 | 32.00 | RW-NH2-116 - 3143 | -61.50 | 32.00 | RW-NH2-117 - 3144 | -61.00 | 32.00 | RW-NH2-118 - 3145 | -60.50 | 32.00 | RW-NH2-119 - 3146 | -60.00 | 32.00 | RW-NH2-120 - 3147 | -59.50 | 32.00 | RW-NH2-121 - 3148 | -59.00 | 32.00 | RW-NH2-122 - 3149 | -58.50 | 32.00 | RW-NH2-123 - 3150 | -58.00 | 32.00 | RW-NH2-124 - 3151 | -57.50 | 32.00 | RW-NH2-125 - 3152 | -57.00 | 32.00 | RW-NH2-126 - 3153 | -56.50 | 32.00 | RW-NH2-127 - 3154 | -56.00 | 32.00 | RW-NH2-128 - 3155 | -55.50 | 32.00 | RW-NH2-129 - 3156 | -55.00 | 32.00 | RW-NH2-130 - 3157 | -54.50 | 32.00 | RW-NH2-131 - 3158 | -54.00 | 32.00 | RW-NH2-132 - 3159 | -53.50 | 32.00 | RW-NH2-133 - 3160 | -53.00 | 32.00 | RW-NH2-134 - 3161 | -52.50 | 32.00 | RW-NH2-135 - 3162 | -52.00 | 32.00 | RW-NH2-136 - 3163 | -51.50 | 32.00 | RW-NH2-137 - 3164 | -51.00 | 32.00 | RW-NH2-138 - 3165 | -50.50 | 32.00 | RW-NH2-139 - 3166 | -50.00 | 32.00 | RW-NH2-140 - 3167 | -49.50 | 32.00 | RW-NH2-141 - 3168 | -49.00 | 32.00 | RW-NH2-142 - 3169 | -48.50 | 32.00 | RW-NH2-143 - 3170 | -48.00 | 32.00 | RW-NH2-144 - 3171 | -47.50 | 32.00 | RW-NH2-145 - 3172 | -47.00 | 32.00 | RW-NH2-146 - 3173 | -46.50 | 32.00 | RW-NH2-147 - 3174 | -46.00 | 32.00 | RW-NH2-148 - 3175 | -45.50 | 32.00 | RW-NH2-149 - 3176 | -45.00 | 32.00 | RW-NH2-150 - 3177 | -44.50 | 32.00 | RW-NH2-151 - 3178 | -44.00 | 32.00 | RW-NH2-152 - 3179 | -43.50 | 32.00 | RW-NH2-153 - 3180 | -43.00 | 32.00 | RW-NH2-154 - 3181 | -42.50 | 32.00 | RW-NH2-155 - 3182 | -42.00 | 32.00 | RW-NH2-156 - 3183 | -41.50 | 32.00 | RW-NH2-157 - 3184 | -41.00 | 32.00 | RW-NH2-158 - 3185 | -40.50 | 32.00 | RW-NH2-159 - 3186 | -40.00 | 32.00 | RW-NH2-160 - 3187 | -39.50 | 32.00 | RW-NH2-161 - 3188 | -39.00 | 32.00 | RW-NH2-162 - 3189 | -38.50 | 32.00 | RW-NH2-163 - 3190 | -38.00 | 32.00 | RW-NH2-164 - 3191 | -37.50 | 32.00 | RW-NH2-165 - 3192 | -37.00 | 32.00 | RW-NH2-166 - 3193 | -36.50 | 32.00 | RW-NH2-167 - 3194 | -36.00 | 32.00 | RW-NH2-168 - 3195 | -35.50 | 32.00 | RW-NH2-169 - 3196 | -35.00 | 32.00 | RW-NH2-170 - 3197 | -34.50 | 32.00 | RW-NH2-171 - 3198 | -34.00 | 32.00 | RW-NH2-172 - 3199 | -33.50 | 32.00 | RW-NH2-173 - 3200 | -33.00 | 32.00 | RW-NH2-174 - 3201 | -32.50 | 32.00 | RW-NH2-175 - 3202 | -32.00 | 32.00 | RW-NH2-176 - 3203 | -31.50 | 32.00 | RW-NH2-177 - 3204 | -31.00 | 32.00 | RW-NH2-178 - 3205 | -30.50 | 32.00 | RW-NH2-179 - 3206 | -30.00 | 32.00 | RW-NH2-180 - 3207 | -29.50 | 32.00 | RW-NH2-181 - 3208 | -29.00 | 32.00 | RW-NH2-182 - 3209 | -28.50 | 32.00 | RW-NH2-183 - 3210 | -28.00 | 32.00 | RW-NH2-184 - 3211 | -27.50 | 32.00 | RW-NH2-185 - 3212 | -27.00 | 32.00 | RW-NH2-186 - 3213 | -26.50 | 32.00 | RW-NH2-187 - 3214 | -26.00 | 32.00 | RW-NH2-188 - 3215 | -25.50 | 32.00 | RW-NH2-189 - 3216 | -25.00 | 32.00 | RW-NH2-190 - 3217 | -24.50 | 32.00 | RW-NH2-191 - 3218 | -24.00 | 32.00 | RW-NH2-192 - 3219 | -23.50 | 32.00 | RW-NH2-193 - 3220 | -23.00 | 32.00 | RW-NH2-194 - 3221 | -22.50 | 32.00 | RW-NH2-195 - 3222 | -22.00 | 32.00 | RW-NH2-196 - 3223 | -21.50 | 32.00 | RW-NH2-197 - 3224 | -21.00 | 32.00 | RW-NH2-198 - 3225 | -20.50 | 32.00 | RW-NH2-199 - 3226 | -20.00 | 32.00 | RW-NH2-200 - 3227 | -19.50 | 32.00 | RW-NH2-201 - 3228 | -19.00 | 32.00 | RW-NH2-202 - 3229 | -18.50 | 32.00 | RW-NH2-203 - 3230 | -18.00 | 32.00 | RW-NH2-204 - 3231 | -17.50 | 32.00 | RW-NH2-205 - 3232 | -17.00 | 32.00 | RW-NH2-206 - 3233 | -16.50 | 32.00 | RW-NH2-207 - 3234 | -16.00 | 32.00 | RW-NH2-208 - 3235 | -15.50 | 32.00 | RW-NH2-209 - 3236 | -15.00 | 32.00 | RW-NH2-210 - 3237 | -14.50 | 32.00 | RW-NH2-211 - 3238 | -14.00 | 32.00 | RW-NH2-212 - 3239 | -13.50 | 32.00 | RW-NH2-213 - 3240 | -13.00 | 32.00 | RW-NH2-214 - 3241 | -12.50 | 32.00 | RW-NH2-215 - 3242 | -12.00 | 32.00 | RW-NH2-216 - 3243 | -11.50 | 32.00 | RW-NH2-217 - 3244 | -11.00 | 32.00 | RW-NH2-218 - 3245 | -10.50 | 32.00 | RW-NH2-219 - 3246 | -10.00 | 4.50 | RW-NH2-224 - 3247 | -10.00 | 4.00 | RW-NH2-225 - 3248 | -10.00 | 3.50 | RW-NH2-226 - 3249 | -10.00 | 3.00 | RW-NH2-227 - 3250 | -10.00 | 3.00 | RW-NH2-228 - 3251 | -10.50 | 3.00 | RW-NH2-229 - 3252 | -11.00 | 3.00 | RW-NH2-230 - 3253 | -11.50 | 3.00 | RW-NH2-231 - 3254 | -12.00 | 3.00 | RW-NH2-232 - 3255 | -12.50 | 3.00 | RW-NH2-233 - 3256 | -13.00 | 3.00 | RW-NH2-234 - 3257 | -13.50 | 3.00 | RW-NH2-235 - 3258 | -14.00 | 3.00 | RW-NH2-236 - 3259 | -14.50 | 3.00 | RW-NH2-237 - 3260 | -15.00 | 3.00 | RW-NH2-238 - 3261 | -15.50 | 3.00 | RW-NH2-239 - 3262 | -16.00 | 3.00 | RW-NH2-240 - 3263 | -16.50 | 3.00 | RW-NH2-241 - 3264 | -17.00 | 3.00 | RW-NH2-242 - 3265 | -17.50 | 3.00 | RW-NH2-243 - 3266 | -18.00 | 3.00 | RW-NH2-244 - 3267 | -18.50 | 3.00 | RW-NH2-245 - 3268 | -19.00 | 3.00 | RW-NH2-246 - 3269 | -19.50 | 3.00 | RW-NH2-247 - 3270 | -20.00 | 3.00 | RW-NH2-248 - 3271 | -20.50 | 3.00 | RW-NH2-249 - 3272 | -21.00 | 3.00 | RW-NH2-250 - 3273 | -21.50 | 3.00 | RW-NH2-251 - 3274 | -22.00 | 3.00 | RW-NH2-252 - 3275 | -22.50 | 3.00 | RW-NH2-253 - 3276 | -23.00 | 3.00 | RW-NH2-254 - 3277 | -23.50 | 3.00 | RW-NH2-255 - 3278 | -24.00 | 3.00 | RW-NH2-256 - 3279 | -24.50 | 3.00 | RW-NH2-257 - 3280 | -25.00 | 3.00 | RW-NH2-258 - 3281 | -25.50 | 3.00 | RW-NH2-259 - 3282 | -26.00 | 3.00 | RW-NH2-260 - 3283 | -26.50 | 3.00 | RW-NH2-261 - 3284 | -27.00 | 3.00 | RW-NH2-262 - 3285 | -27.50 | 3.00 | RW-NH2-263 - 3286 | -28.00 | 3.00 | RW-NH2-264 - 3287 | -28.50 | 3.00 | RW-NH2-265 - 3288 | -29.00 | 3.00 | RW-NH2-266 - 3289 | -29.50 | 3.00 | RW-NH2-267 - 3290 | -30.00 | 3.00 | RW-NH2-268 - 3291 | -30.50 | 3.00 | RW-NH2-269 - 3292 | -31.00 | 3.00 | RW-NH2-270 - 3293 | -31.50 | 3.00 | RW-NH2-271 - 3294 | -32.00 | 3.00 | RW-NH2-272 - 3295 | -32.50 | 3.00 | RW-NH2-273 - 3296 | -33.00 | 3.00 | RW-NH2-274 - 3297 | -33.50 | 3.00 | RW-NH2-275 - 3298 | -34.00 | 3.00 | RW-NH2-276 - 3299 | -34.50 | 3.00 | RW-NH2-277 - 3300 | -35.00 | 3.00 | RW-NH2-278 - 3301 | -35.50 | 3.00 | RW-NH2-279 - 3302 | -36.00 | 3.00 | RW-NH2-280 - 3303 | -36.50 | 3.00 | RW-NH2-281 - 3304 | -37.00 | 3.00 | RW-NH2-282 - 3305 | -37.50 | 3.00 | RW-NH2-283 - 3306 | -38.00 | 3.00 | RW-NH2-284 - 3307 | -38.50 | 3.00 | RW-NH2-285 - 3308 | -39.00 | 3.00 | RW-NH2-286 - 3309 | -39.50 | 3.00 | RW-NH2-287 - 3310 | -40.00 | 3.00 | RW-NH2-288 - 3311 | -40.50 | 3.00 | RW-NH2-289 - 3312 | -41.00 | 3.00 | RW-NH2-290 - 3313 | -41.50 | 3.00 | RW-NH2-291 - 3314 | -42.00 | 3.00 | RW-NH2-292 - 3315 | -42.50 | 3.00 | RW-NH2-293 - 3316 | -43.00 | 3.00 | RW-NH2-294 - 3317 | -43.50 | 3.00 | RW-NH2-295 - 3318 | -44.00 | 3.00 | RW-NH2-296 - 3319 | -44.50 | 3.00 | RW-NH2-297 - 3320 | -42.00 | 3.00 | RW-NH2-298 - 3321 | -42.50 | 3.00 | RW-NH2-299 - 3322 | -43.00 | 3.00 | RW-NH2-300 - 3323 | -43.50 | 3.00 | RW-NH2-301 - 3324 | -44.00 | 3.00 | RW-NH2-302 - 3325 | -44.50 | 3.00 | RW-NH2-303 - 3326 | -45.00 | 3.00 | RW-NH2-304 - 3327 | -42.50 | 3.00 | RW-NH2-305 - 3328 | -43.00 | 3.00 | RW-NH2-306 - 3329 | -43.50 | 3.00 | RW-NH2-307 - 3330 | -44.00 | 3.00 | RW-NH2-308 - 3331 | -44.50 | 3.00 | RW-NH2-309 - 3332 | -45.00 | 3.00 | RW-NH2-310 - 3333 | -45.50 | 3.00 | RW-NH2-311 - 3334 | -46.00 | 3.00 | RW-NH2-312 - 3335 | -46.50 | 3.00 | RW-NH2-313 - 3336 | -47.00 | 3.00 | RW-NH2-314 - 3337 | -47.50 | 3.00 | RW-NH2-315 - 3338 | -48.00 | 3.00 | RW-NH2-316 - 3339 | -45.50 | 3.00 | RW-NH2-317 - 3340 | -46.00 | 3.00 | RW-NH2-318 - 3341 | -46.50 | 3.00 | RW-NH2-319 - 3342 | -47.00 | 3.00 | RW-NH2-320 - 3343 | -47.50 | 3.00 | RW-NH2-321 - 3344 | -48.00 | 3.00 | RW-NH2-322 - 3345 | -48.50 | 3.00 | RW-NH2-323 - 3346 | -49.00 | 3.00 | RW-NH2-324 - 3347 | -49.50 | 3.00 | RW-NH2-325 - 3348 | -50.00 | 3.00 | RW-NH2-326 - 3349 | -50.50 | 3.00 | RW-NH2-327 - 3350 | -78.00 | 3.00 | RW-NH2-330 - 3351 | -78.50 | 3.00 | RW-NH2-331 - 3352 | -79.00 | 3.00 | RW-NH2-332 - 3353 | -79.50 | 3.00 | RW-NH2-333 - 3354 | -80.00 | 3.00 | RW-NH2-334 - 3355 | -80.50 | 3.00 | RW-NH2-335 - 3356 | -81.00 | 3.00 | RW-NH2-336 - 3357 | -81.50 | 3.00 | RW-NH2-337 - 3358 | -82.00 | 3.00 | RW-NH2-338 - 3359 | -82.50 | 3.00 | RW-NH2-339 - 3360 | -83.00 | 3.00 | RW-NH2-340 - 3361 | -83.50 | 3.00 | RW-NH2-341 - 3362 | -84.00 | 3.00 | RW-NH2-342 - 3363 | -84.50 | 3.00 | RW-NH2-343 - 3364 | -85.00 | 3.00 | RW-NH2-344 - 3365 | -85.50 | 3.00 | RW-NH2-345 - 3366 | -86.00 | 3.00 | RW-NH2-346 - 3367 | -86.50 | 3.00 | RW-NH2-347 - 3368 | -87.00 | 3.00 | RW-NH2-348 - 3369 | -87.50 | 3.00 | RW-NH2-349 - 3370 | -88.00 | 3.00 | RW-NH2-350 - 3371 | -88.50 | 3.00 | RW-NH2-351 - 3372 | -89.00 | 3.00 | RW-NH2-352 - 3373 | -89.50 | 3.00 | RW-NH2-353 - 3374 | -90.00 | 3.00 | RW-NH2-354 - 3375 | -90.50 | 3.00 | RW-NH2-355 - 3376 | -91.00 | 3.00 | RW-NH2-356 - 3377 | -91.50 | 3.00 | RW-NH2-357 - 3378 | -92.00 | 3.00 | RW-NH2-358 - 3379 | -92.50 | 3.00 | RW-NH2-359 - 3380 | -93.00 | 3.00 | RW-NH2-360 - 3381 | -93.50 | 3.00 | RW-NH2-361 - 3382 | -94.00 | 3.00 | RW-NH2-362 - 3383 | -94.50 | 3.00 | RW-NH2-363 - 3384 | -95.00 | 3.00 | RW-NH2-364 - 3385 | -95.50 | 3.00 | RW-NH2-365 - 3386 | -96.00 | 3.00 | RW-NH2-366 - 3387 | -96.50 | 3.00 | RW-NH2-367 - 3388 | -97.00 | 3.00 | RW-NH2-368 - 3389 | -97.50 | 3.00 | RW-NH2-369 - 3390 | -98.00 | 3.00 | RW-NH2-370 - 3391 | -30.00 | 3.00 | RW-NH2-371 - 3392 | -30.00 | 3.50 | RW-NH2-372 - 3393 | -30.00 | 4.00 | RW-NH2-373 - 3394 | -30.00 | 4.50 | RW-NH2-374 - 3395 | -30.00 | 5.00 | RW-NH2-375 - 3396 | -30.00 | 5.50 | RW-NH2-376 - 3397 | -30.00 | 6.00 | RW-NH2-377 - 3398 | -30.00 | 6.50 | RW-NH2-378 - 3399 | -30.00 | 7.00 | RW-NH2-379 - 3400 | -30.00 | 7.50 | RW-NH2-380 - 3401 | -30.00 | 8.00 | RW-NH2-381 - 3402 | -30.00 | 8.50 | RW-NH2-382 - 3403 | -30.00 | 9.00 | RW-NH2-383 - 3404 | -30.00 | 9.50 | RW-NH2-384 - 3405 | -30.00 | 10.00 | RW-NH2-385 - 3406 | -30.00 | 10.50 | RW-NH2-386 - 3407 | -30.00 | 11.00 | RW-NH2-387 - 3408 | -30.00 | 11.50 | RW-NH2-388 - 3409 | -30.00 | 12.00 | RW-NH2-389 - 3410 | -30.00 | 12.50 | RW-NH2-390 - 3411 | -30.00 | 13.00 | RW-NH2-391 - 3412 | -30.00 | 13.50 | RW-NH2-392 - 3413 | -30.00 | 14.00 | RW-NH2-393 - 3414 | -30.00 | 14.50 | RW-NH2-394 - 3415 | -30.00 | 15.00 | RW-NH2-395 - 3416 | -30.00 | 15.50 | RW-NH2-396 - 3417 | -30.00 | 16.00 | RW-NH2-397 - 3418 | -30.00 | 16.50 | RW-NH2-398 - 3419 | -30.00 | 17.00 | RW-NH2-399 - 3420 | -30.00 | 17.50 | RW-NH2-400 - 3421 | -30.00 | 18.00 | RW-NH2-401 - 3422 | -30.00 | 18.50 | RW-NH2-402 - 3423 | -30.00 | 19.00 | RW-NH2-403 - 3424 | -30.00 | 19.50 | RW-NH2-404 - 3425 | -30.00 | 20.00 | RW-NH2-405 - 3426 | -30.00 | 20.50 | RW-NH2-406 - 3427 | -30.00 | 21.00 | RW-NH2-407 - 3428 | -30.00 | 21.50 | RW-NH2-408 - 3429 | -30.00 | 22.00 | RW-NH2-409 - 3430 | -30.00 | 22.50 | RW-NH2-410 - 3431 | -30.00 | 23.00 | RW-NH2-411 - 3432 | -30.00 | 23.50 | RW-NH2-412 - 3433 | -30.00 | 24.00 | RW-NH2-413 - 3434 | -30.00 | 24.50 | RW-NH2-414 - 3435 | -30.00 | 25.00 | RW-NH2-415 - 3436 | -30.00 | 25.60 | RW-NH2-416 - 3437 | -30.00 | 26.00 | RW-NH2-417 - 3438 | -30.00 | 26.50 | RW-NH2-418 - 3439 | -30.00 | 27.00 | RW-NH2-419 - 3440 | -30.00 | 27.50 | RW-NH2-420 - 3441 | -30.00 | 28.00 | RW-NH2-421 - 3442 | -30.00 | 28.50 | RW-NH2-422 - 3443 | -30.00 | 29.00 | RW-NH2-423 - 3444 | -30.00 | 29.50 | RW-NH2-424 - 3445 | -30.00 | 30.00 | RW-NH2-425 - 3446 | -30.00 | 30.50 | RW-NH2-426 - 3447 | -30.00 | 31.00 | RW-NH2-427 - 3448 | -30.00 | 31.50 | RW-NH2-428 - 3449 | -30.00 | 32.00 | RW-NH2-429 - 3450 | -141.00 | -19.50 | RW-NH1-51 - 3451 | -141.00 | -19.00 | RW-NH1-52 - 3452 | -141.00 | -18.50 | RW-NH1-53 - 3453 | -141.00 | -18.00 | RW-NH1-54 - 3454 | -141.00 | -17.50 | RW-NH1-55 - 3455 | -141.00 | -17.00 | RW-NH1-56 - 3456 | -141.00 | -16.50 | RW-NH1-57 - 3457 | -141.00 | -16.00 | RW-NH1-58 - 3458 | -141.00 | -15.50 | RW-NH1-59 - 3459 | -141.00 | -15.00 | RW-NH1-60 - 3460 | -141.00 | -14.50 | RW-NH1-61 - 3461 | -141.00 | -14.00 | RW-NH1-62 - 3462 | -141.00 | -13.50 | RW-NH1-63 - 3463 | -141.00 | -13.00 | RW-NH1-64 - 3464 | -141.00 | -12.50 | RW-NH1-65 - 3465 | -141.00 | -12.00 | RW-NH1-66 - 3466 | -141.00 | -11.50 | RW-NH1-67 - 3467 | -141.00 | -11.00 | RW-NH1-68 - 3468 | -141.00 | -10.50 | RW-NH1-69 - 3469 | -141.00 | -10.00 | RW-NH1-70 - 3470 | -141.00 | -9.50 | RW-NH1-71 - 3471 | -141.00 | -9.00 | RW-NH1-72 - 3472 | -141.00 | -8.50 | RW-NH1-73 - 3473 | -141.00 | -8.00 | RW-NH1-74 - 3474 | -141.00 | -7.50 | RW-NH1-75 - 3475 | -141.00 | -7.00 | RW-NH1-76 - 3476 | -141.00 | -6.50 | RW-NH1-77 - 3477 | -141.00 | -6.00 | RW-NH1-78 - 3478 | -141.00 | -5.50 | RW-NH1-79 - 3479 | -141.00 | -5.00 | RW-NH1-80 - 3480 | -141.00 | -4.50 | RW-NH1-81 - 3481 | -141.00 | -4.00 | RW-NH1-82 - 3482 | -141.00 | -3.50 | RW-NH1-83 - 3483 | -141.00 | -3.00 | RW-NH1-84 - 3484 | -141.00 | -2.50 | RW-NH1-85 - 3485 | -141.00 | -2.00 | RW-NH1-86 - 3486 | -141.00 | -1.50 | RW-NH1-87 - 3487 | -141.00 | -1.00 | RW-NH1-88 - 3488 | -141.00 | -0.50 | RW-NH1-89 - 3489 | -141.00 | 0.00 | RW-NH1-90 - 3490 | -141.00 | 0.50 | RW-NH1-91 - 3491 | -141.00 | 1.00 | RW-NH1-92 - 3492 | -141.00 | 1.50 | RW-NH1-93 - 3493 | -141.00 | 2.00 | RW-NH1-94 - 3494 | -141.00 | 2.50 | RW-NH1-95 - 3495 | -141.00 | 3.00 | RW-NH1-96 - 3496 | -141.00 | 3.50 | RW-NH1-97 - 3497 | -141.00 | 4.00 | RW-NH1-98 - 3498 | -141.00 | 4.50 | RW-NH1-99 - 3499 | -141.00 | 5.00 | RW-NH1-100 - 3500 | -141.00 | 5.50 | RW-NH1-101 - 3501 | -141.00 | 6.00 | RW-NH1-102 - 3502 | -141.00 | 6.50 | RW-NH1-103 - 3503 | -141.00 | 7.00 | RW-NH1-104 - 3504 | -141.00 | 7.50 | RW-NH1-105 - 3505 | -141.00 | 8.00 | RW-NH1-106 - 3506 | -141.00 | 8.50 | RW-NH1-107 - 3507 | -141.00 | 9.00 | RW-NH1-108 - 3508 | -141.00 | 9.50 | RW-NH1-109 - 3509 | -141.00 | 10.00 | RW-NH1-110 - 3510 | -141.00 | 10.50 | RW-NH1-111 - 3511 | -141.00 | 11.00 | RW-NH1-112 - 3512 | -141.00 | 11.50 | RW-NH1-113 - 3513 | -141.00 | 12.00 | RW-NH1-114 - 3514 | -141.00 | 12.50 | RW-NH1-115 - 3515 | -141.00 | 13.00 | RW-NH1-116 - 3516 | -141.00 | 13.50 | RW-NH1-117 - 3517 | -141.00 | 14.00 | RW-NH1-118 - 3518 | -141.00 | 14.50 | RW-NH1-119 - 3519 | -141.00 | 15.00 | RW-NH1-120 - 3520 | -141.00 | 15.50 | RW-NH1-121 - 3521 | -141.00 | 16.00 | RW-NH1-122 - 3522 | -141.00 | 16.50 | RW-NH1-123 - 3523 | -141.00 | 17.00 | RW-NH1-124 - 3524 | -141.00 | 17.50 | RW-NH1-125 - 3525 | -141.00 | 18.00 | RW-NH1-126 - 3526 | -141.00 | 18.50 | RW-NH1-127 - 3527 | -141.00 | 19.00 | RW-NH1-128 - 3528 | -141.00 | 19.50 | RW-NH1-129 - 3529 | -141.00 | 20.00 | RW-NH1-130 - 3530 | -141.00 | 20.50 | RW-NH1-131 - 3531 | -141.00 | 21.00 | RW-NH1-132 - 3532 | -141.00 | 21.50 | RW-NH1-133 - 3533 | -141.00 | 22.00 | RW-NH1-134 - 3534 | -141.00 | 22.50 | RW-NH1-135 - 3535 | -141.00 | 23.00 | RW-NH1-136 - 3536 | -141.00 | 23.50 | RW-NH1-137 - 3537 | -141.00 | 24.00 | RW-NH1-138 - 3538 | -141.00 | 24.50 | RW-NH1-139 - 3539 | -141.00 | 25.00 | RW-NH1-140 - 3540 | -141.00 | 25.50 | RW-NH1-141 - 3541 | -141.00 | 26.00 | RW-NH1-142 - 3542 | -141.00 | 26.50 | RW-NH1-143 - 3543 | -141.00 | 27.00 | RW-NH1-144 - 3544 | -141.00 | 27.50 | RW-NH1-145 - 3545 | -141.00 | 28.00 | RW-NH1-146 - 3546 | -141.00 | 28.50 | RW-NH1-147 - 3547 | -141.00 | 29.00 | RW-NH1-148 - 3548 | -141.00 | 29.50 | RW-NH1-149 - 3549 | -141.00 | 30.00 | RW-NH1-150 - 3550 | -141.00 | 30.50 | RW-NH1-151 - 3551 | -141.00 | 31.00 | RW-NH1-152 - 3552 | -141.00 | 31.50 | RW-NH1-153 - 3553 | -141.00 | 32.00 | RW-NH1-154 - 3554 | -141.00 | 32.50 | RW-NH1-155 - 3555 | -141.00 | 32.80 | RW-NH1-156 - 3556 | -141.00 | 32.80 | RW-NH1-157 - 3557 | -140.50 | 32.80 | RW-NH1-158 - 3558 | -140.00 | 32.80 | RW-NH1-159 - 3559 | -139.50 | 32.80 | RW-NH1-160 - 3560 | -139.00 | 32.80 | RW-NH1-161 - 3561 | -138.50 | 32.80 | RW-NH1-162 - 3562 | -138.00 | 32.80 | RW-NH1-163 - 3563 | -137.50 | 32.80 | RW-NH1-164 - 3564 | -137.00 | 32.80 | RW-NH1-165 - 3565 | -136.50 | 32.80 | RW-NH1-166 - 3566 | -136.00 | 32.80 | RW-NH1-167 - 3567 | -135.50 | 32.80 | RW-NH1-168 - 3568 | -135.00 | 32.80 | RW-NH1-169 - 3569 | -134.50 | 32.80 | RW-NH1-170 - 3570 | -134.00 | 32.80 | RW-NH1-171 - 3571 | -133.50 | 32.80 | RW-NH1-172 - 3572 | -133.00 | 32.80 | RW-NH1-173 - 3573 | -132.50 | 32.80 | RW-NH1-174 - 3574 | -132.00 | 32.80 | RW-NH1-175 - 3575 | -131.50 | 32.80 | RW-NH1-176 - 3576 | -131.00 | 32.80 | RW-NH1-177 - 3577 | -130.50 | 32.80 | RW-NH1-178 - 3578 | -130.00 | 32.80 | RW-NH1-179 - 3579 | -129.50 | 32.80 | RW-NH1-180 - 3580 | -129.00 | 32.80 | RW-NH1-181 - 3581 | -128.50 | 32.80 | RW-NH1-182 - 3582 | -128.00 | 32.80 | RW-NH1-183 - 3583 | -127.50 | 32.80 | RW-NH1-184 - 3584 | -127.00 | 32.80 | RW-NH1-185 - 3585 | -126.50 | 32.80 | RW-NH1-186 - 3586 | -126.00 | 32.80 | RW-NH1-187 - 3587 | -125.50 | 32.80 | RW-NH1-188 - 3588 | -125.00 | 32.80 | RW-NH1-189 - 3589 | -124.50 | 32.80 | RW-NH1-190 - 3590 | -124.00 | 32.80 | RW-NH1-191 - 3591 | -123.50 | 32.80 | RW-NH1-192 - 3592 | -123.00 | 32.80 | RW-NH1-193 - 3593 | -122.50 | 32.80 | RW-NH1-194 - 3594 | -122.00 | 32.80 | RW-NH1-195 - 3595 | -121.50 | 32.80 | RW-NH1-196 - 3596 | -121.00 | 32.80 | RW-NH1-197 - 3597 | -120.50 | 32.80 | RW-NH1-198 - 3598 | -120.00 | 32.80 | RW-NH1-199 - 3599 | -119.50 | 32.80 | RW-NH1-200 - 3600 | -119.00 | 32.80 | RW-NH1-201 - 3601 | -118.50 | 32.80 | RW-NH1-202 - 3602 | -81.25 | 31.00 | RW-NH1-206 - 3603 | -81.00 | 31.00 | RW-NH1-207 - 3604 | -80.50 | 31.00 | RW-NH1-208 - 3605 | -80.00 | 31.00 | RW-NH1-209 - 3606 | -79.50 | 31.00 | RW-NH1-210 - 3607 | -79.00 | 31.00 | RW-NH1-211 - 3608 | -78.50 | 31.00 | RW-NH1-212 - 3609 | -78.00 | 31.00 | RW-NH1-213 - 3610 | -77.50 | 31.00 | RW-NH1-214 - 3611 | -77.00 | 31.00 | RW-NH1-215 - 3612 | -76.50 | 31.00 | RW-NH1-216 - 3613 | -76.00 | 31.00 | RW-NH1-217 - 3614 | -75.50 | 31.00 | RW-NH1-218 - 3615 | -75.00 | 31.00 | RW-NH1-219 - 3616 | -74.50 | 31.00 | RW-NH1-220 - 3617 | -74.00 | 31.00 | RW-NH1-221 - 3618 | -73.50 | 31.00 | RW-NH1-222 - 3619 | -73.00 | 31.00 | RW-NH1-223 - 3620 | -72.50 | 31.00 | RW-NH1-224 - 3621 | -72.00 | 31.00 | RW-NH1-225 - 3622 | -71.50 | 31.00 | RW-NH1-226 - 3623 | -71.00 | 31.00 | RW-NH1-227 - 3624 | -70.50 | 31.00 | RW-NH1-228 - 3625 | -70.00 | 31.00 | RW-NH1-229 - 3626 | -70.00 | 31.00 | RW-NH1-230 - 3627 | -70.00 | 30.50 | RW-NH1-231 - 3628 | -70.00 | 30.00 | RW-NH1-232 - 3629 | -70.00 | 29.50 | RW-NH1-233 - 3630 | -70.00 | 29.00 | RW-NH1-234 - 3631 | -70.00 | 28.50 | RW-NH1-235 - 3632 | -70.00 | 28.00 | RW-NH1-236 - 3633 | -70.00 | 27.50 | RW-NH1-237 - 3634 | -70.00 | 27.00 | RW-NH1-238 - 3635 | -70.00 | 26.50 | RW-NH1-239 - 3636 | -70.00 | 26.00 | RW-NH1-240 - 3637 | -70.00 | 25.50 | RW-NH1-241 - 3638 | -70.00 | 25.00 | RW-NH1-242 - 3639 | -70.00 | 24.50 | RW-NH1-243 - 3640 | -70.00 | 24.00 | RW-NH1-244 - 3641 | -70.00 | 23.50 | RW-NH1-245 - 3642 | -70.00 | 23.00 | RW-NH1-246 - 3643 | -70.00 | 22.50 | RW-NH1-247 - 3644 | -70.00 | 22.00 | RW-NH1-248 - 3645 | -70.00 | 21.50 | RW-NH1-249 - 3646 | -70.00 | 21.00 | RW-NH1-250 - 3647 | -70.00 | 20.50 | RW-NH1-251 - 3648 | -70.00 | 20.00 | RW-NH1-252 - 3649 | -70.00 | 18.00 | RW-NH1-253 - 3650 | -70.00 | 17.50 | RW-NH1-254 - 3651 | -70.00 | 17.00 | RW-NH1-255 - 3652 | -70.00 | 16.50 | RW-NH1-256 - 3653 | -70.00 | 16.00 | RW-NH1-257 - 3654 | -70.00 | 15.50 | RW-NH1-258 - 3655 | -70.00 | 15.00 | RW-NH1-259 - 3656 | -70.00 | 14.50 | RW-NH1-260 - 3657 | -70.00 | 14.00 | RW-NH1-261 - 3658 | -70.00 | 13.50 | RW-NH1-262 - 3659 | -70.00 | 13.00 | RW-NH1-263 - 3660 | -70.00 | 12.50 | RW-NH1-264 - 3661 | -71.00 | -19.50 | RW-NH1-266 - 3662 | -71.50 | -19.50 | RW-NH1-267 - 3663 | -72.00 | -19.50 | RW-NH1-268 - 3664 | -72.50 | -19.50 | RW-NH1-269 - 3665 | -73.00 | -19.50 | RW-NH1-270 - 3666 | -73.50 | -19.50 | RW-NH1-271 - 3667 | -74.00 | -19.50 | RW-NH1-272 - 3668 | -74.50 | -19.50 | RW-NH1-273 - 3669 | -75.00 | -19.50 | RW-NH1-274 - 3670 | -75.50 | -19.50 | RW-NH1-275 - 3671 | -76.00 | -19.50 | RW-NH1-276 - 3672 | -76.50 | -19.50 | RW-NH1-277 - 3673 | -77.00 | -19.50 | RW-NH1-278 - 3674 | -77.50 | -19.50 | RW-NH1-279 - 3675 | -78.00 | -19.50 | RW-NH1-280 - 3676 | -78.50 | -19.50 | RW-NH1-281 - 3677 | -79.00 | -19.50 | RW-NH1-282 - 3678 | -79.50 | -19.50 | RW-NH1-283 - 3679 | -80.00 | -19.50 | RW-NH1-284 - 3680 | -80.50 | -19.50 | RW-NH1-285 - 3681 | -81.00 | -19.50 | RW-NH1-286 - 3682 | -81.50 | -19.50 | RW-NH1-287 - 3683 | -82.00 | -19.50 | RW-NH1-288 - 3684 | -82.50 | -19.50 | RW-NH1-289 - 3685 | -83.00 | -19.50 | RW-NH1-290 - 3686 | -83.50 | -19.50 | RW-NH1-291 - 3687 | -84.00 | -19.50 | RW-NH1-292 - 3688 | -84.50 | -19.50 | RW-NH1-293 - 3689 | -85.00 | -19.50 | RW-NH1-294 - 3690 | -85.50 | -19.50 | RW-NH1-295 - 3691 | -86.00 | -19.50 | RW-NH1-296 - 3692 | -86.50 | -19.50 | RW-NH1-297 - 3693 | -87.00 | -19.50 | RW-NH1-298 - 3694 | -87.50 | -19.50 | RW-NH1-299 - 3695 | -88.00 | -19.50 | RW-NH1-300 - 3696 | -88.50 | -19.50 | RW-NH1-301 - 3697 | -89.00 | -19.50 | RW-NH1-302 - 3698 | -89.50 | -19.50 | RW-NH1-303 - 3699 | -90.00 | -19.50 | RW-NH1-304 - 3700 | -90.50 | -19.50 | RW-NH1-305 - 3701 | -91.00 | -19.50 | RW-NH1-306 - 3702 | -91.50 | -19.50 | RW-NH1-307 - 3703 | -92.00 | -19.50 | RW-NH1-308 - 3704 | -92.50 | -19.50 | RW-NH1-309 - 3705 | -93.00 | -19.50 | RW-NH1-310 - 3706 | -93.50 | -19.50 | RW-NH1-311 - 3707 | -94.00 | -19.50 | RW-NH1-312 - 3708 | -94.50 | -19.50 | RW-NH1-313 - 3709 | -95.00 | -19.50 | RW-NH1-314 - 3710 | -95.50 | -19.50 | RW-NH1-315 - 3711 | -96.00 | -19.50 | RW-NH1-316 - 3712 | -96.50 | -19.50 | RW-NH1-317 - 3713 | -97.00 | -19.50 | RW-NH1-318 - 3714 | -97.50 | -19.50 | RW-NH1-319 - 3715 | -98.00 | -19.50 | RW-NH1-320 - 3716 | -98.50 | -19.50 | RW-NH1-321 - 3717 | -99.00 | -19.50 | RW-NH1-322 - 3718 | -99.50 | -19.50 | RW-NH1-323 - 3719 | -100.00 | -19.50 | RW-NH1-324 - 3720 | -100.50 | -19.50 | RW-NH1-325 - 3721 | -101.00 | -19.50 | RW-NH1-326 - 3722 | -101.50 | -19.50 | RW-NH1-327 - 3723 | -102.00 | -19.50 | RW-NH1-328 - 3724 | -102.50 | -19.50 | RW-NH1-329 - 3725 | -103.00 | -19.50 | RW-NH1-330 - 3726 | -103.50 | -19.50 | RW-NH1-331 - 3727 | -104.00 | -19.50 | RW-NH1-332 - 3728 | -104.50 | -19.50 | RW-NH1-333 - 3729 | -102.00 | -19.50 | RW-NH1-334 - 3730 | -102.50 | -19.50 | RW-NH1-335 - 3731 | -103.00 | -19.50 | RW-NH1-336 - 3732 | -103.50 | -19.50 | RW-NH1-337 - 3733 | -104.00 | -19.50 | RW-NH1-338 - 3734 | -104.50 | -19.50 | RW-NH1-339 - 3735 | -105.00 | -19.50 | RW-NH1-340 - 3736 | -102.50 | -19.50 | RW-NH1-341 - 3737 | -103.00 | -19.50 | RW-NH1-342 - 3738 | -103.50 | -19.50 | RW-NH1-343 - 3739 | -104.00 | -19.50 | RW-NH1-344 - 3740 | -104.50 | -19.50 | RW-NH1-345 - 3741 | -105.00 | -19.50 | RW-NH1-346 - 3742 | -105.50 | -19.50 | RW-NH1-347 - 3743 | -106.00 | -19.50 | RW-NH1-348 - 3744 | -106.50 | -19.50 | RW-NH1-349 - 3745 | -107.00 | -19.50 | RW-NH1-350 - 3746 | -107.50 | -19.50 | RW-NH1-351 - 3747 | -108.00 | -19.50 | RW-NH1-352 - 3748 | -105.50 | -19.50 | RW-NH1-353 - 3749 | -106.00 | -19.50 | RW-NH1-354 - 3750 | -106.50 | -19.50 | RW-NH1-355 - 3751 | -107.00 | -19.50 | RW-NH1-356 - 3752 | -107.50 | -19.50 | RW-NH1-357 - 3753 | -108.00 | -19.50 | RW-NH1-358 - 3754 | -108.50 | -19.50 | RW-NH1-359 - 3755 | -109.00 | -19.50 | RW-NH1-360 - 3756 | -109.50 | -19.50 | RW-NH1-361 - 3757 | -110.00 | -19.50 | RW-NH1-362 - 3758 | -110.50 | -19.50 | RW-NH1-363 - 3759 | -111.00 | -19.50 | RW-NH1-364 - 3760 | -111.50 | -19.50 | RW-NH1-365 - 3761 | -112.00 | -19.50 | RW-NH1-366 - 3762 | -112.50 | -19.50 | RW-NH1-367 - 3763 | -113.00 | -19.50 | RW-NH1-368 - 3764 | -113.50 | -19.50 | RW-NH1-369 - 3765 | -114.00 | -19.50 | RW-NH1-370 - 3766 | -114.50 | -19.50 | RW-NH1-371 - 3767 | -115.00 | -19.50 | RW-NH1-372 - 3768 | -115.50 | -19.50 | RW-NH1-373 - 3769 | -116.00 | -19.50 | RW-NH1-374 - 3770 | -116.50 | -19.50 | RW-NH1-375 - 3771 | -117.00 | -19.50 | RW-NH1-376 - 3772 | -117.50 | -19.50 | RW-NH1-377 - 3773 | -118.00 | -19.50 | RW-NH1-378 - 3774 | -118.50 | -19.50 | RW-NH1-379 - 3775 | -119.00 | -19.50 | RW-NH1-380 - 3776 | -119.50 | -19.50 | RW-NH1-381 - 3777 | -120.00 | -19.50 | RW-NH1-382 - 3778 | -120.50 | -19.50 | RW-NH1-383 - 3779 | -121.00 | -19.50 | RW-NH1-384 - 3780 | -121.50 | -19.50 | RW-NH1-385 - 3781 | -122.00 | -19.50 | RW-NH1-386 - 3782 | -122.50 | -19.50 | RW-NH1-387 - 3783 | -123.00 | -19.50 | RW-NH1-388 - 3784 | -123.50 | -19.50 | RW-NH1-389 - 3785 | -124.00 | -19.50 | RW-NH1-390 - 3786 | -124.50 | -19.50 | RW-NH1-391 - 3787 | -125.00 | -19.50 | RW-NH1-392 - 3788 | -125.50 | -19.50 | RW-NH1-393 - 3789 | -126.00 | -19.50 | RW-NH1-394 - 3790 | -126.50 | -19.50 | RW-NH1-395 - 3791 | -127.00 | -19.50 | RW-NH1-396 - 3792 | -127.50 | -19.50 | RW-NH1-397 - 3793 | -128.00 | -19.50 | RW-NH1-398 - 3794 | -128.50 | -19.50 | RW-NH1-399 - 3795 | -129.00 | -19.50 | RW-NH1-400 - 3796 | -129.50 | -19.50 | RW-NH1-401 - 3797 | -130.00 | -19.50 | RW-NH1-402 - 3798 | -130.50 | -19.50 | RW-NH1-403 - 3799 | -131.00 | -19.50 | RW-NH1-404 - 3800 | -131.50 | -19.50 | RW-NH1-405 - 3801 | -132.00 | -19.50 | RW-NH1-406 - 3802 | -132.50 | -19.50 | RW-NH1-407 - 3803 | -133.00 | -19.50 | RW-NH1-408 - 3804 | -133.50 | -19.50 | RW-NH1-409 - 3805 | -134.00 | -19.50 | RW-NH1-410 - 3806 | -134.50 | -19.50 | RW-NH1-411 - 3807 | -135.00 | -19.50 | RW-NH1-412 - 3808 | -135.50 | -19.50 | RW-NH1-413 - 3809 | -136.00 | -19.50 | RW-NH1-414 - 3810 | -136.50 | -19.50 | RW-NH1-415 - 3811 | -137.00 | -19.50 | RW-NH1-416 - 3812 | -137.50 | -19.50 | RW-NH1-417 - 3813 | -138.00 | -19.50 | RW-NH1-418 - 3814 | -138.50 | -19.50 | RW-NH1-419 - 3815 | -139.00 | -19.50 | RW-NH1-420 - 3816 | -139.50 | -19.50 | RW-NH1-421 - 3817 | -140.00 | -19.50 | RW-NH1-422 - 3818 | -140.50 | -19.50 | RW-NH1-423 - 3819 | -141.00 | -19.50 | RW-NH1-424 - 3820 | 0.00 | 0.00 | STOPSTRING - - Initial time : 2021/03/23 12:00:00 UTC - Ice field time : date and time not set. -(wav_comp_nuopc:wavinit_ufs): WW3 timesteps from mod_def 3600.0 1800.0 1800.0 30.0 -(wav_comp_nuopc:wavinit_ufs): WW3 timesteps 3600.0 1800.0 1800.0 30.0 -(wav_comp_nuopc:wavinit_ufs): WW3 additional restart field : 1 ice - - - | input | output | - |-----------------------|------------------| - step | pass | date time | b w l c t r i i1 i5 d | g p t r b f c r2 | - --------|------|---------------------|-----------------------|------------------| - --------+------+---------------------+---------------------------+--------------+ - 0| 1| 2021/03/23 12:00:00 | F F I | X X | - 1| 1| 13:00:00 | X X | X X | - --------+------+---------------------+-----------------------+------------------+ - 2| 2| 14:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 3| 3| 15:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 4| 4| 16:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 5| 5| 17:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 6| 6| 18:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 7| 7| 19:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 8| 8| 20:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 9| 9| 21:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 10| 10| 22:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 11| 11| 23:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - Writing restart file ./WW3_RESTART/ufs.cpld.ww3.r.2021-03-24-00000.nc - 12| 12| 2021/03/24 00:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 13| 13| 01:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 14| 14| 02:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 15| 15| 03:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 16| 16| 04:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 17| 17| 05:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 18| 18| 06:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 19| 19| 07:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 20| 20| 08:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 21| 21| 09:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 22| 22| 10:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 23| 23| 11:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - Writing restart file ./WW3_RESTART/ufs.cpld.ww3.r.2021-03-24-43200.nc - 24| 24| 12:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 25| 25| 13:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 26| 26| 14:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 27| 27| 15:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 28| 28| 16:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 29| 29| 17:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 30| 30| 18:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 31| 31| 19:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 32| 32| 20:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 33| 33| 21:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 34| 34| 22:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 35| 35| 23:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - Writing restart file ./WW3_RESTART/ufs.cpld.ww3.r.2021-03-25-00000.nc - 36| 36| 2021/03/25 00:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 37| 37| 01:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 38| 38| 02:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 39| 39| 03:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 40| 40| 04:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 41| 41| 05:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 42| 42| 06:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 43| 43| 07:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 44| 44| 08:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 45| 45| 09:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 46| 46| 10:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 47| 47| 11:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - Writing restart file ./WW3_RESTART/ufs.cpld.ww3.r.2021-03-25-43200.nc - 48| 48| 12:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 49| 49| 13:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 50| 50| 14:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 51| 51| 15:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 52| 52| 16:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 53| 53| 17:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 54| 54| 18:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 55| 55| 19:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 56| 56| 20:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 57| 57| 21:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 58| 58| 22:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 59| 59| 23:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - Writing restart file ./WW3_RESTART/ufs.cpld.ww3.r.2021-03-26-00000.nc - 60| 60| 2021/03/26 00:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 61| 61| 01:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 62| 62| 02:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 63| 63| 03:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 64| 64| 04:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 65| 65| 05:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 66| 66| 06:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 67| 67| 07:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 68| 68| 08:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 69| 69| 09:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 70| 70| 10:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 71| 71| 11:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - Writing restart file ./WW3_RESTART/ufs.cpld.ww3.r.2021-03-26-43200.nc - 72| 72| 12:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 73| 73| 13:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 74| 74| 14:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 75| 75| 15:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 76| 76| 16:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 77| 77| 17:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 78| 78| 18:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 79| 79| 19:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 80| 80| 20:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 81| 81| 21:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 82| 82| 22:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 83| 83| 23:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - Writing restart file ./WW3_RESTART/ufs.cpld.ww3.r.2021-03-27-00000.nc - 84| 84| 2021/03/27 00:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 85| 85| 01:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 86| 86| 02:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 87| 87| 03:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 88| 88| 04:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 89| 89| 05:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 90| 90| 06:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 91| 91| 07:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 92| 92| 08:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 93| 93| 09:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 94| 94| 10:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 95| 95| 11:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - Writing restart file ./WW3_RESTART/ufs.cpld.ww3.r.2021-03-27-43200.nc - 96| 96| 12:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 97| 97| 13:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 98| 98| 14:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 99| 99| 15:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 100| 100| 16:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 101| 101| 17:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 102| 102| 18:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 103| 103| 19:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 104| 104| 20:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 105| 105| 21:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 106| 106| 22:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 107| 107| 23:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - Writing restart file ./WW3_RESTART/ufs.cpld.ww3.r.2021-03-28-00000.nc - 108| 108| 2021/03/28 00:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 109| 109| 01:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 110| 110| 02:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 111| 111| 03:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 112| 112| 04:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 113| 113| 05:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 114| 114| 06:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 115| 115| 07:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 116| 116| 08:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 117| 117| 09:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 118| 118| 10:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - 119| 119| 11:00:00 | X X U | X X | - --------+------+---------------------+-----------------------+------------------+ - Writing restart file ./WW3_RESTART/ufs.cpld.ww3.r.2021-03-28-43200.nc - 120| 120| 12:00:00 | X X U | L L | - --------+------+---------------------+-----------------------+------------------+ -(ww3_comp_nuopc) ------------------------------------------------------------------------- -(ww3_comp_nuopc) WW3: end of main integration loop -(ww3_comp_nuopc) ------------------------------------------------------------------------- diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f000.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f000.log deleted file mode 100644 index c41e8358..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f000.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.120000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f001.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f001.log deleted file mode 100644 index 605351c7..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f001.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.130000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f002.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f002.log deleted file mode 100644 index 9dae2c3c..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f002.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.140000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f003.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f003.log deleted file mode 100644 index 71fa13ab..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f003.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.150000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f004.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f004.log deleted file mode 100644 index 895adc75..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f004.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.160000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f005.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f005.log deleted file mode 100644 index b806c69c..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f005.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.170000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f006.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f006.log deleted file mode 100644 index 85ef023b..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f006.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.180000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f007.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f007.log deleted file mode 100644 index 9cb16580..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f007.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.190000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f008.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f008.log deleted file mode 100644 index 9cf7eb35..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f008.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.200000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f009.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f009.log deleted file mode 100644 index cbcf5703..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f009.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.210000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f010.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f010.log deleted file mode 100644 index 4b94c227..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f010.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.220000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f011.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f011.log deleted file mode 100644 index d104fca5..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f011.log +++ /dev/null @@ -1 +0,0 @@ - The 20210323.230000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f012.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f012.log deleted file mode 100644 index a4d640d4..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f012.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.000000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f013.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f013.log deleted file mode 100644 index dab3b4eb..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f013.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.010000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f014.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f014.log deleted file mode 100644 index 6b855bb7..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f014.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.020000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f015.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f015.log deleted file mode 100644 index c6e8427d..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f015.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.030000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f016.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f016.log deleted file mode 100644 index ede366b7..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f016.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.040000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f017.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f017.log deleted file mode 100644 index 6259be3c..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f017.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.050000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f018.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f018.log deleted file mode 100644 index 00d13926..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f018.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.060000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f019.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f019.log deleted file mode 100644 index 33f92a84..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f019.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.070000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f020.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f020.log deleted file mode 100644 index aa084f79..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f020.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.080000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f021.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f021.log deleted file mode 100644 index 3ff7db7e..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f021.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.090000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f022.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f022.log deleted file mode 100644 index 938afdc4..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f022.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.100000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f023.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f023.log deleted file mode 100644 index d4bb52f0..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f023.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.110000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f024.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f024.log deleted file mode 100644 index e03acec2..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f024.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.120000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f025.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f025.log deleted file mode 100644 index e4bbb1dd..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f025.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.130000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f026.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f026.log deleted file mode 100644 index 6a92a828..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f026.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.140000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f027.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f027.log deleted file mode 100644 index 35ddcf4f..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f027.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.150000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f028.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f028.log deleted file mode 100644 index 9a2bdc25..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f028.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.160000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f029.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f029.log deleted file mode 100644 index ac32b722..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f029.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.170000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f030.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f030.log deleted file mode 100644 index 6d8afa11..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f030.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.180000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f031.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f031.log deleted file mode 100644 index dbca7ca8..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f031.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.190000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f032.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f032.log deleted file mode 100644 index 7cccfe68..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f032.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.200000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f033.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f033.log deleted file mode 100644 index e37916a5..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f033.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.210000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f034.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f034.log deleted file mode 100644 index 89237397..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f034.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.220000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f035.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f035.log deleted file mode 100644 index 79ffaa2b..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f035.log +++ /dev/null @@ -1 +0,0 @@ - The 20210324.230000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f036.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f036.log deleted file mode 100644 index 4a82d291..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f036.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.000000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f037.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f037.log deleted file mode 100644 index a97b7840..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f037.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.010000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f038.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f038.log deleted file mode 100644 index 68ac3895..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f038.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.020000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f039.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f039.log deleted file mode 100644 index 675d2536..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f039.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.030000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f040.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f040.log deleted file mode 100644 index 260e5115..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f040.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.040000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f041.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f041.log deleted file mode 100644 index 0c9c4fe3..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f041.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.050000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f042.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f042.log deleted file mode 100644 index 24112e7a..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f042.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.060000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f043.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f043.log deleted file mode 100644 index ed02192b..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f043.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.070000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f044.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f044.log deleted file mode 100644 index b0f3ed7d..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f044.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.080000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f045.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f045.log deleted file mode 100644 index 305812c9..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f045.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.090000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f046.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f046.log deleted file mode 100644 index dd48eebe..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f046.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.100000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f047.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f047.log deleted file mode 100644 index 7f17253b..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f047.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.110000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f048.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f048.log deleted file mode 100644 index 341cf5ab..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f048.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.120000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f051.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f051.log deleted file mode 100644 index 7ecd32a4..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f051.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.150000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f054.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f054.log deleted file mode 100644 index 0690d460..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f054.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.180000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f057.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f057.log deleted file mode 100644 index c5cd7ae7..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f057.log +++ /dev/null @@ -1 +0,0 @@ - The 20210325.210000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f060.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f060.log deleted file mode 100644 index 7afe6417..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f060.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.000000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f063.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f063.log deleted file mode 100644 index 814696fa..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f063.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.030000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f066.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f066.log deleted file mode 100644 index 5c935058..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f066.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.060000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f069.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f069.log deleted file mode 100644 index 9388cbaf..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f069.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.090000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f072.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f072.log deleted file mode 100644 index 1946e33e..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f072.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.120000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f075.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f075.log deleted file mode 100644 index 304b54d7..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f075.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.150000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f078.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f078.log deleted file mode 100644 index 43a6575a..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f078.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.180000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f081.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f081.log deleted file mode 100644 index fb6d1d38..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f081.log +++ /dev/null @@ -1 +0,0 @@ - The 20210326.210000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f084.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f084.log deleted file mode 100644 index 42e0989c..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f084.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.000000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f087.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f087.log deleted file mode 100644 index 86907709..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f087.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.030000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f090.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f090.log deleted file mode 100644 index a5d68361..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f090.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.060000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f093.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f093.log deleted file mode 100644 index de5a3582..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f093.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.090000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f096.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f096.log deleted file mode 100644 index 72e4540e..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f096.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.120000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f099.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f099.log deleted file mode 100644 index 7ef403d1..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f099.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.150000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f102.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f102.log deleted file mode 100644 index 62f8a340..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f102.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.180000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f105.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f105.log deleted file mode 100644 index b24d3102..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f105.log +++ /dev/null @@ -1 +0,0 @@ - The 20210327.210000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f108.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f108.log deleted file mode 100644 index ee2079d7..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f108.log +++ /dev/null @@ -1 +0,0 @@ - The 20210328.000000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f111.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f111.log deleted file mode 100644 index 87841a00..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f111.log +++ /dev/null @@ -1 +0,0 @@ - The 20210328.030000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f114.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f114.log deleted file mode 100644 index 8c884120..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f114.log +++ /dev/null @@ -1 +0,0 @@ - The 20210328.060000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f117.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f117.log deleted file mode 100644 index ff695c6d..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f117.log +++ /dev/null @@ -1 +0,0 @@ - The 20210328.090000.out_grd.ww3 file has been successfully written! diff --git a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f120.log b/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f120.log deleted file mode 100644 index 3de174ed..00000000 --- a/ci/error_logs/HECULES_PR_LOGS/model_wave_history_gfs.wave.t12z.uglo_100km.f120.log +++ /dev/null @@ -1 +0,0 @@ - The 20210328.120000.out_grd.ww3 file has been successfully written!